Changes in Varnish 5.1¶
We have a couple of new and interesting features in Varnish 5.1, and we have a lot of smaller improvements and bugfixes all over the place, in total we have made about 750 commits since Varnish 5.0, so this is just some of the highlights.
Probably the biggest change in Varnish 5.1 is that a couple of very significant contributors to Varnish have changed jobs, and therefore stopped being active contributors to the Varnish Project.
Per Buer was one of the first people who realized that Varnish was not just “Some program for a couple of Nordic newspapers”, and he started the company Varnish Software, which is one of the major sponsors of the Varnish Project.
Lasse Karstensen got roped into Varnish Software by Per, and in addition to his other duties, he has taken care of the projects system administration and release engineering for most of the 11 years we have been around now.
Per & Lasse: “Thanks” doesn’t even start to cover it, and we wish you all the best for the future!
Startup CLI command file¶
The new ‘-I cli_file’ option to varnishd will make it much more practical to use the VCL labels introduced in Varnish 5.0.
The cli commands in the file will be executed before the worker process starts, so it could for instance contain:
vcl.load panic /etc/varnish_panic.vcl
vcl.load siteA0 /etc/varnish_siteA.vcl
vcl.load siteB0 /etc/varnish_siteB.vcl
vcl.load siteC0 /etc/varnish_siteC.vcl
vcl.label siteA siteA0
vcl.label siteB siteB0
vcl.label siteC siteC0
vcl.load main /etc/varnish_main.vcl
vcl.use main
If a command in the file is prefixed with ‘-’, failure will not abort the startup.
Related to this change we have reordered the argument checking so that argument problems are reported more consistently.
In case you didn’t hear about them yet, labelling VCL programs allows you to branch out to other VCLs in the main::vcl_recv{}, which in the above example could look like:
sub vcl_recv {
if (req.http.host ~ "asite.example.com$") {
return(vcl(siteA));
}
if (req.http.host ~ "bsite.example.com$") {
return(vcl(siteB));
}
if (req.http.host ~ "csite.example.com$") {
return(vcl(siteC));
}
// Main site processing ...
}
Universal VCL return(fail)¶
It is now possible to return(fail)
anywhere in VCL,
including inside VMODs. This will cause VCL processing
to terminate forthright.
In addition to return(fail)
, this mechanism will be
used to handle all failure conditions without a safe
fallback, for instance workspace exhaustion, too many
headers etc. (This is a work in progress, there is a
lot of code to review before we are done.)
In vcl_init{}
failing causes the vcl.load
to fail,
this is nothing new for this sub-routine.
A failure in any of the client side VCL methods (vcl_recv{}
,
vcl_hash{}
…) except vcl_synth{}
, sends the request
to vcl_synth{}
with a 503, and reason “VCL failed”.
A failure on the backend side (vcl_backend_*{}
) causes the
fetch to fail.
(VMOD writers should use the new VRT_fail(ctx, format_string, ...)
function which logs a SLT_VCL_Error record.)
Progress on HTTP/2 support¶
HTTP/2 support is better than in 5.0, and is now enabled and survives pretty well on our own varnish-cache.org website, but there are still things missing, most notably windows and priority, which may be fatal to more complex websites.
We expect HTTP/2 support to be production ready in the autumn 2017 release of Varnish-Cache, but that requires a testing and feedback from real-world applications.
So if you have a chance to test our HTTP/2 code, by all means do so, please report any crashes, bugs or other trouble back to us.
To enable HTTP/2 you need to param.set feature +http2
but due
to internet-politics, you will only see HTTP/2 traffic if you have
an SSL proxy in front of Varnish which advertises HTTP2 with ALPN.
For the hitch SSL proxy, add the argument --alpn-protos="h2,http/1.1"
Hit-For-Pass has returned¶
As hinted in Changes in Varnish 5.0, we have restored the
possibility of invoking the old hit-for-pass feature in VCL. The
treatment of uncacheable content that was new in version 5.0, which we
have taken to calling “hit-for-miss”, remains the default. Now you can
choose hit-for-pass with return(pass(DURATION))
from
vcl_backend_response
, setting the duration of the hit-for-pass
state in the argument to pass
. For example:
return(pass(120s))
.
To recap: when beresp.uncacheable
is set to true
in
vcl_backend_response
, Varnish makes a note of it with a minimal
object in the cache, and finds that information again on the next
lookup for the same object. In essence, the cache is used to remember
that the last backend response was not cacheable. In that case,
Varnish proceeds as with a cache miss, so that the response may become
cacheable on subsequent requests. The difference is that Varnish does
not perform request coalescing, as it does for ordinary misses, when a
response has been marked uncacheable. For ordinary misses, when there
are requests pending for the same object at the same time, only one
fetch is executed at a time, since the response may be cached, in
which case the cached response may be used for the remaining
requests. But this is not done for “hit-for-miss” objects, since they
are known to have been uncacheable on the previous fetch.
builtin.vcl
sets beresp.uncacheable
to true
when a number
of conditions hold for a backend response that indicate that it should
not be cached, for example if the TTL has been determined to be 0
(perhaps due to a Cache-Control
header), or if a Set-Cookie
header is present in the response. So hit-for-miss is the default
for uncacheable backend responses.
A consequence of this is that fetches for uncacheable responses cannot
be conditional in the default case. That is, the backend request may
not include the headers If-Modified-Since
or If-None-Match
,
which might cause the backend to return status “304 Not Modified” with
no response body. Since the response to a cache miss might be cached,
there has to be a body to cache, and this is true of hit-for-miss as
well. If either of those two headers were present in the client
request, they are removed from the backend request for a miss or
hit-for-miss.
Since conditional backend requests and the 304 response may be
critical to performance for non-cacheable content, especially if the
response body is large, we have made the old hit-for-pass feature
available again, with return(pass(DURATION))
in VCL.
As with hit-for-miss, Varnish uses the cache to make a note of
hit-for-pass objects, and finds them again on subsequent lookups. The
requests are then processed as for ordinary passes (return(pass)
from vcl_recv
) – there is no request coalescing, and the response
will not be cached, even if it might have been otherwise.
If-Modified-Since
or If-None-Match
headers in the client
request are passed along in the backend request, and a backend
response with status 304 and no body is passed back to the client.
The hit-for-pass state of an object lasts for the time given as the
DURATION in the previous return from vcl_backend_response
. After
the “hit-for-pass TTL” elapses, the next request will be an ordinary
miss. So a hit-for-pass object cannot become cacheable again until
that much time has passed.
304 Not Modified responses after a pass¶
Related to the previous topic, there has been a change in the way Varnish handles a very specific case: deciding whether to send a “304 Not Modified” response to the client after a pass, when the backend had the opportunity to send a 304 response, but chose not to by sending a 200 response status instead.
Previously, Varnish went along with the backend when this happened,
sending the 200 response together with the response body to the
client. This was the case even if the backend set the response headers
ETag
and/or Last-Modified
so that, when compared to the
request headers If-None-Match
and If-Modified-Since
, a 304
response would seem to be warranted. Since those headers are passed
back to the client, the result could appear a bit odd from the
client’s perspective – the client used the request headers to ask if
the response was unmodified, and the response headers seem to indicate
that it wasn’t, and yet the response status suggests that it was.
Now the decision to send a 304 client response status is made solely at delivery time, based on the contents of the client request headers and the headers in the response that Varnish is preparing to send, regardless of whether the backend fetch was a pass. So Varnish may send a 304 client response after a pass, even though the backend chose not to, having seen the same request headers (if the response headers permit it).
We made this change for consistency – for hits, misses, hit-for-miss, hit-for-pass, and now pass, the decision to send a 304 client response is based solely on the contents of client request headers and the response headers.
You can restore the previous behavior – don’t send a 304 client
response on pass if the backend didn’t – with VCL means, either by
removing the ETag
or Last-Modified
headers in
vcl_backend_response
, or by removing the If-* client request
headers in vcl_pass
.
VXID in VSL queries¶
The Varnish Shared Log (VSL) became much more powerful starting Varnish 4.0 and hasn’t changed much since. Changes usually consist in adding new log records when new feature are introduced, or when we realize that some missing piece of information could really help troubleshooting.
Varnish UTilities (VUT) relying on the VSL usually share the same -q
option for querying, which allows to filter transactions based on log
records. For example you could be looking for figures on a specific
domain:
varnishtop -i ReqURL -q 'ReqHeader:Host eq www.example.com'
While options like -i
and -q
were until now both limited to log
records, it also meant you could only query a specific transaction using
the X-Varnish
header. Depending on the nature of the transaction
(client or backend side) the syntax is not the same and you can’t match
a session.
For instance, we are looking for the transaction 1234 that occurred very recently and we would like to collect everything from the same session. We have two options:
# client side
varnishlog -d -g session -q 'RespHeader:X-Varnish[1] == 1234'
# backend side
varnishlog -d -g session -q 'BereqHeader:X-Varnish == 1234'
There was no simple way to match any transaction using its id until the
introduction of vxid
as a possible left-hand side of a -q
query
expression:
# client side
varnishlog -d -g session -q 'vxid == 1234'
# backend side
varnishlog -d -g session -q 'vxid == 1234'
# session
varnishlog -d -g session -q 'vxid == 1234'
Another use case is the collection of non-transactional logs. With raw grouping the output is organized differently and each record starts with its transaction id or zero for non-transactional logs:
# before 5.1
varnishlog -g raw | awk '$1 == 0'
# from now on
varnishlog -g raw -q 'vxid == 0'
This should offer you a more concise, and more consistent means to filter
transactions with varnishlog
and other VUTs.
Project tool improvements¶
We have spent a fair amount of time on the tools we use internally in the project.
The varnishtest
program has been improved in many small ways,
in particular it is now much easier to execute and examine
results from other programs with the shell
and process
commands. It might break existing test cases if you were already
using varnishtest
.
The project now has KISS web-backend which summarizes
make distcheck
results from various platforms:
http://varnish-cache.org/vtest/
If you want Varnish to be tested on a platform not already covered, all you need to do is run the tools/vtest.sh script from the source tree. We would love to see more platforms covered (arm64, ppc, mips) and OS/X would also be nice.
We also publish our code-coverage status now:
http://varnish-cache.org/gcov/
Our goal is 90+% coverage, but we need to start implementing
terminal emulation in varnishtest
before we can test the curses(1)
based programs (top/stat/hist) comprehensively, so they currently
drag us down.