This post is an introduction to LHIC, which is something I've been working on recently.
Also known as a reverse proxy cache, web accelerator, etc – it's a specific type of shared cache that is shared by all clients, and is generally within the same organisational boundary as the origin server.
It's a layer, and as such it may actually consist of one or many peered 'instances', this research does not define how such a peering mechanism would work between instances, and this is definitely an area for future work. For our purposes here we'll consider the gateway cache layer to be one complete component.
The primary objective for a gateway cache is to minimize demand on origin servers.
All caches work by leveraging one or more of the 3 principal caching mechanisms:
- Expiration
- Validation
- Invalidation
The downsides to expiration are primarily that it's inefficient, and difficult to manage.
It's inefficient because the expiry period is always limited in length to the resource's greatest potential volatility. This is particularly inefficient for resources which depend on human interaction and have periodic and dynamic volatility.
It's difficult to manage because the more efficiency you try to squeeze out of it, the greater the risk to the integrity of the cached information - this puts pressure on questions that are already very difficult to answer such as; What should the rules be? Where are those rules stored? How are those rules governed over time?
Ensuring freshness is a property of the validation mechanism which does have significant benefits, however this is at the expense of the server side which will still handle each request and incur processing and I/O costs. This is therefore not useful for gateway caching since the primary objective is minimizing demands on the server.
Using a combination of both expiration and validation will effectively give you the best of both worlds, but you will still inherit the problems that come with the expiration mechanism.
Invalidation-based caching works by keeping responses cached right up until their resource's state is altered by a client request. Therefore, in order to rely on invalidation exclusively, the cache must intermediate interactions from all clients, and is therefore a mechanism that is only really suited to gateway caches, and not to normal shared (i.e. forward proxy) or client-side caches.
In HTTP terms an invalidating request is any non-safe request that receives a 200 or 300 response, this sequence diagram demonstrates how invalidation can work in practice.
All REST constraints play important roles in enabling cache-ability in general, however the main enabler of invalidation is the uniform interface, and specifically self-descriptive messages. The uniform interface and self-descriptiveness of messages empower layering in REST and crucially it enables intermediaries like caches to make assertions about client-server interactions. It is these assertions that are the key to the invalidation mechanism.
What are the benefits of invalidation-based caching?
Invalidation-based caches have self control; they are governed naturally and dynamically by client-server interaction. This makes them much easier to manage, ensures freshness, and operates with best-case efficiency in which responses are only invalidated when absolutely necessary. This best-case efficiency results in responses being cached for the longest possible period minimising contact with origin server, and bandwidth consumed.
So.. why isn't this really used? Because there are common problems when building systems with HTTP, that cause the mechanism to fail.
There are other types of problem and variations on these two, they just happen to be the most common.
In a perfect world resources are granular and don't share state. At all. So - in the perfect world example above, the collection is simply a series of links. This does, however, require any client to make several subsequent requests for each item resource. This behaviour is generally considered overly 'chatty' and inefficient and therefore in the real world clear identification of resources and their state is traded-away for network efficiency gains.
This trade-off has consequences for invalidation..
How would an intermediary answer that question?
The right answer should be "None." or at least "I don't know."
Given that URI's are essentially opaque it should, as far as intermediaries are concerned, have no effect. Using assumptions of perceived URI hierarchy is brittle and restrictive - what happens if this item belongs to more than one composite collection?
It's common practice to treat representations as resources in their own right, and expose them with their own URIs. This creates the same kind of situation as the composite resource problem in which these 'representation resources' share state invisibly.
You could propose to solve this by making assumptions using 'dot notation' however this is again ignoring the opacity of URIs and is brittle and restrictive.
There are other examples in which the existing approach to invalidation is made impossible, but they all revolve around the same core problem:
Composite and split resources are problems because they reduce visibility.
Resources share state, and are therefore dependent, but the uniform interface lacks the capability to express this as control-data; and it is therefore not visible to intermediaries.
Link headers can be used to "beef up" the uniform interface by expressing these invisible dependencies as link relations.
Standardising the link relations allows these links to be used as control data within the uniform interface; thus increasing self-descriptiveness of messages and visibility.
This is named "Link Header-based Invalidation of Caches" (LHIC). There are two types of LHIC:
LHIC-I is a simple mechanism that can be thought of as "pointing out" affected resources in the response. This gives the origin server dynamic control over the invalidation.
In order to secure the LHIC-I mechanism from DoS attacks in which any/all cached objects could be indicated for invalidation, it is likely that a same-domain policy would have to be adopted.
The purpose of this type of link relation is to simply increase visibility of the invalidating interaction itself.
LHIC-II is a more complex mechanism that can be thought of as dependent resources "latching on" to one another. This effectively creates a 'dependency graph' within a gateway cache which can be queried against each invalidating request. LHIC-II is therefore capable of allowing invalidation to cascade along a chain of dependencies, whereas LHIC-I is only capable of handling first level dependencies.
The purpose of this type of link relation is to increase overall visibility in the system, ahead of an invalidation taking place.
LHIC-II does not suffer the drawbacks of LHIC-I, however it is more complex to implement and does not allow dynamic control of invalidation by origin servers.
The optimal approach is to implement both methods; this allows for both dynamic control by origin servers, and cascading invalidation.
Conclusion..
LHIC injects lost visibility back into the web
The resulting caching mechanism is
- Very efficient
- Ensures freshness
- Easily managed
- Leverages existing specs
Thoughts, comments, suggestions all welcome! :)