Sciweavers

ERLANG
2006
ACM

Concurrent caching

14 years 6 months ago
Concurrent caching
A concurrent cache design is presented which allows cached data to be spread across a cluster of computers. The implementation s persistent storage from cache storage and abstracts the cache behaviour so that the user can experiment with cache size and replacement policy to optimize performance for a given system, even if the production data store is not available. Using processes to implement cached objects allows for runtime configurability and adaptive use policies as well as parallelization to optimize resource access efficiency.
Jay Nelson
Added 13 Jun 2010
Updated 13 Jun 2010
Type Conference
Year 2006
Where ERLANG
Authors Jay Nelson
Comments (0)