"Caching is a strategy to have the frequently accessed data in the memory to reduce the latency. whenever a client request the data from server, it is first accessed into cache and if not available then it is getting fetched from the storage and stored into cache. As cache is having limited memory, so amount of data can be stored in cache is less. Data will be flushed out from cache based on a criteria which is termed as caching strategy. There could be different mechanisms under which a data can"
Archna M. - "Caching is a strategy to have the frequently accessed data in the memory to reduce the latency. whenever a client request the data from server, it is first accessed into cache and if not available then it is getting fetched from the storage and stored into cache. As cache is having limited memory, so amount of data can be stored in cache is less. Data will be flushed out from cache based on a criteria which is termed as caching strategy. There could be different mechanisms under which a data can"See full answer