Skip to main content


The naive approach to implement caching is to just store everything in an in memory collection such as a hashtable. After all it works on my machine.

I've worked on systems in the past that used this technique but:

  • Bring in two processes and this falls apart
  • No Time to Live (TTL)
  • No cache eviction, memory will grow until it crashes the process

This sort of caching meant the system needed daily restarts due to each worker process starting to eat up more and more RAM. At the time I didn't realise this was the problem as to why daily restarts were required. These were automated so the team just sort of forgot about the problem after a while. This never felt right.

"Improper use of caching is the major cause of memory leaks, which turn into horrors like daily server restarts" - @mtnygard in Release It!.

Scale this system up, and daily becomes twice daily and so on. In a global market where software shouldn't be constrained by time zones or "working hours" this is wrong.


There are numerous easy ways to solve these problems depending on the application in question.

Don't Roll your Own, Try a Third Party

Easy. Just use an off the shelf solution that solves the problems above plus includes a whole host of additional features.

Use your Standard Library

For example .NET includes caching functionality within the System.Runtime.Caching namespace. While there are limitations to this, it will work for some scenarios and solves some of the problems above.

Soft References

I've overlooked soft references in the past but for caching they can be incredibly useful. Use soft references for anything that isn't important or that can be recalculated. An example would be content displayed within an MVC view using the web servers session. Here if each item stored is a weak reference we introduce some benefits.

  • Stops your web server running of of memory - references will be reclaimed if memory starts to become a bottleneck.
  • Greater scalability with the same amount of memory - great for a sudden spike in traffic.

A web server's session being full of references that won't expire for a set period is a common cause of downtime. If soft references are used all we need to do is perform a simple conditional check prior to retrieval from the session. Different languages have similar features, e.g. Weak References in .NET.


Caching isn't always the best solution, in some cases pre-computation can be much easier and offer better performance. In other words at least some users will experience a slow response until the cache is warm, other techniques can be used to avoid this completely. I will expand on pre-computation in a future post.


More information can be found in the excellent book Release It!


Popular posts from this blog

Constant Object Anti Pattern

Most constants are used to remove magic numbers or variables that lack context. A classic example would be code littered with the number 7. What does this refer to exactly? If this was replaced with DaysInWeek or similar, much clarity is provided. You can determine that code performing offsets would be adding days, rather than a mysterious number seven.Sadly a common pattern which uses constants is the use of a single constant file or object. The beauty of constants is clarity, and the obvious fact such variables are fixed. When a constant container is used, constants are simply lumped together. These can grow in size and often become a dumping ground for all values within the application.A disadvantage of this pattern is the actual value is hidden. While a friendly variable name is great, there will come a time where you will want to know the actual value. This forces you to navigate, if only to peek at the value within the constant object. A solution is to simple perform a refactor …

Three Steps to Code Quality via TDD

Common complaints and problems that I've both encountered and hear other developers raise when it comes to the practice of Test Driven Development are: Impossible to refactor without all the tests breakingMinor changes require hours of changes to test codeTest setup is huge, slow to write and difficult to understandThe use of test doubles (mocks, stubs and fakes is confusing)Over the next three posts I will demonstrate three easy steps that can resolve the problems above. In turn this will allow developers to gain one of the benefits that TDD promises - the ability to refactor your code mercifully in order to improve code quality.StepsStop Making Everything PublicLimit the Amount of Dependencies you Use A Unit is Not Always a Method or ClassCode quality is a tricky subject and highly subjective, however if you follow the three guidelines above you should have the ability to radically change implementation details and therefore improve code quality when needed.

DRY vs DAMP in Tests

In the previous post I mentioned that duplication in tests is not always bad. Sometimes duplication becomes a problem. Tests can become large or virtually identically excluding a few lines. Changes to these tests can take a while and increase the maintenance overhead. At this point, DRY violations need to be resolved.SolutionsTest HelpersA common solution is to extract common functionality into setup methods or other helper utilities. While this will remove and reduce duplication this can make tests a bit harder to read as the test is now split amongst unrelated components. There is a limit to how useful such extractions can help as each test may need to do something slightly differently.DAMP - Descriptive and Meaningful PhrasesDescriptive and Meaningful Phrases is the alter ego of DRY. DAMP tests often use the builder pattern to construct the System Under Test. This allows calls to be chained in a fluent API style, similar to the Page Object Pattern. Internally the implementation wil…