I’ve been thinking about caching lately because our application’s performance started degrading as we scaled. We were hitting the database for every request, and response times were suffering. That’s when I realized we needed a more sophisticated approach to caching - something beyond the basic key-value storage most developers implement.
Have you ever noticed how some applications feel lightning-fast while others struggle with every click? The difference often comes down to intelligent caching strategies.
Let me show you how we transformed our application’s performance using Redis with Node.js. We started with a simple memory cache but quickly realized we needed something more robust for our distributed architecture.
// Basic memory cache implementation
class SimpleCache {
constructor() {
this.cache = new Map();
this.ttl = new Map();
}
set(key, value, ttlMs = 300000) {
this.cache.set(key, value);
this.ttl.set(key, Date.now() + ttlMs);
}
get(key) {
const expiry = this.ttl.get(key);
if (expiry && Date.now() > expiry) {
this.cache.delete(key);
this.ttl.delete(key);
return null;
}
return this.cache.get(key);
}
}
But what happens when your application scales across multiple servers? Memory caching alone isn’t enough. That’s where Redis comes in - providing a shared cache layer that all your application instances can access.
The cache-aside pattern became our foundation. When data is requested, we first check Redis. If it’s not there, we fetch from the database and populate the cache. This simple pattern reduced our database load by 80%.
// Cache-aside implementation
async function getUserWithCache(userId) {
const cacheKey = `user:${userId}`;
let user = await redis.get(cacheKey);
if (!user) {
user = await database.users.findById(userId);
if (user) {
await redis.setex(cacheKey, 300, JSON.stringify(user));
}
} else {
user = JSON.parse(user);
}
return user;
}
Cache invalidation is where things get interesting. How do you ensure cached data stays fresh when the underlying data changes? We implemented a tagging system that tracks relationships between cached items.
// Cache invalidation with tags
async function updateProduct(productId, updates) {
// Update in database
await database.products.update(productId, updates);
// Invalidate related cache entries
const tags = [`product:${productId}`, 'products:list'];
await redis.del(...tags);
// Also invalidate user carts containing this product
const cartKeys = await redis.keys('cart:*');
for (const key of cartKeys) {
const cart = await redis.get(key);
if (cart && JSON.parse(cart).items.some(item => item.productId === productId)) {
await redis.del(key);
}
}
}
When your cache becomes critical to performance, you need to plan for failure. We implemented circuit breakers to prevent cache failures from taking down the entire application.
// Circuit breaker for cache operations
class CacheCircuitBreaker {
constructor(failureThreshold = 5, resetTimeout = 60000) {
this.failureThreshold = failureThreshold;
this.resetTimeout = resetTimeout;
this.failureCount = 0;
this.lastFailureTime = null;
this.state = 'CLOSED';
}
async execute(operation) {
if (this.state === 'OPEN') {
if (Date.now() - this.lastFailureTime > this.resetTimeout) {
this.state = 'HALF_OPEN';
} else {
throw new Error('Cache circuit breaker is OPEN');
}
}
try {
const result = await operation();
if (this.state === 'HALF_OPEN') {
this.state = 'CLOSED';
this.failureCount = 0;
}
return result;
} catch (error) {
this.failureCount++;
this.lastFailureTime = Date.now();
if (this.failureCount >= this.failureThreshold) {
this.state = 'OPEN';
}
throw error;
}
}
}
Monitoring cache performance is crucial. We track hit rates, response times, and memory usage to identify bottlenecks before they become problems.
Did you know that proper pipelining can reduce Redis latency by up to 70%? Instead of waiting for each command to complete before sending the next, we batch operations together.
// Redis pipelining for performance
async function cacheMultipleUsers(userIds) {
const pipeline = redis.pipeline();
userIds.forEach(userId => {
pipeline.get(`user:${userId}`);
});
const results = await pipeline.exec();
return results.map(([error, result]) =>
error ? null : JSON.parse(result)
);
}
For truly massive scale, Redis clustering distributes data across multiple nodes. This provides both horizontal scaling and high availability. We configured our cluster with automatic failover, so if one node goes down, others take over seamlessly.
But what about write-heavy applications? That’s where write-through caching shines. We write to both cache and database simultaneously, ensuring data consistency while maintaining performance.
The journey from basic memory caching to distributed patterns transformed our application. Response times dropped from seconds to milliseconds, and our infrastructure became resilient to traffic spikes.
What caching challenges are you facing in your projects? Have you considered how different cache patterns might solve your performance issues?
I’d love to hear about your experiences with caching strategies. If this approach helped you improve your application’s performance, please share your results in the comments below. Don’t forget to like and share this with your team - better caching could be the performance boost your application needs.