If you're running a high-traffic Node.js application, you've probably hit a bottleneck where a single process just isnβt enough. Since Node.js runs on a single thread, how do we take advantage of multi-core CPUs?
The answer: Clustering. π
π₯ What is the Cluster Module?
The Cluster module in Node.js allows you to spawn multiple child processes (workers), each handling requests independently but sharing the same port.
Think of it as hiring multiple chefs in a kitchen instead of relying on just one. More workers = better performance! πͺ
π Basic Cluster Setup
Letβs set up a clustered Node.js server:
import cluster from "cluster";
import http from "http";
import os from "os";
const numCPUs = os.cpus().length;
if (cluster.isPrimary) {
console.log(`π Master process ${process.pid} is running`);
// Fork workers
for (let i = 0; i < numCPUs; i++) {
cluster.fork();
}
// Restart workers if they die
cluster.on("exit", (worker) => {
console.log(`π Worker ${worker.process.pid} died. Spawning a new one...`);
cluster.fork();
});
} else {
// Worker processes create HTTP server
http
.createServer((req, res) => {
res.writeHead(200);
res.end(`Handled by Worker ${process.pid}\n`);
})
.listen(3000);
console.log(`β‘ Worker ${process.pid} started`);
}
π Whatβs Happening?
- The primary process (
cluster.isPrimary
) spawns workers = one per CPU core. - Each worker process creates an HTTP server.
- If a worker crashes, the master replaces it.
- All workers share the same port (3000).
β‘ Testing the Cluster
Run the server:
node cluster-server.mjs
Now open multiple terminal tabs and run:
curl http://localhost:3000
Each request gets handled by a different worker, proving weβre using multiple processes! π
π Worker Communication
Workers donβt share memory, but they can send messages to each other and the master.
Example: Worker sends a message to the master
if (cluster.isPrimary) {
cluster.on("message", (worker, message) => {
console.log(`π© Message from Worker ${worker.process.pid}:`, message);
});
cluster.fork(); // Start a worker
} else {
process.send({ status: "Hello from Worker!" });
}
Each worker can communicate with the master process, allowing coordination and data exchange.
π― Real-World Use Case: Scaling an API Server
Imagine you're running a REST API that serves thousands of requests per second.
Without clustering:
- The Node.js process handles all requests alone, causing slow responses under high traffic.
- Requests queue up, increasing latency.
- CPU utilization is low because it only runs on a single core.
With clustering:
- The API server spawns multiple workers, distributing the load.
- More requests are handled in parallel, reducing response time.
- CPU utilization increases, making the most of available resources.
How to Apply Clustering to an Express.js API
Letβs modify our setup to handle API requests using Express.js.
import cluster from "cluster";
import os from "os";
import express from "express";
const numCPUs = os.cpus().length;
if (cluster.isPrimary) {
console.log(`π Master ${process.pid} is running`);
for (let i = 0; i < numCPUs; i++) {
cluster.fork();
}
cluster.on("exit", (worker) => {
console.log(`π Worker ${worker.process.pid} died. Restarting...`);
cluster.fork();
});
} else {
const app = express();
app.get("/", (req, res) => {
res.send(`Handled by Worker ${process.pid}`);
});
app.listen(3000, () => {
console.log(`β‘ Worker ${process.pid} is listening on port 3000`);
});
}
Benefits
β
Handles more concurrent requests β Each worker runs on a separate core.
β
Prevents downtime β If a worker crashes, another takes over.
β
Improves performance β Spreads workload across CPU cores.
π― Why Use Clustering?
- Maximizes CPU Utilization π₯ β Run multiple Node.js instances on all cores.
- Handles More Requests π β Distributes load across multiple processes.
- Prevents Crashes π β If one worker dies, others keep running.
- Great for Scaling π β Essential for high-traffic applications.
β οΈ Limitations of Clustering
- Workers donβt share memory β Use databases (e.g., Redis) for shared state.
- Not ideal for CPU-heavy tasks β Consider worker threads instead.
- Limited scaling β For multiple servers, use load balancers (NGINX, PM2, Kubernetes).
π₯ Final Thoughts
The Cluster module is a simple and powerful way to scale Node.js applications. Itβs perfect for handling high traffic and ensuring better fault tolerance.