Quick and Dirty Node.js Process (Job) Queue
I recently began experimenting with nodejs for a small web scraping project. I wrote a tiny script that goes out to lots of URLs and downloads files to disk. The simple solution was to iterate through the list and send a request to load the URL and download the page.
Too Many Open Files
Unfortunately, there are limits on the amount of simultaneous exec()
calls you can make. Since running an external command via exec()
is non-blocking, making too many back-to-back calls of it will result in the following:
node.js:201
throw e; // process.nextTick error, or error event on first tick
^
Error: spawn EMFILE
at errnoException (child_process.js:481:11)
at ChildProcess.spawn (child_process.js:444:11)
at child_process.js:342:9
at Object.execFile (child_process.js:252:15)
at child_process.js:220:18
Maximum Simultaneous Calls
To solve this, I implemented something like the following:
var queue = [];
var MAX = 20; // only allow 20 simultaneous exec calls
var count = 0; // holds how many execs are running
var urls = [...] // long list of urls
// our callback for each exec call
function wget_callback(err, stdout, stderr) {
count -= 1;
if (queue.length > 0 && count < MAX) { // get next item in the queue!
count += 1;
var url = queue.shift();
exec('wget '+url, wget_callback);
}
}
urls.forEach( function(url) {
if (count < MAX) { // go get the file!
count += 1;
exec('wget '+url, wget_callback);
} else { // queue it up...
queue.push(url);
}
});
This will only allow so many exec() calls to simultaneously run. The rest of the URLs will be stored in a queue until a slot becomes available for them. Checking (and shifting) the queue is done in the callback function wget_callback()
. I fetch the next URL to download out of the queue
only if there are no more than MAX
exec()
calls already running. I keep track of how many calls are currently running using count
and increment/decrement accordingly.
I’m sure there are tons of libraries that do this, but I decided to implement a quick and dirty solution to this problem and thought I’d share!