DRMacIver's Notebook
Auto-parallelising test-case reduction
Auto-parallelising test-case reduction
There’s a new parallel test-case reducer called half-empty.
It uses what is a new to me approach to parallelisation but is apparently also what C-Reduce does, which they call pessimistic parallelism. The basic idea is that you parallelise based on the assumption that what you’re going to try isn’t actually going to work, you fork a background process to check if it does actually work, and the main calculation just proceeds as if it was false. If your assumption that it rarely works mostly holds true, this lets you turn what seem like highly sequential processes into highly parallel ones.
It occurred to me that you could fairly easily automate this in a way that lets you write a test-case reducer exactly as if it were sequential but have it magically automatically parallelised in the background.
The way it would work is this: You have some test case reducer state object which has a cached version of the predicate. You wrap a reduction pass in some special function (say a decorator in Python). Now when you call the predicate from within the reduction pass what happens is as follows:
- If the result has already been cached, use that.
- If the result hasn’t been cached, return false and queue the result for background computation, which will update the cache when it is finished.
- If at any point a backgrounded job returns true instead of false, clear the queue, wait for the current computations to finish, and then restart the reduction pass from the beginning.
This makes a couple of assumptions:
- Running the full reduction pass is cheaper than running the predicate.
- The predicate will rarely return true.
Ways it might be useful to patch this:
- Keep the queue size bounded. When the test function calls the predicate and the queue is at capacity, have it block until the queue is emptier.
- If the pessimistic assumption does not hold, e.g. say if at least 5 of the last 10 predicate calls were true, run the predicate in the foreground instead of backgrounding it.
If you want a really cheeky approach (I don’t think this will work), here’s a neat trick you could try: Use some sort of classifier (language inference, machine learning, whatever) to predict the result of the predicate and use that, invalidating when the parallel computation gets it wrong.. You could even just be very lazy and just predict whichever outcome is the most common.
Actually speaking of language learning, this approach would also work well for L* with a bit more tracking of dependencies, which would give you a fully parallel language learner.