hadoop - Any python library for parallel and distributed tasks? -
i looking python library can distribute tasks across few servers. task similar can parallelized subprocess
library in single machine.
i know can setup hadoop system such purposes. hadoop heavy weight. in case, use shared network disk data i/o, , don't need fancy failure recover. in mapreduce's terminology, need mappers, no aggregators or reducers.
any such library in python? thanks!
try using celery.
celery asynchronous task queue/job queue based on distributed message passing. focused on real-time operation, supports scheduling well.
the execution units, called tasks, executed concurrently on single or more worker servers using multiprocessing, eventlet, or gevent. tasks can execute asynchronously (in background) or synchronously (wait until ready).
Comments
Post a Comment