hadoop - Any python library for parallel and distributed tasks? -


i looking python library can distribute tasks across few servers. task similar can parallelized subprocess library in single machine.

i know can setup hadoop system such purposes. hadoop heavy weight. in case, use shared network disk data i/o, , don't need fancy failure recover. in mapreduce's terminology, need mappers, no aggregators or reducers.

any such library in python? thanks!

try using celery.

celery asynchronous task queue/job queue based on distributed message passing. focused on real-time operation, supports scheduling well.

the execution units, called tasks, executed concurrently on single or more worker servers using multiprocessing, eventlet, or gevent. tasks can execute asynchronously (in background) or synchronously (wait until ready).


Comments

Popular posts from this blog

html5 - What is breaking my page when printing? -

html - Unable to style the color of bullets in a list -

c# - must be a non-abstract type with a public parameterless constructor in redis -