Using multiprocessing with a decorated function results in a PicklingError

This is happening because g is actually defined as a nested function in MultiProcFunctions, which means it's not actually importable from the top-level of that module, which means it won't pickle properly. Now, we're actually pretty clearly defining g in the top-level of __main__ module though, when we do this:

g = MultiProcDecorator(sq)

So, it really should be picklable. We can make it work by explicitly setting the __module__ of g to be "__main__":

g = MultiProcDecorator(sq)
g.__module__ = "__main__"  # Fix the __module__

This will allow the pickling process to work, since it will look for the definition of g in __main__, where it is defined at the top-level, rather than MultiProcFunctions, where it is only defined in a nested scope.

Edit:

Note that you could also make the change in the decorator itself:

def MultiProcDecorator(f,*args):

    """
    Takes a function f, and formats it so that results are saved to a shared dict
    """

    def g(procnum,return_dict,*args):
        result = f(*args)
        return_dict[procnum] = result
    g.__module__ = "__main__"

    return g

This probably makes more sense for you, since this decorator is strictly meant to be using for multiprocessing purposes.


Trying dano's trick seem to only works in Python 2. When trying in Python 3, I get the following error:

pickle.PicklingError: Can't pickle <function serialize at 0x7f7a1ac1fd08>: it's not the same object as __main__.orig_fn

I solved this issue by "decorating" function from worker's init:

from functools import wraps
import sys

def worker_init(fn, *args):
    @wraps(fn)
    def wrapper(x):
        # wrapper logic
        pass

    setattr(sys.modules[fn.__module__], fn.__name__, wrapper)

pool = mp.Pool(initializer=worker_init, initargs=[orig_fn, *args])
# ...