""" This module extends standard Python's pickle module so that it is capable of writing more efficient pickle files that contain Panda objects with shared pointers. In particular, a single Python structure that contains many NodePaths into the same scene graph will write the NodePaths correctly when used with this pickle module, so that when it is unpickled later, the NodePaths will still reference into the same scene graph together. If you use the standard pickle module instead, the NodePaths will each duplicate its own copy of its scene graph. This is necessary because the standard pickle module doesn't provide a mechanism for sharing context between different objects written to the same pickle stream, so each NodePath has to write itself without knowing about the other NodePaths that will also be writing to the same stream. This replacement module solves this problem by defining a __reduce_persist__() replacement method for __reduce__(), which accepts a pointer to the Pickler object itself, allowing for shared context between all objects written by that Pickler. Unfortunately, cPickle cannot be supported, because it does not support extensions of this nature. """ from types import * from copy_reg import dispatch_table from pandac.PandaModules import BamWriter, BamReader # A funny replacement for "import pickle" so we don't get confused # with the local pickle.py. pickle = __import__('pickle') class Pickler(pickle.Pickler): def __init__(self, *args, **kw): self.bamWriter = BamWriter() pickle.Pickler.__init__(self, *args, **kw) # We have to duplicate most of the save() method, so we can add # support for __reduce_persist__(). def save(self, obj): # Check for persistent id (defined by a subclass) pid = self.persistent_id(obj) if pid: self.save_pers(pid) return # Check the memo x = self.memo.get(id(obj)) if x: self.write(self.get(x[0])) return # Check the type dispatch table t = type(obj) f = self.dispatch.get(t) if f: f(self, obj) # Call unbound method with explicit self return # Check for a class with a custom metaclass; treat as regular class try: issc = issubclass(t, TypeType) except TypeError: # t is not a class (old Boost; see SF #502085) issc = 0 if issc: self.save_global(obj) return # Check copy_reg.dispatch_table reduce = dispatch_table.get(t) if reduce: rv = reduce(obj) else: # New code: check for a __reduce_persist__ method, then # fall back to standard methods. reduce = getattr(obj, "__reduce_persist__", None) if reduce: rv = reduce(self) else: # Check for a __reduce_ex__ method, fall back to __reduce__ reduce = getattr(obj, "__reduce_ex__", None) if reduce: rv = reduce(self.proto) else: reduce = getattr(obj, "__reduce__", None) if reduce: rv = reduce() else: raise PicklingError("Can't pickle %r object: %r" % (t.__name__, obj)) # Check for string returned by reduce(), meaning "save as global" if type(rv) is StringType: self.save_global(obj, rv) return # Assert that reduce() returned a tuple if type(rv) is not TupleType: raise PicklingError("%s must return string or tuple" % reduce) # Assert that it returned an appropriately sized tuple l = len(rv) if not (2 <= l <= 5): raise PicklingError("Tuple returned by %s must have " "two to five elements" % reduce) # Save the reduce() output and finally memoize the object self.save_reduce(obj=obj, *rv) class Unpickler(pickle.Unpickler): def __init__(self, *args, **kw): self.bamReader = BamReader() pickle.Unpickler.__init__(self, *args, **kw) # Duplicate the load_reduce() function, to provide a special case # for the reduction function. def load_reduce(self): stack = self.stack args = stack.pop() func = stack[-1] # If the function name ends with "Persist", then assume the # function wants the Unpickler as the first parameter. if func.__name__.endswith('Persist'): value = func(self, *args) else: # Otherwise, use the existing pickle convention. value = func(*args) stack[-1] = value pickle.Unpickler.dispatch[pickle.REDUCE] = load_reduce # Shorthands try: from cStringIO import StringIO except ImportError: from StringIO import StringIO def dump(obj, file, protocol=None): Pickler(file, protocol).dump(obj) def dumps(obj, protocol=None): file = StringIO() Pickler(file, protocol).dump(obj) return file.getvalue() def load(file): return Unpickler(file).load() def loads(str): file = StringIO(str) return Unpickler(file).load()