What is the correct (or best) way to subclass the Python set class, adding a new instance variable?
My favorite way to wrap methods of a built-in collection:
class Fooset(set):
def __init__(self, s=(), foo=None):
super(Fooset,self).__init__(s)
if foo is None and hasattr(s, 'foo'):
foo = s.foo
self.foo = foo
@classmethod
def _wrap_methods(cls, names):
def wrap_method_closure(name):
def inner(self, *args):
result = getattr(super(cls, self), name)(*args)
if isinstance(result, set) and not hasattr(result, 'foo'):
result = cls(result, foo=self.foo)
return result
inner.fn_name = name
setattr(cls, name, inner)
for name in names:
wrap_method_closure(name)
Fooset._wrap_methods(['__ror__', 'difference_update', '__isub__',
'symmetric_difference', '__rsub__', '__and__', '__rand__', 'intersection',
'difference', '__iand__', 'union', '__ixor__',
'symmetric_difference_update', '__or__', 'copy', '__rxor__',
'intersection_update', '__xor__', '__ior__', '__sub__',
])
Essentially the same thing you're doing in your own answer, but with fewer loc. It's also easy to put in a metaclass if you want to do the same thing with lists and dicts as well.
I think that the recommended way to do this is not to subclass directly from the built-in set
, but rather to make use of the Abstract Base Class Set
available in collections.abc.
Using the ABC Set gives you some methods for free as a mix-in so you can have a minimal Set class by defining only __contains__()
, __len__()
and __iter__()
. If you want some of the nicer set methods like intersection()
and difference()
, you probably do have to wrap them.
Here's my attempt (this one happens to be a frozenset-like, but you can inherit from MutableSet
to get a mutable version):
from collections.abc import Set, Hashable
class CustomSet(Set, Hashable):
"""An example of a custom frozenset-like object using
Abstract Base Classes.
"""
__hash__ = Set._hash
wrapped_methods = ('difference',
'intersection',
'symetric_difference',
'union',
'copy')
def __repr__(self):
return "CustomSet({0})".format(list(self._set))
def __new__(cls, iterable=None):
selfobj = super(CustomSet, cls).__new__(CustomSet)
selfobj._set = frozenset() if iterable is None else frozenset(iterable)
for method_name in cls.wrapped_methods:
setattr(selfobj, method_name, cls._wrap_method(method_name, selfobj))
return selfobj
@classmethod
def _wrap_method(cls, method_name, obj):
def method(*args, **kwargs):
result = getattr(obj._set, method_name)(*args, **kwargs)
return CustomSet(result)
return method
def __getattr__(self, attr):
"""Make sure that we get things like issuperset() that aren't provided
by the mix-in, but don't need to return a new set."""
return getattr(self._set, attr)
def __contains__(self, item):
return item in self._set
def __len__(self):
return len(self._set)
def __iter__(self):
return iter(self._set)
Sadly, set does not follow the rules and __new__
is not called to make new set
objects, even though they keep the type. This is clearly a bug in Python (issue #1721812, which will not be fixed in the 2.x sequence). You should never be able to get an object of type X without calling the type
object that creates X objects! If set.__or__
is not going to call __new__
it is formally obligated to return set
objects instead of subclass objects.
But actually, noting the post by nosklo above, your original behavior does not make any sense. The Set.__or__
operator should not be reusing either of the source objects to construct its result, it should be whipping up a new one, in which case its foo
should be "default"
!
So, practically, anyone doing this should have to overload those operators so that they would know which copy of foo
gets used. If it is not dependent on the Foosets being combined, you can make it a class default, in which case it will get honored, because the new object thinks it is of the subclass type.
What I mean is, your example would work, sort of, if you did this:
class Fooset(set):
foo = 'default'
def __init__(self, s = []):
if isinstance(s, Fooset):
self.foo = s.foo
f = Fooset([1,2,5])
assert (f|f).foo == 'default'