I'm designing a bloom filter and I'm wondering what the most performant bit array implementation is in Python.
The nice thing about Python is that it can handle arbitrary length integers out of the box and that's what I use now, but I don't know enough about Python internals to know if that's the most performant way to do it in Python.
I found bitarray
but it handles a lot of other things like slicing, which I don't need. I only need the &
and |
and <<
operations.
The built-in
int
is pretty nicely optimized, and it already supports&
,|
, and<<
.There's at least one alternative implementation of arbitrary-length integers, based on GMP, known as
gmpy2
. (There's also the originalgmpy
,PyGMP
,Sophie
, and a few other wrappers around the same library, but I doubt they'd have any real performance differences.)And there are two major implementations of the "bit array" idea,
bitarray
(the one you linked) andbitstring
, as well as a few libraries likeintbitset
that give you a set-like interface instead (which should also work for your uses).So, let's toss them all together and compare:
On my Mac, running Python.org 64-bit CPython 3.3.2, here's what I get:
That seems to be enough to summarily dismiss
bitstring
.Also, while I didn't show
intbitset
here, because neither it nor any similar libraries I found work with Python 3, from a variety of comparisons (usingintbitset.intbitset([i for i, bit in enumerate(bin(n)[2:]) if bit != '0'])
) it's anywhere from 14 to 70 times slower thanint
in every test I throw at it, so I've summarily dismissed it as well.So let's try the other three with more reps:
And the numbers hold up.
bitarray
is nowhere near as fast as built-inint
. It's also more clumsy to use (note that what should be an "alternate constructor" classmethod is an instance method, there's no fast and easy way to convert from or to an int, and the reason I was only testingx | x
andx & x
is that there are limitations on the operators). If you need an integer as an array of bits, it's great; if you need to do C-style bit-munging on an integer, that's not what it's best at.As for
gmpy2
, it seems to be about a third faster than the nativeint
. What if we make the numbers a whole lot bigger, like 1.5kbits?And here are the numbers with Apple Python 2.7.5:
So, is it worth it? It's less friendly to use, it's slower rather than faster on some other operations that you didn't ask about, it requires a third-party C library which is LGPL-licensed, it has much worse error-handling behavior in memory overflow cases (as in, your app may segfault instead of raising), and there's at least one person on StackOverflow who is going to show up and tell you that GMP sucks whenever it gets mentioned (I believe because of a bug in an older version).
But if you need that extra speed, maybe it's worth it.
On the other hand, here's PyPy, 3.2.3/2.1.0b1 and PyPy 2.7.3/2.1.0, using the native
int
type—compare with the 0.19562570203561336 and 0.2890629768371582 results from gmpy2 above:So, switching from CPython to PyPy gives you almost as much benefit as switching from
int
togmpy2.mpz
in Python 3, and significantly more benefit in Python 2. And it will almost certainly accelerate the rest of your code as well.