This implements sparse multidimensional arrays on top of NumPy and Scipy.sparse. It generalizes the scipy.sparse.coo_matrix layout but extends beyond just rows and columns to an arbitrary number of dimensions.
The original motivation is for machine learning algorithms, but it is intended for somewhat general use.
- Reshape
- Transpose
- Tensordot
- Reductions (sum, max)
- Slicing with integers, lists, and slices (with no step value)
A "does not support" list is hard to build because it is infinitely long. However the following things are in scope, relatively doable, and not yet built (help welcome).
- Arithmetic like add
- NumPy ufunc support (where zero is maintained)
- Concatenation and stacking
- Smooth interaction with numpy arrays, scipy.sparse arrays, etc. for binary operations
- Incremental buliding of arrays and inplace updates
- Parallel computing
In [1]: import numpy as np
In [2]: n = 1000
In [3]: ndims = 4
In [4]: nnz = 1000000
In [5]: coords = np.random.randint(0, n - 1, size=(ndims, nnz))
In [6]: data = np.random.random(nnz)
In [7]: import sparse
In [8]: x = sparse.COO(coords, data, shape=((n,) * ndims))
In [9]: x
Out[9]: <COO: shape=(1000, 1000, 1000, 1000), dtype=float64, nnz=1000000>
In [10]: x.nbytes
Out[10]: 40000000
In [11]: %time y = sparse.tensordot(x, x, axes=((3, 0), (1, 2)))
CPU times: user 1.52 s, sys: 20 ms, total: 1.54 s
Wall time: 1.54 s
In [12]: y
Out[12]: <COO: shape=(1000, 1000, 1000, 1000), dtype=float64, nnz=1001588>
In [13]: %time z = y.sum(axis=(0, 1, 2))
CPU times: user 408 ms, sys: 408 ms, total: 816 ms
Wall time: 818 ms
In [14]: z
Out[14]: <COO: shape=(1000,), dtype=float64, nnz=999>
In [15]: z.todense()
Out[15]:
array([ 244.0671803 , 246.38455787, 243.43383158, 256.46068737,
261.18598416, 256.36439011, 271.74177584, 238.56059193,
...
Scipy.sparse implements decent 2-d sparse matrix objects for the standard layouts, notably for our purposes CSR, CSC, and COO <https://en.wikipedia.org/wiki/Sparse_matrix>. However it doesn't include support for sparse arrays of greater than 2 dimensions.
This library extends the COO layout, which stores the row index, column index, and value of every element:
row | col | data |
---|---|---|
0 | 0 | 10 |
0 | 2 | 13 |
1 | 3 | 9 |
3 | 8 | 21 |
It is straightforward to extend the COO layout to an arbitrary number of dimensions:
dim1 | dim2 | dim3 | ... | data |
---|---|---|---|---|
0 | 0 | 0 | . | 10 |
0 | 0 | 3 | . | 13 |
0 | 2 | 2 | . | 9 |
3 | 1 | 4 | . | 21 |
This makes it easy to store a multidimensional sparse array, but we still need to reimplement all of the array operations like transpose, reshape, slicing, tensordot, reductions, etc., which can be quite challenging in general.
Fortunately in many cases we can leverage the existing SciPy.sparse algorithms if we can intelligently transpose and reshape our multi-dimensional array into an appropriate 2-d sparse matrix, perform a modified sparse matrix operation, and then reshape and transpose back. These reshape and transpose operations can all be done at numpy speeds by modifying the arrays of coordinates. After scipy.sparse runs its operations (coded in C) then we can convert back to using the same path of reshapings and transpositions in reverse.
This approach is not novel; it has been around in the multidimensional array
community for a while. It is also how some operations in numpy work. For example
the numpy.tensordot
function performs transposes and reshapes so that it can
use the numpy.dot
function for matrix multiplication which is backed by
fast BLAS implementations. The sparse.tensordot
code is very slight
modification of numpy.tensordot
, replacing numpy.dot
with
scipy.sprarse.csr_matrix.dot
.
This is licensed under New BSD-3