@@ -747,12 +747,16 @@ cpdef sparse_nanpow(ndarray x, SparseIndex xindex,
747
747
cdef inline tuple sparse_nancombine(ndarray x, SparseIndex xindex,
748
748
ndarray y, SparseIndex yindex,
749
749
double_func op):
750
- if isinstance (xindex, BlockIndex):
751
- return block_nanop(x, xindex.to_block_index(),
752
- y, yindex.to_block_index(), op)
753
- elif isinstance (xindex, IntIndex):
754
- return int_nanop(x, xindex.to_int_index(),
755
- y, yindex.to_int_index(), op)
750
+ # block_nanop is up to 40x slower and I don't know why yet
751
+ return int_nanop(x, xindex.to_int_index(),
752
+ y, yindex.to_int_index(), op)
753
+
754
+ # if isinstance(xindex, BlockIndex):
755
+ # return block_nanop(x, xindex.to_block_index(),
756
+ # y, yindex.to_block_index(), op)
757
+ # elif isinstance(xindex, IntIndex):
758
+ # return int_nanop(x, xindex.to_int_index(),
759
+ # y, yindex.to_int_index(), op)
756
760
757
761
cpdef sparse_add(ndarray x, SparseIndex xindex, float64_t xfill,
758
762
ndarray y, SparseIndex yindex, float64_t yfill):
0 commit comments