| std::vector<at::Tensor> index_max( | |
| at::Tensor index_vals, | |
| at::Tensor indices, | |
| int A_num_block, | |
| int B_num_block | |
| ) { | |
| return index_max_kernel( | |
| index_vals, | |
| indices, | |
| A_num_block, | |
| B_num_block | |
| ); | |
| } | |
| at::Tensor mm_to_sparse( | |
| at::Tensor dense_A, | |
| at::Tensor dense_B, | |
| at::Tensor indices | |
| ) { | |
| return mm_to_sparse_kernel( | |
| dense_A, | |
| dense_B, | |
| indices | |
| ); | |
| } | |
| at::Tensor sparse_dense_mm( | |
| at::Tensor sparse_A, | |
| at::Tensor indices, | |
| at::Tensor dense_B, | |
| int A_num_block | |
| ) { | |
| return sparse_dense_mm_kernel( | |
| sparse_A, | |
| indices, | |
| dense_B, | |
| A_num_block | |
| ); | |
| } | |
| at::Tensor reduce_sum( | |
| at::Tensor sparse_A, | |
| at::Tensor indices, | |
| int A_num_block, | |
| int B_num_block | |
| ) { | |
| return reduce_sum_kernel( | |
| sparse_A, | |
| indices, | |
| A_num_block, | |
| B_num_block | |
| ); | |
| } | |
| at::Tensor scatter( | |
| at::Tensor dense_A, | |
| at::Tensor indices, | |
| int B_num_block | |
| ) { | |
| return scatter_kernel( | |
| dense_A, | |
| indices, | |
| B_num_block | |
| ); | |
| } | |
| PYBIND11_MODULE(TORCH_EXTENSION_NAME, m) { | |
| m.def("index_max", &index_max, "index_max (CUDA)"); | |
| m.def("mm_to_sparse", &mm_to_sparse, "mm_to_sparse (CUDA)"); | |
| m.def("sparse_dense_mm", &sparse_dense_mm, "sparse_dense_mm (CUDA)"); | |
| m.def("reduce_sum", &reduce_sum, "reduce_sum (CUDA)"); | |
| m.def("scatter", &scatter, "scatter (CUDA)"); | |
| } | |