| | |
| | |
| | |
| | |
| |
|
| | #include <torch/extension.h> |
| |
|
| | |
| | void rope_2d_cuda( torch::Tensor tokens, const torch::Tensor pos, const float base, const float fwd ); |
| |
|
| | void rope_2d_cpu( torch::Tensor tokens, const torch::Tensor positions, const float base, const float fwd ) |
| | { |
| | const int B = tokens.size(0); |
| | const int N = tokens.size(1); |
| | const int H = tokens.size(2); |
| | const int D = tokens.size(3) / 4; |
| |
|
| | auto tok = tokens.accessor<float, 4>(); |
| | auto pos = positions.accessor<int64_t, 3>(); |
| |
|
| | for (int b = 0; b < B; b++) { |
| | for (int x = 0; x < 2; x++) { |
| | for (int n = 0; n < N; n++) { |
| | |
| | |
| | const int p = pos[b][n][x]; |
| |
|
| | for (int h = 0; h < H; h++) { |
| | for (int d = 0; d < D; d++) { |
| | |
| | float u = tok[b][n][h][d+0+x*2*D]; |
| | float v = tok[b][n][h][d+D+x*2*D]; |
| |
|
| | |
| | const float inv_freq = fwd * p / powf(base, d/float(D)); |
| | float c = cosf(inv_freq); |
| | float s = sinf(inv_freq); |
| |
|
| | |
| | tok[b][n][h][d+0+x*2*D] = u*c - v*s; |
| | tok[b][n][h][d+D+x*2*D] = v*c + u*s; |
| | } |
| | } |
| | } |
| | } |
| | } |
| | } |
| |
|
| | void rope_2d( torch::Tensor tokens, |
| | const torch::Tensor positions, |
| | const float base, |
| | const float fwd ) |
| | { |
| | TORCH_CHECK(tokens.dim() == 4, "tokens must have 4 dimensions"); |
| | TORCH_CHECK(positions.dim() == 3, "positions must have 3 dimensions"); |
| | TORCH_CHECK(tokens.size(0) == positions.size(0), "batch size differs between tokens & positions"); |
| | TORCH_CHECK(tokens.size(1) == positions.size(1), "seq_length differs between tokens & positions"); |
| | TORCH_CHECK(positions.size(2) == 2, "positions.shape[2] must be equal to 2"); |
| | TORCH_CHECK(tokens.is_cuda() == positions.is_cuda(), "tokens and positions are not on the same device" ); |
| |
|
| | if (tokens.is_cuda()) |
| | rope_2d_cuda( tokens, positions, base, fwd ); |
| | else |
| | rope_2d_cpu( tokens, positions, base, fwd ); |
| | } |
| |
|
| | PYBIND11_MODULE(TORCH_EXTENSION_NAME, m) { |
| | m.def("rope_2d", &rope_2d, "RoPE 2d forward/backward"); |
| | } |
| |
|