ZTWHHH commited on
Commit
5072694
·
verified ·
1 Parent(s): ccf2005

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +10 -0
  2. llava/lib/python3.10/__pycache__/_markupbase.cpython-310.pyc +0 -0
  3. llava/lib/python3.10/__pycache__/_strptime.cpython-310.pyc +0 -0
  4. llava/lib/python3.10/__pycache__/abc.cpython-310.pyc +0 -0
  5. llava/lib/python3.10/__pycache__/asyncore.cpython-310.pyc +0 -0
  6. llava/lib/python3.10/__pycache__/cgitb.cpython-310.pyc +0 -0
  7. llava/lib/python3.10/__pycache__/configparser.cpython-310.pyc +0 -0
  8. llava/lib/python3.10/__pycache__/glob.cpython-310.pyc +0 -0
  9. llava/lib/python3.10/__pycache__/gzip.cpython-310.pyc +0 -0
  10. llava/lib/python3.10/__pycache__/imaplib.cpython-310.pyc +0 -0
  11. llava/lib/python3.10/__pycache__/pydoc.cpython-310.pyc +0 -0
  12. llava/lib/python3.10/__pycache__/rlcompleter.cpython-310.pyc +0 -0
  13. llava/lib/python3.10/__pycache__/sre_constants.cpython-310.pyc +0 -0
  14. llava/lib/python3.10/__pycache__/traceback.cpython-310.pyc +0 -0
  15. llava/lib/python3.10/__pycache__/types.cpython-310.pyc +0 -0
  16. llava/lib/python3.10/__pycache__/wave.cpython-310.pyc +0 -0
  17. llava/lib/python3.10/__pycache__/zipfile.cpython-310.pyc +0 -0
  18. llava/lib/python3.10/lib-dynload/_codecs_hk.cpython-310-x86_64-linux-gnu.so +3 -0
  19. llava/lib/python3.10/lib-dynload/_codecs_tw.cpython-310-x86_64-linux-gnu.so +3 -0
  20. llava/lib/python3.10/lib-dynload/_ctypes.cpython-310-x86_64-linux-gnu.so +3 -0
  21. llava/lib/python3.10/lib-dynload/_lzma.cpython-310-x86_64-linux-gnu.so +3 -0
  22. llava/lib/python3.10/lib-dynload/_socket.cpython-310-x86_64-linux-gnu.so +3 -0
  23. llava/lib/python3.10/lib-dynload/_ssl.cpython-310-x86_64-linux-gnu.so +3 -0
  24. llava/lib/python3.10/lib-dynload/_xxsubinterpreters.cpython-310-x86_64-linux-gnu.so +3 -0
  25. llava/lib/python3.10/lib-dynload/_zoneinfo.cpython-310-x86_64-linux-gnu.so +3 -0
  26. llava/lib/python3.10/lib-dynload/readline.cpython-310-x86_64-linux-gnu.so +3 -0
  27. llava/lib/python3.10/test/__pycache__/__init__.cpython-310.pyc +0 -0
  28. llava/lib/python3.10/test/__pycache__/test_script_helper.cpython-310.pyc +0 -0
  29. llava/lib/python3.10/test/__pycache__/test_support.cpython-310.pyc +0 -0
  30. llava/lib/python3.10/test/support/__pycache__/__init__.cpython-310.pyc +0 -0
  31. llava_next/lib/python3.10/site-packages/triton/third_party/cuda/bin/ptxas +3 -0
  32. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_batch_norm_impl_index_backward_compositeimplicitautograd_dispatch.h +23 -0
  33. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_cdist_forward_cpu_dispatch.h +23 -0
  34. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_foreach_acos_native.h +25 -0
  35. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_foreach_sinh_cuda_dispatch.h +24 -0
  36. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_foreach_trunc.h +44 -0
  37. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_linalg_slogdet.h +39 -0
  38. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_log_softmax_native.h +26 -0
  39. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_masked_scale_cuda_dispatch.h +23 -0
  40. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_native_multi_head_attention_cpu_dispatch.h +23 -0
  41. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_neg_view.h +30 -0
  42. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_prelu_kernel_native.h +23 -0
  43. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_thnn_differentiable_lstm_cell_backward.h +30 -0
  44. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_to_dense_ops.h +39 -0
  45. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_unsafe_index_ops.h +28 -0
  46. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/arange_cpu_dispatch.h +24 -0
  47. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/atan2_cpu_dispatch.h +26 -0
  48. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/bitwise_or.h +67 -0
  49. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/cholesky_solve_native.h +22 -0
  50. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/cos_cpu_dispatch.h +26 -0
.gitattributes CHANGED
@@ -440,3 +440,13 @@ llava/lib/python3.10/lib2to3/tests/__pycache__/test_fixers.cpython-310.pyc filte
440
  parrot/lib/libtsan.so.0 filter=lfs diff=lfs merge=lfs -text
441
  parrot/lib/python3.10/site-packages/torch/lib/libc10.so filter=lfs diff=lfs merge=lfs -text
442
  parrot/lib/libtsan.so.0.0.0 filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
440
  parrot/lib/libtsan.so.0 filter=lfs diff=lfs merge=lfs -text
441
  parrot/lib/python3.10/site-packages/torch/lib/libc10.so filter=lfs diff=lfs merge=lfs -text
442
  parrot/lib/libtsan.so.0.0.0 filter=lfs diff=lfs merge=lfs -text
443
+ llava_next/lib/python3.10/site-packages/triton/third_party/cuda/bin/ptxas filter=lfs diff=lfs merge=lfs -text
444
+ llava/lib/python3.10/lib-dynload/_ctypes.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
445
+ llava/lib/python3.10/lib-dynload/_codecs_hk.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
446
+ llava/lib/python3.10/lib-dynload/_zoneinfo.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
447
+ llava/lib/python3.10/lib-dynload/_ssl.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
448
+ llava/lib/python3.10/lib-dynload/readline.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
449
+ llava/lib/python3.10/lib-dynload/_xxsubinterpreters.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
450
+ llava/lib/python3.10/lib-dynload/_socket.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
451
+ llava/lib/python3.10/lib-dynload/_lzma.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
452
+ llava/lib/python3.10/lib-dynload/_codecs_tw.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
llava/lib/python3.10/__pycache__/_markupbase.cpython-310.pyc ADDED
Binary file (7.57 kB). View file
 
llava/lib/python3.10/__pycache__/_strptime.cpython-310.pyc ADDED
Binary file (15.9 kB). View file
 
llava/lib/python3.10/__pycache__/abc.cpython-310.pyc ADDED
Binary file (6.75 kB). View file
 
llava/lib/python3.10/__pycache__/asyncore.cpython-310.pyc ADDED
Binary file (16.3 kB). View file
 
llava/lib/python3.10/__pycache__/cgitb.cpython-310.pyc ADDED
Binary file (10.3 kB). View file
 
llava/lib/python3.10/__pycache__/configparser.cpython-310.pyc ADDED
Binary file (45.5 kB). View file
 
llava/lib/python3.10/__pycache__/glob.cpython-310.pyc ADDED
Binary file (6.11 kB). View file
 
llava/lib/python3.10/__pycache__/gzip.cpython-310.pyc ADDED
Binary file (18.8 kB). View file
 
llava/lib/python3.10/__pycache__/imaplib.cpython-310.pyc ADDED
Binary file (42.3 kB). View file
 
llava/lib/python3.10/__pycache__/pydoc.cpython-310.pyc ADDED
Binary file (85.4 kB). View file
 
llava/lib/python3.10/__pycache__/rlcompleter.cpython-310.pyc ADDED
Binary file (6.21 kB). View file
 
llava/lib/python3.10/__pycache__/sre_constants.cpython-310.pyc ADDED
Binary file (6.35 kB). View file
 
llava/lib/python3.10/__pycache__/traceback.cpython-310.pyc ADDED
Binary file (22 kB). View file
 
llava/lib/python3.10/__pycache__/types.cpython-310.pyc ADDED
Binary file (9.78 kB). View file
 
llava/lib/python3.10/__pycache__/wave.cpython-310.pyc ADDED
Binary file (17.8 kB). View file
 
llava/lib/python3.10/__pycache__/zipfile.cpython-310.pyc ADDED
Binary file (61.8 kB). View file
 
llava/lib/python3.10/lib-dynload/_codecs_hk.cpython-310-x86_64-linux-gnu.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d5028632a8d7f167721ab2e0de42fb9e1c4ea5a80161005dabfb5c559f784dcb
3
+ size 194368
llava/lib/python3.10/lib-dynload/_codecs_tw.cpython-310-x86_64-linux-gnu.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:148bfe2a46df83fdece3e335e31ff78e92a273a1bfe660764988c88682ea3066
3
+ size 143232
llava/lib/python3.10/lib-dynload/_ctypes.cpython-310-x86_64-linux-gnu.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:679b8c5b2ea999168ab98fd52076ae0bfde863226dcb5d24431a7bb16dafaa9a
3
+ size 548912
llava/lib/python3.10/lib-dynload/_lzma.cpython-310-x86_64-linux-gnu.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ebfdcab2537ab089e99839348e41c28e8ce2b63d58d095f94c9f80a753748b0b
3
+ size 145376
llava/lib/python3.10/lib-dynload/_socket.cpython-310-x86_64-linux-gnu.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:457b4b1f4e960408aea7e2a1cbb92a8840ac68a682ecf5baf3450ee0e796fc00
3
+ size 283680
llava/lib/python3.10/lib-dynload/_ssl.cpython-310-x86_64-linux-gnu.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f93473eb2cfe128149c6d2973086402faa935d68856a103c752d0b513925432e
3
+ size 499736
llava/lib/python3.10/lib-dynload/_xxsubinterpreters.cpython-310-x86_64-linux-gnu.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:462ae2906ffebca406891ed2955376243f777336816acb18c6efc32aa1249e8d
3
+ size 149936
llava/lib/python3.10/lib-dynload/_zoneinfo.cpython-310-x86_64-linux-gnu.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eb38a8e751825aa95af436b17da44762c3f446a02c33ee193788d4ab02124017
3
+ size 160600
llava/lib/python3.10/lib-dynload/readline.cpython-310-x86_64-linux-gnu.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:96bcc678065c5ffab0324a9b557a7511f7fda79d85d84fafb7f2012261f4cee7
3
+ size 112992
llava/lib/python3.10/test/__pycache__/__init__.cpython-310.pyc ADDED
Binary file (383 Bytes). View file
 
llava/lib/python3.10/test/__pycache__/test_script_helper.cpython-310.pyc ADDED
Binary file (5.97 kB). View file
 
llava/lib/python3.10/test/__pycache__/test_support.cpython-310.pyc ADDED
Binary file (23.2 kB). View file
 
llava/lib/python3.10/test/support/__pycache__/__init__.cpython-310.pyc ADDED
Binary file (57.9 kB). View file
 
llava_next/lib/python3.10/site-packages/triton/third_party/cuda/bin/ptxas ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eb9c87832293aa88d4e6e58ecd0d77e25deb93da5c1a8bb91de7d7cd5b0281ac
3
+ size 20495600
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_batch_norm_impl_index_backward_compositeimplicitautograd_dispatch.h ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+ // @generated by torchgen/gen.py from DispatchKeyFunction.h
3
+
4
+ // NB: The implementing C++ file is RegisterDispatchKey.cpp
5
+
6
+ // The only #includes we need are for custom classes that have defaults in the C++ API
7
+ #include <c10/core/MemoryFormat.h>
8
+ #include <c10/core/Scalar.h>
9
+ #include <ATen/core/Reduction.h>
10
+
11
+ // Forward declarations of any types needed in the operator signatures.
12
+ // We can't directly include these classes because it will cause circular include dependencies.
13
+ // This file is included by TensorBody.h, which defines the Tensor class.
14
+ #include <ATen/core/ATen_fwd.h>
15
+
16
+ namespace at {
17
+
18
+ namespace compositeimplicitautograd {
19
+
20
+ TORCH_API ::std::tuple<at::Tensor,at::Tensor,at::Tensor> _batch_norm_impl_index_backward(int64_t impl_index, const at::Tensor & input, const at::Tensor & grad_output, const ::std::optional<at::Tensor> & weight, const ::std::optional<at::Tensor> & running_mean, const ::std::optional<at::Tensor> & running_var, const ::std::optional<at::Tensor> & save_mean, const ::std::optional<at::Tensor> & save_var_transform, bool train, double eps, ::std::array<bool,3> output_mask, const at::Tensor & reservedSpace);
21
+
22
+ } // namespace compositeimplicitautograd
23
+ } // namespace at
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_cdist_forward_cpu_dispatch.h ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+ // @generated by torchgen/gen.py from DispatchKeyFunction.h
3
+
4
+ // NB: The implementing C++ file is RegisterDispatchKey.cpp
5
+
6
+ // The only #includes we need are for custom classes that have defaults in the C++ API
7
+ #include <c10/core/MemoryFormat.h>
8
+ #include <c10/core/Scalar.h>
9
+ #include <ATen/core/Reduction.h>
10
+
11
+ // Forward declarations of any types needed in the operator signatures.
12
+ // We can't directly include these classes because it will cause circular include dependencies.
13
+ // This file is included by TensorBody.h, which defines the Tensor class.
14
+ #include <ATen/core/ATen_fwd.h>
15
+
16
+ namespace at {
17
+
18
+ namespace cpu {
19
+
20
+ TORCH_API at::Tensor _cdist_forward(const at::Tensor & x1, const at::Tensor & x2, double p, ::std::optional<int64_t> compute_mode);
21
+
22
+ } // namespace cpu
23
+ } // namespace at
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_foreach_acos_native.h ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from NativeFunction.h
4
+
5
+ #include <c10/core/Scalar.h>
6
+ #include <c10/core/Storage.h>
7
+ #include <c10/core/TensorOptions.h>
8
+ #include <c10/util/Deprecated.h>
9
+ #include <c10/util/Optional.h>
10
+ #include <c10/core/QScheme.h>
11
+ #include <ATen/core/Reduction.h>
12
+ #include <ATen/core/Tensor.h>
13
+ #include <tuple>
14
+ #include <vector>
15
+
16
+
17
+ namespace at {
18
+ namespace native {
19
+ TORCH_API ::std::vector<at::Tensor> foreach_tensor_acos_slow(at::TensorList self);
20
+ TORCH_API void _foreach_acos_out(at::TensorList self, at::TensorList out);
21
+ TORCH_API void foreach_tensor_acos_slow_(at::TensorList self);
22
+ TORCH_API ::std::vector<at::Tensor> foreach_tensor_acos_cuda(at::TensorList self);
23
+ TORCH_API void foreach_tensor_acos_cuda_(at::TensorList self);
24
+ } // namespace native
25
+ } // namespace at
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_foreach_sinh_cuda_dispatch.h ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+ // @generated by torchgen/gen.py from DispatchKeyFunction.h
3
+
4
+ // NB: The implementing C++ file is RegisterDispatchKey.cpp
5
+
6
+ // The only #includes we need are for custom classes that have defaults in the C++ API
7
+ #include <c10/core/MemoryFormat.h>
8
+ #include <c10/core/Scalar.h>
9
+ #include <ATen/core/Reduction.h>
10
+
11
+ // Forward declarations of any types needed in the operator signatures.
12
+ // We can't directly include these classes because it will cause circular include dependencies.
13
+ // This file is included by TensorBody.h, which defines the Tensor class.
14
+ #include <ATen/core/ATen_fwd.h>
15
+
16
+ namespace at {
17
+
18
+ namespace cuda {
19
+
20
+ TORCH_API ::std::vector<at::Tensor> _foreach_sinh(at::TensorList self);
21
+ TORCH_API void _foreach_sinh_(at::TensorList self);
22
+
23
+ } // namespace cuda
24
+ } // namespace at
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_foreach_trunc.h ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from Function.h
4
+
5
+ #include <ATen/Context.h>
6
+ #include <ATen/DeviceGuard.h>
7
+ #include <ATen/TensorUtils.h>
8
+ #include <ATen/TracerMode.h>
9
+ #include <ATen/core/Generator.h>
10
+ #include <ATen/core/Reduction.h>
11
+ #include <ATen/core/Tensor.h>
12
+ #include <c10/core/Scalar.h>
13
+ #include <c10/core/Storage.h>
14
+ #include <c10/core/TensorOptions.h>
15
+ #include <c10/util/Deprecated.h>
16
+ #include <c10/util/Optional.h>
17
+
18
+
19
+
20
+ #include <ATen/ops/_foreach_trunc_ops.h>
21
+
22
+ namespace at {
23
+
24
+
25
+ // aten::_foreach_trunc(Tensor[] self) -> Tensor[]
26
+ inline ::std::vector<at::Tensor> _foreach_trunc(at::TensorList self) {
27
+ return at::_ops::_foreach_trunc::call(self);
28
+ }
29
+
30
+ // aten::_foreach_trunc_(Tensor(a!)[] self) -> ()
31
+ inline void _foreach_trunc_(at::TensorList self) {
32
+ return at::_ops::_foreach_trunc_::call(self);
33
+ }
34
+
35
+ // aten::_foreach_trunc.out(Tensor[] self, *, Tensor(a!)[] out) -> ()
36
+ inline void _foreach_trunc_out(at::TensorList out, at::TensorList self) {
37
+ return at::_ops::_foreach_trunc_out::call(self, out);
38
+ }
39
+ // aten::_foreach_trunc.out(Tensor[] self, *, Tensor(a!)[] out) -> ()
40
+ inline void _foreach_trunc_outf(at::TensorList self, at::TensorList out) {
41
+ return at::_ops::_foreach_trunc_out::call(self, out);
42
+ }
43
+
44
+ }
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_linalg_slogdet.h ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from Function.h
4
+
5
+ #include <ATen/Context.h>
6
+ #include <ATen/DeviceGuard.h>
7
+ #include <ATen/TensorUtils.h>
8
+ #include <ATen/TracerMode.h>
9
+ #include <ATen/core/Generator.h>
10
+ #include <ATen/core/Reduction.h>
11
+ #include <ATen/core/Tensor.h>
12
+ #include <c10/core/Scalar.h>
13
+ #include <c10/core/Storage.h>
14
+ #include <c10/core/TensorOptions.h>
15
+ #include <c10/util/Deprecated.h>
16
+ #include <c10/util/Optional.h>
17
+
18
+
19
+
20
+ #include <ATen/ops/_linalg_slogdet_ops.h>
21
+
22
+ namespace at {
23
+
24
+
25
+ // aten::_linalg_slogdet(Tensor A) -> (Tensor sign, Tensor logabsdet, Tensor LU, Tensor pivots)
26
+ inline ::std::tuple<at::Tensor,at::Tensor,at::Tensor,at::Tensor> _linalg_slogdet(const at::Tensor & A) {
27
+ return at::_ops::_linalg_slogdet::call(A);
28
+ }
29
+
30
+ // aten::_linalg_slogdet.sign(Tensor A, *, Tensor(a!) sign, Tensor(b!) logabsdet, Tensor(c!) LU, Tensor(d!) pivots) -> (Tensor(a!) sign, Tensor(b!) logabsdet, Tensor(c!) LU, Tensor(d!) pivots)
31
+ inline ::std::tuple<at::Tensor &,at::Tensor &,at::Tensor &,at::Tensor &> _linalg_slogdet_out(at::Tensor & sign, at::Tensor & logabsdet, at::Tensor & LU, at::Tensor & pivots, const at::Tensor & A) {
32
+ return at::_ops::_linalg_slogdet_sign::call(A, sign, logabsdet, LU, pivots);
33
+ }
34
+ // aten::_linalg_slogdet.sign(Tensor A, *, Tensor(a!) sign, Tensor(b!) logabsdet, Tensor(c!) LU, Tensor(d!) pivots) -> (Tensor(a!) sign, Tensor(b!) logabsdet, Tensor(c!) LU, Tensor(d!) pivots)
35
+ inline ::std::tuple<at::Tensor &,at::Tensor &,at::Tensor &,at::Tensor &> _linalg_slogdet_outf(const at::Tensor & A, at::Tensor & sign, at::Tensor & logabsdet, at::Tensor & LU, at::Tensor & pivots) {
36
+ return at::_ops::_linalg_slogdet_sign::call(A, sign, logabsdet, LU, pivots);
37
+ }
38
+
39
+ }
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_log_softmax_native.h ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from NativeFunction.h
4
+
5
+ #include <c10/core/Scalar.h>
6
+ #include <c10/core/Storage.h>
7
+ #include <c10/core/TensorOptions.h>
8
+ #include <c10/util/Deprecated.h>
9
+ #include <c10/util/Optional.h>
10
+ #include <c10/core/QScheme.h>
11
+ #include <ATen/core/Reduction.h>
12
+ #include <ATen/core/Tensor.h>
13
+ #include <tuple>
14
+ #include <vector>
15
+ #include <ATen/ops/_log_softmax_meta.h>
16
+
17
+ namespace at {
18
+ namespace native {
19
+ struct TORCH_API structured_log_softmax_cpu_out : public at::meta::structured__log_softmax {
20
+ void impl(const at::Tensor & self, int64_t dim, bool half_to_float, const at::Tensor & out);
21
+ };
22
+ struct TORCH_API structured_log_softmax_cuda_out : public at::meta::structured__log_softmax {
23
+ void impl(const at::Tensor & self, int64_t dim, bool half_to_float, const at::Tensor & out);
24
+ };
25
+ } // namespace native
26
+ } // namespace at
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_masked_scale_cuda_dispatch.h ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+ // @generated by torchgen/gen.py from DispatchKeyFunction.h
3
+
4
+ // NB: The implementing C++ file is RegisterDispatchKey.cpp
5
+
6
+ // The only #includes we need are for custom classes that have defaults in the C++ API
7
+ #include <c10/core/MemoryFormat.h>
8
+ #include <c10/core/Scalar.h>
9
+ #include <ATen/core/Reduction.h>
10
+
11
+ // Forward declarations of any types needed in the operator signatures.
12
+ // We can't directly include these classes because it will cause circular include dependencies.
13
+ // This file is included by TensorBody.h, which defines the Tensor class.
14
+ #include <ATen/core/ATen_fwd.h>
15
+
16
+ namespace at {
17
+
18
+ namespace cuda {
19
+
20
+ TORCH_API at::Tensor _masked_scale(const at::Tensor & self, const at::Tensor & mask, double scale);
21
+
22
+ } // namespace cuda
23
+ } // namespace at
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_native_multi_head_attention_cpu_dispatch.h ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+ // @generated by torchgen/gen.py from DispatchKeyFunction.h
3
+
4
+ // NB: The implementing C++ file is RegisterDispatchKey.cpp
5
+
6
+ // The only #includes we need are for custom classes that have defaults in the C++ API
7
+ #include <c10/core/MemoryFormat.h>
8
+ #include <c10/core/Scalar.h>
9
+ #include <ATen/core/Reduction.h>
10
+
11
+ // Forward declarations of any types needed in the operator signatures.
12
+ // We can't directly include these classes because it will cause circular include dependencies.
13
+ // This file is included by TensorBody.h, which defines the Tensor class.
14
+ #include <ATen/core/ATen_fwd.h>
15
+
16
+ namespace at {
17
+
18
+ namespace cpu {
19
+
20
+ TORCH_API ::std::tuple<at::Tensor,at::Tensor> _native_multi_head_attention(const at::Tensor & query, const at::Tensor & key, const at::Tensor & value, int64_t embed_dim, int64_t num_head, const at::Tensor & qkv_weight, const at::Tensor & qkv_bias, const at::Tensor & proj_weight, const at::Tensor & proj_bias, const ::std::optional<at::Tensor> & mask={}, bool need_weights=true, bool average_attn_weights=true, ::std::optional<int64_t> mask_type=::std::nullopt);
21
+
22
+ } // namespace cpu
23
+ } // namespace at
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_neg_view.h ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from Function.h
4
+
5
+ #include <ATen/Context.h>
6
+ #include <ATen/DeviceGuard.h>
7
+ #include <ATen/TensorUtils.h>
8
+ #include <ATen/TracerMode.h>
9
+ #include <ATen/core/Generator.h>
10
+ #include <ATen/core/Reduction.h>
11
+ #include <ATen/core/Tensor.h>
12
+ #include <c10/core/Scalar.h>
13
+ #include <c10/core/Storage.h>
14
+ #include <c10/core/TensorOptions.h>
15
+ #include <c10/util/Deprecated.h>
16
+ #include <c10/util/Optional.h>
17
+
18
+
19
+
20
+ #include <ATen/ops/_neg_view_ops.h>
21
+
22
+ namespace at {
23
+
24
+
25
+ // aten::_neg_view(Tensor(a) self) -> Tensor(a)
26
+ inline at::Tensor _neg_view(const at::Tensor & self) {
27
+ return at::_ops::_neg_view::call(self);
28
+ }
29
+
30
+ }
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_prelu_kernel_native.h ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from NativeFunction.h
4
+
5
+ #include <c10/core/Scalar.h>
6
+ #include <c10/core/Storage.h>
7
+ #include <c10/core/TensorOptions.h>
8
+ #include <c10/util/Deprecated.h>
9
+ #include <c10/util/Optional.h>
10
+ #include <c10/core/QScheme.h>
11
+ #include <ATen/core/Reduction.h>
12
+ #include <ATen/core/Tensor.h>
13
+ #include <tuple>
14
+ #include <vector>
15
+
16
+
17
+ namespace at {
18
+ namespace native {
19
+ TORCH_API at::Tensor _prelu_kernel(const at::Tensor & self, const at::Tensor & weight);
20
+ TORCH_API at::Tensor mkldnn_prelu(const at::Tensor & self, const at::Tensor & weight);
21
+ TORCH_API at::Tensor _prelu_kernel_quantized_cpu(const at::Tensor & self, const at::Tensor & weight);
22
+ } // namespace native
23
+ } // namespace at
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_thnn_differentiable_lstm_cell_backward.h ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from Function.h
4
+
5
+ #include <ATen/Context.h>
6
+ #include <ATen/DeviceGuard.h>
7
+ #include <ATen/TensorUtils.h>
8
+ #include <ATen/TracerMode.h>
9
+ #include <ATen/core/Generator.h>
10
+ #include <ATen/core/Reduction.h>
11
+ #include <ATen/core/Tensor.h>
12
+ #include <c10/core/Scalar.h>
13
+ #include <c10/core/Storage.h>
14
+ #include <c10/core/TensorOptions.h>
15
+ #include <c10/util/Deprecated.h>
16
+ #include <c10/util/Optional.h>
17
+
18
+
19
+
20
+ #include <ATen/ops/_thnn_differentiable_lstm_cell_backward_ops.h>
21
+
22
+ namespace at {
23
+
24
+
25
+ // aten::_thnn_differentiable_lstm_cell_backward(Tensor? grad_hy, Tensor? grad_cy, Tensor input_gates, Tensor hidden_gates, Tensor? input_bias, Tensor? hidden_bias, Tensor cx, Tensor cy) -> (Tensor, Tensor, Tensor, Tensor, Tensor)
26
+ inline ::std::tuple<at::Tensor,at::Tensor,at::Tensor,at::Tensor,at::Tensor> _thnn_differentiable_lstm_cell_backward(const ::std::optional<at::Tensor> & grad_hy, const ::std::optional<at::Tensor> & grad_cy, const at::Tensor & input_gates, const at::Tensor & hidden_gates, const ::std::optional<at::Tensor> & input_bias, const ::std::optional<at::Tensor> & hidden_bias, const at::Tensor & cx, const at::Tensor & cy) {
27
+ return at::_ops::_thnn_differentiable_lstm_cell_backward::call(grad_hy, grad_cy, input_gates, hidden_gates, input_bias, hidden_bias, cx, cy);
28
+ }
29
+
30
+ }
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_to_dense_ops.h ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from Operator.h
4
+
5
+ #include <tuple>
6
+ #include <vector>
7
+
8
+ // Forward declarations of any types needed in the operator signatures.
9
+ // We can't directly include these classes because it will cause circular include dependencies.
10
+ // This file is included by TensorBody.h, which defines the Tensor class.
11
+ #include <ATen/core/ATen_fwd.h>
12
+
13
+ namespace at {
14
+ namespace _ops {
15
+
16
+
17
+ struct TORCH_API _to_dense {
18
+ using schema = at::Tensor (const at::Tensor &, ::std::optional<at::ScalarType>, ::std::optional<bool>);
19
+ using ptr_schema = schema*;
20
+ // See Note [static constexpr char* members for windows NVCC]
21
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(name, "aten::_to_dense")
22
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(overload_name, "")
23
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(schema_str, "_to_dense(Tensor self, ScalarType? dtype=None, bool? masked_grad=None) -> Tensor")
24
+ static at::Tensor call(const at::Tensor & self, ::std::optional<at::ScalarType> dtype, ::std::optional<bool> masked_grad);
25
+ static at::Tensor redispatch(c10::DispatchKeySet dispatchKeySet, const at::Tensor & self, ::std::optional<at::ScalarType> dtype, ::std::optional<bool> masked_grad);
26
+ };
27
+
28
+ struct TORCH_API _to_dense_out {
29
+ using schema = at::Tensor & (const at::Tensor &, ::std::optional<at::ScalarType>, ::std::optional<bool>, at::Tensor &);
30
+ using ptr_schema = schema*;
31
+ // See Note [static constexpr char* members for windows NVCC]
32
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(name, "aten::_to_dense")
33
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(overload_name, "out")
34
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(schema_str, "_to_dense.out(Tensor self, ScalarType? dtype=None, bool? masked_grad=None, *, Tensor(a!) out) -> Tensor(a!)")
35
+ static at::Tensor & call(const at::Tensor & self, ::std::optional<at::ScalarType> dtype, ::std::optional<bool> masked_grad, at::Tensor & out);
36
+ static at::Tensor & redispatch(c10::DispatchKeySet dispatchKeySet, const at::Tensor & self, ::std::optional<at::ScalarType> dtype, ::std::optional<bool> masked_grad, at::Tensor & out);
37
+ };
38
+
39
+ }} // namespace at::_ops
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_unsafe_index_ops.h ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from Operator.h
4
+
5
+ #include <tuple>
6
+ #include <vector>
7
+
8
+ // Forward declarations of any types needed in the operator signatures.
9
+ // We can't directly include these classes because it will cause circular include dependencies.
10
+ // This file is included by TensorBody.h, which defines the Tensor class.
11
+ #include <ATen/core/ATen_fwd.h>
12
+
13
+ namespace at {
14
+ namespace _ops {
15
+
16
+
17
+ struct TORCH_API _unsafe_index_Tensor {
18
+ using schema = at::Tensor (const at::Tensor &, const c10::List<::std::optional<at::Tensor>> &);
19
+ using ptr_schema = schema*;
20
+ // See Note [static constexpr char* members for windows NVCC]
21
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(name, "aten::_unsafe_index")
22
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(overload_name, "Tensor")
23
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(schema_str, "_unsafe_index.Tensor(Tensor self, Tensor?[] indices) -> Tensor")
24
+ static at::Tensor call(const at::Tensor & self, const c10::List<::std::optional<at::Tensor>> & indices);
25
+ static at::Tensor redispatch(c10::DispatchKeySet dispatchKeySet, const at::Tensor & self, const c10::List<::std::optional<at::Tensor>> & indices);
26
+ };
27
+
28
+ }} // namespace at::_ops
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/arange_cpu_dispatch.h ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+ // @generated by torchgen/gen.py from DispatchKeyFunction.h
3
+
4
+ // NB: The implementing C++ file is RegisterDispatchKey.cpp
5
+
6
+ // The only #includes we need are for custom classes that have defaults in the C++ API
7
+ #include <c10/core/MemoryFormat.h>
8
+ #include <c10/core/Scalar.h>
9
+ #include <ATen/core/Reduction.h>
10
+
11
+ // Forward declarations of any types needed in the operator signatures.
12
+ // We can't directly include these classes because it will cause circular include dependencies.
13
+ // This file is included by TensorBody.h, which defines the Tensor class.
14
+ #include <ATen/core/ATen_fwd.h>
15
+
16
+ namespace at {
17
+
18
+ namespace cpu {
19
+
20
+ TORCH_API at::Tensor & arange_out(at::Tensor & out, const at::Scalar & start, const at::Scalar & end, const at::Scalar & step);
21
+ TORCH_API at::Tensor & arange_outf(const at::Scalar & start, const at::Scalar & end, const at::Scalar & step, at::Tensor & out);
22
+
23
+ } // namespace cpu
24
+ } // namespace at
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/atan2_cpu_dispatch.h ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+ // @generated by torchgen/gen.py from DispatchKeyFunction.h
3
+
4
+ // NB: The implementing C++ file is RegisterDispatchKey.cpp
5
+
6
+ // The only #includes we need are for custom classes that have defaults in the C++ API
7
+ #include <c10/core/MemoryFormat.h>
8
+ #include <c10/core/Scalar.h>
9
+ #include <ATen/core/Reduction.h>
10
+
11
+ // Forward declarations of any types needed in the operator signatures.
12
+ // We can't directly include these classes because it will cause circular include dependencies.
13
+ // This file is included by TensorBody.h, which defines the Tensor class.
14
+ #include <ATen/core/ATen_fwd.h>
15
+
16
+ namespace at {
17
+
18
+ namespace cpu {
19
+
20
+ TORCH_API at::Tensor atan2(const at::Tensor & self, const at::Tensor & other);
21
+ TORCH_API at::Tensor & atan2_out(at::Tensor & out, const at::Tensor & self, const at::Tensor & other);
22
+ TORCH_API at::Tensor & atan2_outf(const at::Tensor & self, const at::Tensor & other, at::Tensor & out);
23
+ TORCH_API at::Tensor & atan2_(at::Tensor & self, const at::Tensor & other);
24
+
25
+ } // namespace cpu
26
+ } // namespace at
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/bitwise_or.h ADDED
@@ -0,0 +1,67 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from Function.h
4
+
5
+ #include <ATen/Context.h>
6
+ #include <ATen/DeviceGuard.h>
7
+ #include <ATen/TensorUtils.h>
8
+ #include <ATen/TracerMode.h>
9
+ #include <ATen/core/Generator.h>
10
+ #include <ATen/core/Reduction.h>
11
+ #include <ATen/core/Tensor.h>
12
+ #include <c10/core/Scalar.h>
13
+ #include <c10/core/Storage.h>
14
+ #include <c10/core/TensorOptions.h>
15
+ #include <c10/util/Deprecated.h>
16
+ #include <c10/util/Optional.h>
17
+
18
+
19
+
20
+ #include <ATen/ops/bitwise_or_ops.h>
21
+
22
+ namespace at {
23
+
24
+
25
+ // aten::bitwise_or.Tensor_out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)
26
+ inline at::Tensor & bitwise_or_out(at::Tensor & out, const at::Tensor & self, const at::Tensor & other) {
27
+ return at::_ops::bitwise_or_Tensor_out::call(self, other, out);
28
+ }
29
+ // aten::bitwise_or.Tensor_out(Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)
30
+ inline at::Tensor & bitwise_or_outf(const at::Tensor & self, const at::Tensor & other, at::Tensor & out) {
31
+ return at::_ops::bitwise_or_Tensor_out::call(self, other, out);
32
+ }
33
+
34
+ // aten::bitwise_or.Scalar_out(Tensor self, Scalar other, *, Tensor(a!) out) -> Tensor(a!)
35
+ inline at::Tensor & bitwise_or_out(at::Tensor & out, const at::Tensor & self, const at::Scalar & other) {
36
+ return at::_ops::bitwise_or_Scalar_out::call(self, other, out);
37
+ }
38
+ // aten::bitwise_or.Scalar_out(Tensor self, Scalar other, *, Tensor(a!) out) -> Tensor(a!)
39
+ inline at::Tensor & bitwise_or_outf(const at::Tensor & self, const at::Scalar & other, at::Tensor & out) {
40
+ return at::_ops::bitwise_or_Scalar_out::call(self, other, out);
41
+ }
42
+
43
+ // aten::bitwise_or.Scalar(Tensor self, Scalar other) -> Tensor
44
+ inline at::Tensor bitwise_or(const at::Tensor & self, const at::Scalar & other) {
45
+ return at::_ops::bitwise_or_Scalar::call(self, other);
46
+ }
47
+
48
+ // aten::bitwise_or.Scalar_Tensor(Scalar self, Tensor other) -> Tensor
49
+ inline at::Tensor bitwise_or(const at::Scalar & self, const at::Tensor & other) {
50
+ return at::_ops::bitwise_or_Scalar_Tensor::call(self, other);
51
+ }
52
+
53
+ // aten::bitwise_or.Tensor(Tensor self, Tensor other) -> Tensor
54
+ inline at::Tensor bitwise_or(const at::Tensor & self, const at::Tensor & other) {
55
+ return at::_ops::bitwise_or_Tensor::call(self, other);
56
+ }
57
+
58
+ // aten::bitwise_or.Scalar_Tensor_out(Scalar self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)
59
+ inline at::Tensor & bitwise_or_out(at::Tensor & out, const at::Scalar & self, const at::Tensor & other) {
60
+ return at::_ops::bitwise_or_Scalar_Tensor_out::call(self, other, out);
61
+ }
62
+ // aten::bitwise_or.Scalar_Tensor_out(Scalar self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)
63
+ inline at::Tensor & bitwise_or_outf(const at::Scalar & self, const at::Tensor & other, at::Tensor & out) {
64
+ return at::_ops::bitwise_or_Scalar_Tensor_out::call(self, other, out);
65
+ }
66
+
67
+ }
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/cholesky_solve_native.h ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from NativeFunction.h
4
+
5
+ #include <c10/core/Scalar.h>
6
+ #include <c10/core/Storage.h>
7
+ #include <c10/core/TensorOptions.h>
8
+ #include <c10/util/Deprecated.h>
9
+ #include <c10/util/Optional.h>
10
+ #include <c10/core/QScheme.h>
11
+ #include <ATen/core/Reduction.h>
12
+ #include <ATen/core/Tensor.h>
13
+ #include <tuple>
14
+ #include <vector>
15
+
16
+
17
+ namespace at {
18
+ namespace native {
19
+ TORCH_API at::Tensor cholesky_solve(const at::Tensor & self, const at::Tensor & input2, bool upper=false);
20
+ TORCH_API at::Tensor & cholesky_solve_out(const at::Tensor & self, const at::Tensor & input2, bool upper, at::Tensor & out);
21
+ } // namespace native
22
+ } // namespace at
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/cos_cpu_dispatch.h ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+ // @generated by torchgen/gen.py from DispatchKeyFunction.h
3
+
4
+ // NB: The implementing C++ file is RegisterDispatchKey.cpp
5
+
6
+ // The only #includes we need are for custom classes that have defaults in the C++ API
7
+ #include <c10/core/MemoryFormat.h>
8
+ #include <c10/core/Scalar.h>
9
+ #include <ATen/core/Reduction.h>
10
+
11
+ // Forward declarations of any types needed in the operator signatures.
12
+ // We can't directly include these classes because it will cause circular include dependencies.
13
+ // This file is included by TensorBody.h, which defines the Tensor class.
14
+ #include <ATen/core/ATen_fwd.h>
15
+
16
+ namespace at {
17
+
18
+ namespace cpu {
19
+
20
+ TORCH_API at::Tensor cos(const at::Tensor & self);
21
+ TORCH_API at::Tensor & cos_out(at::Tensor & out, const at::Tensor & self);
22
+ TORCH_API at::Tensor & cos_outf(const at::Tensor & self, at::Tensor & out);
23
+ TORCH_API at::Tensor & cos_(at::Tensor & self);
24
+
25
+ } // namespace cpu
26
+ } // namespace at