ZTWHHH commited on
Commit
0bc9fb5
·
verified ·
1 Parent(s): 6da29cc

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +6 -0
  2. llava/lib/libasan.so.6.0.0 +3 -0
  3. llava/lib/libbz2.so +3 -0
  4. llava/lib/libbz2.so.1.0.8 +3 -0
  5. llava/lib/libncurses.so +3 -0
  6. llava/lib/libssl.so.3 +3 -0
  7. llava/lib/libtinfo.so.6 +3 -0
  8. llava/lib/tcl8.6/msgs/af_za.msg +6 -0
  9. llava/lib/tcl8.6/msgs/ar_sy.msg +39 -0
  10. llava/lib/tcl8.6/msgs/en_nz.msg +7 -0
  11. llava/lib/tcl8.6/msgs/eo.msg +54 -0
  12. llava/lib/tcl8.6/msgs/es.msg +52 -0
  13. llava/lib/tcl8.6/msgs/es_gt.msg +6 -0
  14. llava/lib/tcl8.6/msgs/es_hn.msg +6 -0
  15. llava/lib/tcl8.6/msgs/es_ni.msg +6 -0
  16. llava/lib/tcl8.6/msgs/es_pr.msg +6 -0
  17. llava/lib/tcl8.6/msgs/gl.msg +47 -0
  18. llava/lib/tcl8.6/msgs/kok_in.msg +6 -0
  19. llava/lib/tcl8.6/msgs/mr.msg +39 -0
  20. llava/lib/tcl8.6/msgs/nl.msg +50 -0
  21. llava/lib/tcl8.6/msgs/nl_be.msg +7 -0
  22. llava/lib/tcl8.6/msgs/nn.msg +52 -0
  23. llava/lib/tcl8.6/msgs/pt.msg +50 -0
  24. llava/lib/tcl8.6/msgs/ro.msg +52 -0
  25. llava/lib/tcl8.6/msgs/ru_ua.msg +6 -0
  26. llava/lib/tcl8.6/msgs/sw.msg +49 -0
  27. llava/lib/tcl8.6/msgs/th.msg +54 -0
  28. llava/lib/tcl8.6/msgs/zh_sg.msg +8 -0
  29. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_foreach_floor_cuda_dispatch.h +24 -0
  30. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_foreach_log2.h +44 -0
  31. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_foreach_neg_cuda_dispatch.h +24 -0
  32. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_lazy_clone_ops.h +28 -0
  33. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_nnpack_spatial_convolution.h +91 -0
  34. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_prelu_kernel_backward_native.h +22 -0
  35. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_sparse_semi_structured_apply_native.h +21 -0
  36. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_unsafe_index_put_compositeexplicitautograd_dispatch.h +23 -0
  37. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_upsample_bicubic2d_aa_backward_meta_dispatch.h +28 -0
  38. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_upsample_bilinear2d_aa_meta.h +27 -0
  39. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/adaptive_avg_pool3d.h +91 -0
  40. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/adaptive_max_pool2d_backward_meta.h +27 -0
  41. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/atan2_meta_dispatch.h +26 -0
  42. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/batch_norm_gather_stats_compositeexplicitautograd_dispatch.h +24 -0
  43. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/batch_norm_gather_stats_with_counts_compositeexplicitautograd_dispatch.h +24 -0
  44. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/complex_cuda_dispatch.h +24 -0
  45. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/conv1d_ops.h +39 -0
  46. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/cudnn_grid_sampler_ops.h +39 -0
  47. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/cummax.h +53 -0
  48. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/diagonal_native.h +22 -0
  49. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/diff_compositeimplicitautograd_dispatch.h +25 -0
  50. parrot/lib/python3.10/site-packages/torch/include/ATen/ops/embedding_dense_backward.h +91 -0
.gitattributes CHANGED
@@ -403,3 +403,9 @@ llava/lib/libquadmath.so filter=lfs diff=lfs merge=lfs -text
403
  llava/lib/liblzma.so.5 filter=lfs diff=lfs merge=lfs -text
404
  llava/lib/libncurses.so.6 filter=lfs diff=lfs merge=lfs -text
405
  llava/lib/libtsan.so.0.0.0 filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
403
  llava/lib/liblzma.so.5 filter=lfs diff=lfs merge=lfs -text
404
  llava/lib/libncurses.so.6 filter=lfs diff=lfs merge=lfs -text
405
  llava/lib/libtsan.so.0.0.0 filter=lfs diff=lfs merge=lfs -text
406
+ llava/lib/libncurses.so filter=lfs diff=lfs merge=lfs -text
407
+ llava/lib/libtinfo.so.6 filter=lfs diff=lfs merge=lfs -text
408
+ llava/lib/libbz2.so filter=lfs diff=lfs merge=lfs -text
409
+ llava/lib/libasan.so.6.0.0 filter=lfs diff=lfs merge=lfs -text
410
+ llava/lib/libbz2.so.1.0.8 filter=lfs diff=lfs merge=lfs -text
411
+ llava/lib/libssl.so.3 filter=lfs diff=lfs merge=lfs -text
llava/lib/libasan.so.6.0.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a8a7995a4d84a8817af8d1604bef621e99d0622df4eda14f6fe5245735a952e
3
+ size 7575272
llava/lib/libbz2.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4979469ae49ac144f62202f75bbdd69b17197aedb879d633337c8cf7e4aba301
3
+ size 229016
llava/lib/libbz2.so.1.0.8 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4979469ae49ac144f62202f75bbdd69b17197aedb879d633337c8cf7e4aba301
3
+ size 229016
llava/lib/libncurses.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7fa4e5e93804d78660b0eef727cdb4211209e1742e4ad3669348022668d90962
3
+ size 271304
llava/lib/libssl.so.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3480c91df4e0c1a33514955568641405e37924f680e8ba42f494a209640516c6
3
+ size 775712
llava/lib/libtinfo.so.6 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f5679c9d7cc0ec2d6b08c6058198667efe71f657e89dcc0bd7adcf5d6cbc80c5
3
+ size 287080
llava/lib/tcl8.6/msgs/af_za.msg ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ # created by tools/loadICU.tcl -- do not edit
2
+ namespace eval ::tcl::clock {
3
+ ::msgcat::mcset af_ZA DATE_FORMAT "%d %B %Y"
4
+ ::msgcat::mcset af_ZA TIME_FORMAT_12 "%l:%M:%S %P"
5
+ ::msgcat::mcset af_ZA DATE_TIME_FORMAT "%d %B %Y %l:%M:%S %P %z"
6
+ }
llava/lib/tcl8.6/msgs/ar_sy.msg ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # created by tools/loadICU.tcl -- do not edit
2
+ namespace eval ::tcl::clock {
3
+ ::msgcat::mcset ar_SY DAYS_OF_WEEK_ABBREV [list \
4
+ "\u0627\u0644\u0623\u062d\u062f"\
5
+ "\u0627\u0644\u0627\u062b\u0646\u064a\u0646"\
6
+ "\u0627\u0644\u062b\u0644\u0627\u062b\u0627\u0621"\
7
+ "\u0627\u0644\u0623\u0631\u0628\u0639\u0627\u0621"\
8
+ "\u0627\u0644\u062e\u0645\u064a\u0633"\
9
+ "\u0627\u0644\u062c\u0645\u0639\u0629"\
10
+ "\u0627\u0644\u0633\u0628\u062a"]
11
+ ::msgcat::mcset ar_SY MONTHS_ABBREV [list \
12
+ "\u0643\u0627\u0646\u0648\u0646 \u0627\u0644\u062b\u0627\u0646\u064a"\
13
+ "\u0634\u0628\u0627\u0637"\
14
+ "\u0622\u0630\u0627\u0631"\
15
+ "\u0646\u064a\u0633\u0627\u0646"\
16
+ "\u0646\u0648\u0627\u0631"\
17
+ "\u062d\u0632\u064a\u0631\u0627\u0646"\
18
+ "\u062a\u0645\u0648\u0632"\
19
+ "\u0622\u0628"\
20
+ "\u0623\u064a\u0644\u0648\u0644"\
21
+ "\u062a\u0634\u0631\u064a\u0646 \u0627\u0644\u0623\u0648\u0644"\
22
+ "\u062a\u0634\u0631\u064a\u0646 \u0627\u0644\u062b\u0627\u0646\u064a"\
23
+ "\u0643\u0627\u0646\u0648\u0646 \u0627\u0644\u0623\u0648\u0644"\
24
+ ""]
25
+ ::msgcat::mcset ar_SY MONTHS_FULL [list \
26
+ "\u0643\u0627\u0646\u0648\u0646 \u0627\u0644\u062b\u0627\u0646\u064a"\
27
+ "\u0634\u0628\u0627\u0637"\
28
+ "\u0622\u0630\u0627\u0631"\
29
+ "\u0646\u064a\u0633\u0627\u0646"\
30
+ "\u0646\u0648\u0627\u0631\u0627\u0646"\
31
+ "\u062d\u0632\u064a\u0631"\
32
+ "\u062a\u0645\u0648\u0632"\
33
+ "\u0622\u0628"\
34
+ "\u0623\u064a\u0644\u0648\u0644"\
35
+ "\u062a\u0634\u0631\u064a\u0646 \u0627\u0644\u0623\u0648\u0644"\
36
+ "\u062a\u0634\u0631\u064a\u0646 \u0627\u0644\u062b\u0627\u0646\u064a"\
37
+ "\u0643\u0627\u0646\u0648\u0646 \u0627\u0644\u0623\u0648\u0644"\
38
+ ""]
39
+ }
llava/lib/tcl8.6/msgs/en_nz.msg ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ # created by tools/loadICU.tcl -- do not edit
2
+ namespace eval ::tcl::clock {
3
+ ::msgcat::mcset en_NZ DATE_FORMAT "%e/%m/%Y"
4
+ ::msgcat::mcset en_NZ TIME_FORMAT "%H:%M:%S"
5
+ ::msgcat::mcset en_NZ TIME_FORMAT_12 "%I:%M:%S %P %z"
6
+ ::msgcat::mcset en_NZ DATE_TIME_FORMAT "%e/%m/%Y %H:%M:%S %z"
7
+ }
llava/lib/tcl8.6/msgs/eo.msg ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # created by tools/loadICU.tcl -- do not edit
2
+ namespace eval ::tcl::clock {
3
+ ::msgcat::mcset eo DAYS_OF_WEEK_ABBREV [list \
4
+ "di"\
5
+ "lu"\
6
+ "ma"\
7
+ "me"\
8
+ "\u0135a"\
9
+ "ve"\
10
+ "sa"]
11
+ ::msgcat::mcset eo DAYS_OF_WEEK_FULL [list \
12
+ "diman\u0109o"\
13
+ "lundo"\
14
+ "mardo"\
15
+ "merkredo"\
16
+ "\u0135a\u016ddo"\
17
+ "vendredo"\
18
+ "sabato"]
19
+ ::msgcat::mcset eo MONTHS_ABBREV [list \
20
+ "jan"\
21
+ "feb"\
22
+ "mar"\
23
+ "apr"\
24
+ "maj"\
25
+ "jun"\
26
+ "jul"\
27
+ "a\u016dg"\
28
+ "sep"\
29
+ "okt"\
30
+ "nov"\
31
+ "dec"\
32
+ ""]
33
+ ::msgcat::mcset eo MONTHS_FULL [list \
34
+ "januaro"\
35
+ "februaro"\
36
+ "marto"\
37
+ "aprilo"\
38
+ "majo"\
39
+ "junio"\
40
+ "julio"\
41
+ "a\u016dgusto"\
42
+ "septembro"\
43
+ "oktobro"\
44
+ "novembro"\
45
+ "decembro"\
46
+ ""]
47
+ ::msgcat::mcset eo BCE "aK"
48
+ ::msgcat::mcset eo CE "pK"
49
+ ::msgcat::mcset eo AM "atm"
50
+ ::msgcat::mcset eo PM "ptm"
51
+ ::msgcat::mcset eo DATE_FORMAT "%Y-%b-%d"
52
+ ::msgcat::mcset eo TIME_FORMAT "%H:%M:%S"
53
+ ::msgcat::mcset eo DATE_TIME_FORMAT "%Y-%b-%d %H:%M:%S %z"
54
+ }
llava/lib/tcl8.6/msgs/es.msg ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # created by tools/loadICU.tcl -- do not edit
2
+ namespace eval ::tcl::clock {
3
+ ::msgcat::mcset es DAYS_OF_WEEK_ABBREV [list \
4
+ "dom"\
5
+ "lun"\
6
+ "mar"\
7
+ "mi\u00e9"\
8
+ "jue"\
9
+ "vie"\
10
+ "s\u00e1b"]
11
+ ::msgcat::mcset es DAYS_OF_WEEK_FULL [list \
12
+ "domingo"\
13
+ "lunes"\
14
+ "martes"\
15
+ "mi\u00e9rcoles"\
16
+ "jueves"\
17
+ "viernes"\
18
+ "s\u00e1bado"]
19
+ ::msgcat::mcset es MONTHS_ABBREV [list \
20
+ "ene"\
21
+ "feb"\
22
+ "mar"\
23
+ "abr"\
24
+ "may"\
25
+ "jun"\
26
+ "jul"\
27
+ "ago"\
28
+ "sep"\
29
+ "oct"\
30
+ "nov"\
31
+ "dic"\
32
+ ""]
33
+ ::msgcat::mcset es MONTHS_FULL [list \
34
+ "enero"\
35
+ "febrero"\
36
+ "marzo"\
37
+ "abril"\
38
+ "mayo"\
39
+ "junio"\
40
+ "julio"\
41
+ "agosto"\
42
+ "septiembre"\
43
+ "octubre"\
44
+ "noviembre"\
45
+ "diciembre"\
46
+ ""]
47
+ ::msgcat::mcset es BCE "a.C."
48
+ ::msgcat::mcset es CE "d.C."
49
+ ::msgcat::mcset es DATE_FORMAT "%e de %B de %Y"
50
+ ::msgcat::mcset es TIME_FORMAT "%k:%M:%S"
51
+ ::msgcat::mcset es DATE_TIME_FORMAT "%e de %B de %Y %k:%M:%S %z"
52
+ }
llava/lib/tcl8.6/msgs/es_gt.msg ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ # created by tools/loadICU.tcl -- do not edit
2
+ namespace eval ::tcl::clock {
3
+ ::msgcat::mcset es_GT DATE_FORMAT "%e/%m/%Y"
4
+ ::msgcat::mcset es_GT TIME_FORMAT_12 "%I:%M:%S %P"
5
+ ::msgcat::mcset es_GT DATE_TIME_FORMAT "%e/%m/%Y %I:%M:%S %P %z"
6
+ }
llava/lib/tcl8.6/msgs/es_hn.msg ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ # created by tools/loadICU.tcl -- do not edit
2
+ namespace eval ::tcl::clock {
3
+ ::msgcat::mcset es_HN DATE_FORMAT "%m-%d-%Y"
4
+ ::msgcat::mcset es_HN TIME_FORMAT_12 "%I:%M:%S %P"
5
+ ::msgcat::mcset es_HN DATE_TIME_FORMAT "%m-%d-%Y %I:%M:%S %P %z"
6
+ }
llava/lib/tcl8.6/msgs/es_ni.msg ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ # created by tools/loadICU.tcl -- do not edit
2
+ namespace eval ::tcl::clock {
3
+ ::msgcat::mcset es_NI DATE_FORMAT "%m-%d-%Y"
4
+ ::msgcat::mcset es_NI TIME_FORMAT_12 "%I:%M:%S %P"
5
+ ::msgcat::mcset es_NI DATE_TIME_FORMAT "%m-%d-%Y %I:%M:%S %P %z"
6
+ }
llava/lib/tcl8.6/msgs/es_pr.msg ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ # created by tools/loadICU.tcl -- do not edit
2
+ namespace eval ::tcl::clock {
3
+ ::msgcat::mcset es_PR DATE_FORMAT "%m-%d-%Y"
4
+ ::msgcat::mcset es_PR TIME_FORMAT_12 "%I:%M:%S %P"
5
+ ::msgcat::mcset es_PR DATE_TIME_FORMAT "%m-%d-%Y %I:%M:%S %P %z"
6
+ }
llava/lib/tcl8.6/msgs/gl.msg ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # created by tools/loadICU.tcl -- do not edit
2
+ namespace eval ::tcl::clock {
3
+ ::msgcat::mcset gl DAYS_OF_WEEK_ABBREV [list \
4
+ "Dom"\
5
+ "Lun"\
6
+ "Mar"\
7
+ "M\u00e9r"\
8
+ "Xov"\
9
+ "Ven"\
10
+ "S\u00e1b"]
11
+ ::msgcat::mcset gl DAYS_OF_WEEK_FULL [list \
12
+ "Domingo"\
13
+ "Luns"\
14
+ "Martes"\
15
+ "M\u00e9rcores"\
16
+ "Xoves"\
17
+ "Venres"\
18
+ "S\u00e1bado"]
19
+ ::msgcat::mcset gl MONTHS_ABBREV [list \
20
+ "Xan"\
21
+ "Feb"\
22
+ "Mar"\
23
+ "Abr"\
24
+ "Mai"\
25
+ "Xu\u00f1"\
26
+ "Xul"\
27
+ "Ago"\
28
+ "Set"\
29
+ "Out"\
30
+ "Nov"\
31
+ "Dec"\
32
+ ""]
33
+ ::msgcat::mcset gl MONTHS_FULL [list \
34
+ "Xaneiro"\
35
+ "Febreiro"\
36
+ "Marzo"\
37
+ "Abril"\
38
+ "Maio"\
39
+ "Xu\u00f1o"\
40
+ "Xullo"\
41
+ "Agosto"\
42
+ "Setembro"\
43
+ "Outubro"\
44
+ "Novembro"\
45
+ "Decembro"\
46
+ ""]
47
+ }
llava/lib/tcl8.6/msgs/kok_in.msg ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ # created by tools/loadICU.tcl -- do not edit
2
+ namespace eval ::tcl::clock {
3
+ ::msgcat::mcset kok_IN DATE_FORMAT "%d %M %Y"
4
+ ::msgcat::mcset kok_IN TIME_FORMAT_12 "%I:%M:%S %P"
5
+ ::msgcat::mcset kok_IN DATE_TIME_FORMAT "%d %M %Y %I:%M:%S %P %z"
6
+ }
llava/lib/tcl8.6/msgs/mr.msg ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # created by tools/loadICU.tcl -- do not edit
2
+ namespace eval ::tcl::clock {
3
+ ::msgcat::mcset mr DAYS_OF_WEEK_FULL [list \
4
+ "\u0930\u0935\u093f\u0935\u093e\u0930"\
5
+ "\u0938\u094b\u092e\u0935\u093e\u0930"\
6
+ "\u092e\u0902\u0917\u0933\u0935\u093e\u0930"\
7
+ "\u092e\u0902\u0917\u0933\u0935\u093e\u0930"\
8
+ "\u0917\u0941\u0930\u0941\u0935\u093e\u0930"\
9
+ "\u0936\u0941\u0915\u094d\u0930\u0935\u093e\u0930"\
10
+ "\u0936\u0928\u093f\u0935\u093e\u0930"]
11
+ ::msgcat::mcset mr MONTHS_ABBREV [list \
12
+ "\u091c\u093e\u0928\u0947\u0935\u093e\u0930\u0940"\
13
+ "\u092b\u0947\u092c\u0943\u0935\u093e\u0930\u0940"\
14
+ "\u092e\u093e\u0930\u094d\u091a"\
15
+ "\u090f\u092a\u094d\u0930\u093f\u0932"\
16
+ "\u092e\u0947"\
17
+ "\u091c\u0942\u0928"\
18
+ "\u091c\u0941\u0932\u0948"\
19
+ "\u0913\u0917\u0938\u094d\u091f"\
20
+ "\u0938\u0947\u092a\u094d\u091f\u0947\u0902\u092c\u0930"\
21
+ "\u0913\u0915\u094d\u091f\u094b\u092c\u0930"\
22
+ "\u0928\u094b\u0935\u094d\u0939\u0947\u0902\u092c\u0930"\
23
+ "\u0921\u093f\u0938\u0947\u0902\u092c\u0930"]
24
+ ::msgcat::mcset mr MONTHS_FULL [list \
25
+ "\u091c\u093e\u0928\u0947\u0935\u093e\u0930\u0940"\
26
+ "\u092b\u0947\u092c\u0943\u0935\u093e\u0930\u0940"\
27
+ "\u092e\u093e\u0930\u094d\u091a"\
28
+ "\u090f\u092a\u094d\u0930\u093f\u0932"\
29
+ "\u092e\u0947"\
30
+ "\u091c\u0942\u0928"\
31
+ "\u091c\u0941\u0932\u0948"\
32
+ "\u0913\u0917\u0938\u094d\u091f"\
33
+ "\u0938\u0947\u092a\u094d\u091f\u0947\u0902\u092c\u0930"\
34
+ "\u0913\u0915\u094d\u091f\u094b\u092c\u0930"\
35
+ "\u0928\u094b\u0935\u094d\u0939\u0947\u0902\u092c\u0930"\
36
+ "\u0921\u093f\u0938\u0947\u0902\u092c\u0930"]
37
+ ::msgcat::mcset mr AM "BC"
38
+ ::msgcat::mcset mr PM "AD"
39
+ }
llava/lib/tcl8.6/msgs/nl.msg ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # created by tools/loadICU.tcl -- do not edit
2
+ namespace eval ::tcl::clock {
3
+ ::msgcat::mcset nl DAYS_OF_WEEK_ABBREV [list \
4
+ "zo"\
5
+ "ma"\
6
+ "di"\
7
+ "wo"\
8
+ "do"\
9
+ "vr"\
10
+ "za"]
11
+ ::msgcat::mcset nl DAYS_OF_WEEK_FULL [list \
12
+ "zondag"\
13
+ "maandag"\
14
+ "dinsdag"\
15
+ "woensdag"\
16
+ "donderdag"\
17
+ "vrijdag"\
18
+ "zaterdag"]
19
+ ::msgcat::mcset nl MONTHS_ABBREV [list \
20
+ "jan"\
21
+ "feb"\
22
+ "mrt"\
23
+ "apr"\
24
+ "mei"\
25
+ "jun"\
26
+ "jul"\
27
+ "aug"\
28
+ "sep"\
29
+ "okt"\
30
+ "nov"\
31
+ "dec"\
32
+ ""]
33
+ ::msgcat::mcset nl MONTHS_FULL [list \
34
+ "januari"\
35
+ "februari"\
36
+ "maart"\
37
+ "april"\
38
+ "mei"\
39
+ "juni"\
40
+ "juli"\
41
+ "augustus"\
42
+ "september"\
43
+ "oktober"\
44
+ "november"\
45
+ "december"\
46
+ ""]
47
+ ::msgcat::mcset nl DATE_FORMAT "%e %B %Y"
48
+ ::msgcat::mcset nl TIME_FORMAT "%k:%M:%S"
49
+ ::msgcat::mcset nl DATE_TIME_FORMAT "%e %B %Y %k:%M:%S %z"
50
+ }
llava/lib/tcl8.6/msgs/nl_be.msg ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ # created by tools/loadICU.tcl -- do not edit
2
+ namespace eval ::tcl::clock {
3
+ ::msgcat::mcset nl_BE DATE_FORMAT "%d-%m-%y"
4
+ ::msgcat::mcset nl_BE TIME_FORMAT "%T"
5
+ ::msgcat::mcset nl_BE TIME_FORMAT_12 "%T"
6
+ ::msgcat::mcset nl_BE DATE_TIME_FORMAT "%a %d %b %Y %T %z"
7
+ }
llava/lib/tcl8.6/msgs/nn.msg ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # created by tools/loadICU.tcl -- do not edit
2
+ namespace eval ::tcl::clock {
3
+ ::msgcat::mcset nn DAYS_OF_WEEK_ABBREV [list \
4
+ "su"\
5
+ "m\u00e5"\
6
+ "ty"\
7
+ "on"\
8
+ "to"\
9
+ "fr"\
10
+ "lau"]
11
+ ::msgcat::mcset nn DAYS_OF_WEEK_FULL [list \
12
+ "sundag"\
13
+ "m\u00e5ndag"\
14
+ "tysdag"\
15
+ "onsdag"\
16
+ "torsdag"\
17
+ "fredag"\
18
+ "laurdag"]
19
+ ::msgcat::mcset nn MONTHS_ABBREV [list \
20
+ "jan"\
21
+ "feb"\
22
+ "mar"\
23
+ "apr"\
24
+ "mai"\
25
+ "jun"\
26
+ "jul"\
27
+ "aug"\
28
+ "sep"\
29
+ "okt"\
30
+ "nov"\
31
+ "des"\
32
+ ""]
33
+ ::msgcat::mcset nn MONTHS_FULL [list \
34
+ "januar"\
35
+ "februar"\
36
+ "mars"\
37
+ "april"\
38
+ "mai"\
39
+ "juni"\
40
+ "juli"\
41
+ "august"\
42
+ "september"\
43
+ "oktober"\
44
+ "november"\
45
+ "desember"\
46
+ ""]
47
+ ::msgcat::mcset nn BCE "f.Kr."
48
+ ::msgcat::mcset nn CE "e.Kr."
49
+ ::msgcat::mcset nn DATE_FORMAT "%e. %B %Y"
50
+ ::msgcat::mcset nn TIME_FORMAT "%H:%M:%S"
51
+ ::msgcat::mcset nn DATE_TIME_FORMAT "%e. %B %Y %H:%M:%S %z"
52
+ }
llava/lib/tcl8.6/msgs/pt.msg ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # created by tools/loadICU.tcl -- do not edit
2
+ namespace eval ::tcl::clock {
3
+ ::msgcat::mcset pt DAYS_OF_WEEK_ABBREV [list \
4
+ "Dom"\
5
+ "Seg"\
6
+ "Ter"\
7
+ "Qua"\
8
+ "Qui"\
9
+ "Sex"\
10
+ "S\u00e1b"]
11
+ ::msgcat::mcset pt DAYS_OF_WEEK_FULL [list \
12
+ "Domingo"\
13
+ "Segunda-feira"\
14
+ "Ter\u00e7a-feira"\
15
+ "Quarta-feira"\
16
+ "Quinta-feira"\
17
+ "Sexta-feira"\
18
+ "S\u00e1bado"]
19
+ ::msgcat::mcset pt MONTHS_ABBREV [list \
20
+ "Jan"\
21
+ "Fev"\
22
+ "Mar"\
23
+ "Abr"\
24
+ "Mai"\
25
+ "Jun"\
26
+ "Jul"\
27
+ "Ago"\
28
+ "Set"\
29
+ "Out"\
30
+ "Nov"\
31
+ "Dez"\
32
+ ""]
33
+ ::msgcat::mcset pt MONTHS_FULL [list \
34
+ "Janeiro"\
35
+ "Fevereiro"\
36
+ "Mar\u00e7o"\
37
+ "Abril"\
38
+ "Maio"\
39
+ "Junho"\
40
+ "Julho"\
41
+ "Agosto"\
42
+ "Setembro"\
43
+ "Outubro"\
44
+ "Novembro"\
45
+ "Dezembro"\
46
+ ""]
47
+ ::msgcat::mcset pt DATE_FORMAT "%d-%m-%Y"
48
+ ::msgcat::mcset pt TIME_FORMAT "%k:%M:%S"
49
+ ::msgcat::mcset pt DATE_TIME_FORMAT "%d-%m-%Y %k:%M:%S %z"
50
+ }
llava/lib/tcl8.6/msgs/ro.msg ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # created by tools/loadICU.tcl -- do not edit
2
+ namespace eval ::tcl::clock {
3
+ ::msgcat::mcset ro DAYS_OF_WEEK_ABBREV [list \
4
+ "D"\
5
+ "L"\
6
+ "Ma"\
7
+ "Mi"\
8
+ "J"\
9
+ "V"\
10
+ "S"]
11
+ ::msgcat::mcset ro DAYS_OF_WEEK_FULL [list \
12
+ "duminic\u0103"\
13
+ "luni"\
14
+ "mar\u0163i"\
15
+ "miercuri"\
16
+ "joi"\
17
+ "vineri"\
18
+ "s\u00eemb\u0103t\u0103"]
19
+ ::msgcat::mcset ro MONTHS_ABBREV [list \
20
+ "Ian"\
21
+ "Feb"\
22
+ "Mar"\
23
+ "Apr"\
24
+ "Mai"\
25
+ "Iun"\
26
+ "Iul"\
27
+ "Aug"\
28
+ "Sep"\
29
+ "Oct"\
30
+ "Nov"\
31
+ "Dec"\
32
+ ""]
33
+ ::msgcat::mcset ro MONTHS_FULL [list \
34
+ "ianuarie"\
35
+ "februarie"\
36
+ "martie"\
37
+ "aprilie"\
38
+ "mai"\
39
+ "iunie"\
40
+ "iulie"\
41
+ "august"\
42
+ "septembrie"\
43
+ "octombrie"\
44
+ "noiembrie"\
45
+ "decembrie"\
46
+ ""]
47
+ ::msgcat::mcset ro BCE "d.C."
48
+ ::msgcat::mcset ro CE "\u00ee.d.C."
49
+ ::msgcat::mcset ro DATE_FORMAT "%d.%m.%Y"
50
+ ::msgcat::mcset ro TIME_FORMAT "%H:%M:%S"
51
+ ::msgcat::mcset ro DATE_TIME_FORMAT "%d.%m.%Y %H:%M:%S %z"
52
+ }
llava/lib/tcl8.6/msgs/ru_ua.msg ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ # created by tools/loadICU.tcl -- do not edit
2
+ namespace eval ::tcl::clock {
3
+ ::msgcat::mcset ru_UA DATE_FORMAT "%d.%m.%Y"
4
+ ::msgcat::mcset ru_UA TIME_FORMAT "%k:%M:%S"
5
+ ::msgcat::mcset ru_UA DATE_TIME_FORMAT "%d.%m.%Y %k:%M:%S %z"
6
+ }
llava/lib/tcl8.6/msgs/sw.msg ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # created by tools/loadICU.tcl -- do not edit
2
+ namespace eval ::tcl::clock {
3
+ ::msgcat::mcset sw DAYS_OF_WEEK_ABBREV [list \
4
+ "Jpi"\
5
+ "Jtt"\
6
+ "Jnn"\
7
+ "Jtn"\
8
+ "Alh"\
9
+ "Iju"\
10
+ "Jmo"]
11
+ ::msgcat::mcset sw DAYS_OF_WEEK_FULL [list \
12
+ "Jumapili"\
13
+ "Jumatatu"\
14
+ "Jumanne"\
15
+ "Jumatano"\
16
+ "Alhamisi"\
17
+ "Ijumaa"\
18
+ "Jumamosi"]
19
+ ::msgcat::mcset sw MONTHS_ABBREV [list \
20
+ "Jan"\
21
+ "Feb"\
22
+ "Mar"\
23
+ "Apr"\
24
+ "Mei"\
25
+ "Jun"\
26
+ "Jul"\
27
+ "Ago"\
28
+ "Sep"\
29
+ "Okt"\
30
+ "Nov"\
31
+ "Des"\
32
+ ""]
33
+ ::msgcat::mcset sw MONTHS_FULL [list \
34
+ "Januari"\
35
+ "Februari"\
36
+ "Machi"\
37
+ "Aprili"\
38
+ "Mei"\
39
+ "Juni"\
40
+ "Julai"\
41
+ "Agosti"\
42
+ "Septemba"\
43
+ "Oktoba"\
44
+ "Novemba"\
45
+ "Desemba"\
46
+ ""]
47
+ ::msgcat::mcset sw BCE "KK"
48
+ ::msgcat::mcset sw CE "BK"
49
+ }
llava/lib/tcl8.6/msgs/th.msg ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # created by tools/loadICU.tcl -- do not edit
2
+ namespace eval ::tcl::clock {
3
+ ::msgcat::mcset th DAYS_OF_WEEK_ABBREV [list \
4
+ "\u0e2d\u0e32."\
5
+ "\u0e08."\
6
+ "\u0e2d."\
7
+ "\u0e1e."\
8
+ "\u0e1e\u0e24."\
9
+ "\u0e28."\
10
+ "\u0e2a."]
11
+ ::msgcat::mcset th DAYS_OF_WEEK_FULL [list \
12
+ "\u0e27\u0e31\u0e19\u0e2d\u0e32\u0e17\u0e34\u0e15\u0e22\u0e4c"\
13
+ "\u0e27\u0e31\u0e19\u0e08\u0e31\u0e19\u0e17\u0e23\u0e4c"\
14
+ "\u0e27\u0e31\u0e19\u0e2d\u0e31\u0e07\u0e04\u0e32\u0e23"\
15
+ "\u0e27\u0e31\u0e19\u0e1e\u0e38\u0e18"\
16
+ "\u0e27\u0e31\u0e19\u0e1e\u0e24\u0e2b\u0e31\u0e2a\u0e1a\u0e14\u0e35"\
17
+ "\u0e27\u0e31\u0e19\u0e28\u0e38\u0e01\u0e23\u0e4c"\
18
+ "\u0e27\u0e31\u0e19\u0e40\u0e2a\u0e32\u0e23\u0e4c"]
19
+ ::msgcat::mcset th MONTHS_ABBREV [list \
20
+ "\u0e21.\u0e04."\
21
+ "\u0e01.\u0e1e."\
22
+ "\u0e21\u0e35.\u0e04."\
23
+ "\u0e40\u0e21.\u0e22."\
24
+ "\u0e1e.\u0e04."\
25
+ "\u0e21\u0e34.\u0e22."\
26
+ "\u0e01.\u0e04."\
27
+ "\u0e2a.\u0e04."\
28
+ "\u0e01.\u0e22."\
29
+ "\u0e15.\u0e04."\
30
+ "\u0e1e.\u0e22."\
31
+ "\u0e18.\u0e04."\
32
+ ""]
33
+ ::msgcat::mcset th MONTHS_FULL [list \
34
+ "\u0e21\u0e01\u0e23\u0e32\u0e04\u0e21"\
35
+ "\u0e01\u0e38\u0e21\u0e20\u0e32\u0e1e\u0e31\u0e19\u0e18\u0e4c"\
36
+ "\u0e21\u0e35\u0e19\u0e32\u0e04\u0e21"\
37
+ "\u0e40\u0e21\u0e29\u0e32\u0e22\u0e19"\
38
+ "\u0e1e\u0e24\u0e29\u0e20\u0e32\u0e04\u0e21"\
39
+ "\u0e21\u0e34\u0e16\u0e38\u0e19\u0e32\u0e22\u0e19"\
40
+ "\u0e01\u0e23\u0e01\u0e0e\u0e32\u0e04\u0e21"\
41
+ "\u0e2a\u0e34\u0e07\u0e2b\u0e32\u0e04\u0e21"\
42
+ "\u0e01\u0e31\u0e19\u0e22\u0e32\u0e22\u0e19"\
43
+ "\u0e15\u0e38\u0e25\u0e32\u0e04\u0e21"\
44
+ "\u0e1e\u0e24\u0e28\u0e08\u0e34\u0e01\u0e32\u0e22\u0e19"\
45
+ "\u0e18\u0e31\u0e19\u0e27\u0e32\u0e04\u0e21"\
46
+ ""]
47
+ ::msgcat::mcset th BCE "\u0e25\u0e17\u0e35\u0e48"
48
+ ::msgcat::mcset th CE "\u0e04.\u0e28."
49
+ ::msgcat::mcset th AM "\u0e01\u0e48\u0e2d\u0e19\u0e40\u0e17\u0e35\u0e48\u0e22\u0e07"
50
+ ::msgcat::mcset th PM "\u0e2b\u0e25\u0e31\u0e07\u0e40\u0e17\u0e35\u0e48\u0e22\u0e07"
51
+ ::msgcat::mcset th DATE_FORMAT "%e/%m/%Y"
52
+ ::msgcat::mcset th TIME_FORMAT "%k:%M:%S"
53
+ ::msgcat::mcset th DATE_TIME_FORMAT "%e/%m/%Y %k:%M:%S %z"
54
+ }
llava/lib/tcl8.6/msgs/zh_sg.msg ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ # created by tools/loadICU.tcl -- do not edit
2
+ namespace eval ::tcl::clock {
3
+ ::msgcat::mcset zh_SG AM "\u4e0a\u5348"
4
+ ::msgcat::mcset zh_SG PM "\u4e2d\u5348"
5
+ ::msgcat::mcset zh_SG DATE_FORMAT "%d %B %Y"
6
+ ::msgcat::mcset zh_SG TIME_FORMAT_12 "%P %I:%M:%S"
7
+ ::msgcat::mcset zh_SG DATE_TIME_FORMAT "%d %B %Y %P %I:%M:%S %z"
8
+ }
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_foreach_floor_cuda_dispatch.h ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+ // @generated by torchgen/gen.py from DispatchKeyFunction.h
3
+
4
+ // NB: The implementing C++ file is RegisterDispatchKey.cpp
5
+
6
+ // The only #includes we need are for custom classes that have defaults in the C++ API
7
+ #include <c10/core/MemoryFormat.h>
8
+ #include <c10/core/Scalar.h>
9
+ #include <ATen/core/Reduction.h>
10
+
11
+ // Forward declarations of any types needed in the operator signatures.
12
+ // We can't directly include these classes because it will cause circular include dependencies.
13
+ // This file is included by TensorBody.h, which defines the Tensor class.
14
+ #include <ATen/core/ATen_fwd.h>
15
+
16
+ namespace at {
17
+
18
+ namespace cuda {
19
+
20
+ TORCH_API ::std::vector<at::Tensor> _foreach_floor(at::TensorList self);
21
+ TORCH_API void _foreach_floor_(at::TensorList self);
22
+
23
+ } // namespace cuda
24
+ } // namespace at
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_foreach_log2.h ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from Function.h
4
+
5
+ #include <ATen/Context.h>
6
+ #include <ATen/DeviceGuard.h>
7
+ #include <ATen/TensorUtils.h>
8
+ #include <ATen/TracerMode.h>
9
+ #include <ATen/core/Generator.h>
10
+ #include <ATen/core/Reduction.h>
11
+ #include <ATen/core/Tensor.h>
12
+ #include <c10/core/Scalar.h>
13
+ #include <c10/core/Storage.h>
14
+ #include <c10/core/TensorOptions.h>
15
+ #include <c10/util/Deprecated.h>
16
+ #include <c10/util/Optional.h>
17
+
18
+
19
+
20
+ #include <ATen/ops/_foreach_log2_ops.h>
21
+
22
+ namespace at {
23
+
24
+
25
+ // aten::_foreach_log2(Tensor[] self) -> Tensor[]
26
+ inline ::std::vector<at::Tensor> _foreach_log2(at::TensorList self) {
27
+ return at::_ops::_foreach_log2::call(self);
28
+ }
29
+
30
+ // aten::_foreach_log2_(Tensor(a!)[] self) -> ()
31
+ inline void _foreach_log2_(at::TensorList self) {
32
+ return at::_ops::_foreach_log2_::call(self);
33
+ }
34
+
35
+ // aten::_foreach_log2.out(Tensor[] self, *, Tensor(a!)[] out) -> ()
36
+ inline void _foreach_log2_out(at::TensorList out, at::TensorList self) {
37
+ return at::_ops::_foreach_log2_out::call(self, out);
38
+ }
39
+ // aten::_foreach_log2.out(Tensor[] self, *, Tensor(a!)[] out) -> ()
40
+ inline void _foreach_log2_outf(at::TensorList self, at::TensorList out) {
41
+ return at::_ops::_foreach_log2_out::call(self, out);
42
+ }
43
+
44
+ }
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_foreach_neg_cuda_dispatch.h ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+ // @generated by torchgen/gen.py from DispatchKeyFunction.h
3
+
4
+ // NB: The implementing C++ file is RegisterDispatchKey.cpp
5
+
6
+ // The only #includes we need are for custom classes that have defaults in the C++ API
7
+ #include <c10/core/MemoryFormat.h>
8
+ #include <c10/core/Scalar.h>
9
+ #include <ATen/core/Reduction.h>
10
+
11
+ // Forward declarations of any types needed in the operator signatures.
12
+ // We can't directly include these classes because it will cause circular include dependencies.
13
+ // This file is included by TensorBody.h, which defines the Tensor class.
14
+ #include <ATen/core/ATen_fwd.h>
15
+
16
+ namespace at {
17
+
18
+ namespace cuda {
19
+
20
+ TORCH_API ::std::vector<at::Tensor> _foreach_neg(at::TensorList self);
21
+ TORCH_API void _foreach_neg_(at::TensorList self);
22
+
23
+ } // namespace cuda
24
+ } // namespace at
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_lazy_clone_ops.h ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from Operator.h
4
+
5
+ #include <tuple>
6
+ #include <vector>
7
+
8
+ // Forward declarations of any types needed in the operator signatures.
9
+ // We can't directly include these classes because it will cause circular include dependencies.
10
+ // This file is included by TensorBody.h, which defines the Tensor class.
11
+ #include <ATen/core/ATen_fwd.h>
12
+
13
+ namespace at {
14
+ namespace _ops {
15
+
16
+
17
+ struct TORCH_API _lazy_clone {
18
+ using schema = at::Tensor (const at::Tensor &);
19
+ using ptr_schema = schema*;
20
+ // See Note [static constexpr char* members for windows NVCC]
21
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(name, "aten::_lazy_clone")
22
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(overload_name, "")
23
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(schema_str, "_lazy_clone(Tensor self) -> Tensor")
24
+ static at::Tensor call(const at::Tensor & self);
25
+ static at::Tensor redispatch(c10::DispatchKeySet dispatchKeySet, const at::Tensor & self);
26
+ };
27
+
28
+ }} // namespace at::_ops
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_nnpack_spatial_convolution.h ADDED
@@ -0,0 +1,91 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from Function.h
4
+
5
+ #include <ATen/Context.h>
6
+ #include <ATen/DeviceGuard.h>
7
+ #include <ATen/TensorUtils.h>
8
+ #include <ATen/TracerMode.h>
9
+ #include <ATen/core/Generator.h>
10
+ #include <ATen/core/Reduction.h>
11
+ #include <ATen/core/Tensor.h>
12
+ #include <c10/core/Scalar.h>
13
+ #include <c10/core/Storage.h>
14
+ #include <c10/core/TensorOptions.h>
15
+ #include <c10/util/Deprecated.h>
16
+ #include <c10/util/Optional.h>
17
+
18
+
19
+
20
+ #include <ATen/ops/_nnpack_spatial_convolution_ops.h>
21
+
22
+ namespace at {
23
+
24
+
25
+ // aten::_nnpack_spatial_convolution(Tensor input, Tensor weight, Tensor? bias, SymInt[2] padding, SymInt[2] stride=1) -> Tensor
26
+ inline at::Tensor _nnpack_spatial_convolution(const at::Tensor & input, const at::Tensor & weight, const ::std::optional<at::Tensor> & bias, at::IntArrayRef padding, at::IntArrayRef stride=1) {
27
+ return at::_ops::_nnpack_spatial_convolution::call(input, weight, bias, c10::fromIntArrayRefSlow(padding), c10::fromIntArrayRefSlow(stride));
28
+ }
29
+ namespace symint {
30
+ template <typename T, typename = std::enable_if_t<std::is_same<T, int64_t>::value>>
31
+ at::Tensor _nnpack_spatial_convolution(const at::Tensor & input, const at::Tensor & weight, const ::std::optional<at::Tensor> & bias, at::IntArrayRef padding, at::IntArrayRef stride=1) {
32
+ return at::_ops::_nnpack_spatial_convolution::call(input, weight, bias, c10::fromIntArrayRefSlow(padding), c10::fromIntArrayRefSlow(stride));
33
+ }
34
+ }
35
+
36
+ // aten::_nnpack_spatial_convolution(Tensor input, Tensor weight, Tensor? bias, SymInt[2] padding, SymInt[2] stride=1) -> Tensor
37
+ inline at::Tensor _nnpack_spatial_convolution_symint(const at::Tensor & input, const at::Tensor & weight, const ::std::optional<at::Tensor> & bias, c10::SymIntArrayRef padding, c10::SymIntArrayRef stride=c10::SymInt(1)) {
38
+ return at::_ops::_nnpack_spatial_convolution::call(input, weight, bias, padding, stride);
39
+ }
40
+ namespace symint {
41
+ template <typename T, typename = std::enable_if_t<std::is_same<T, c10::SymInt>::value>>
42
+ at::Tensor _nnpack_spatial_convolution(const at::Tensor & input, const at::Tensor & weight, const ::std::optional<at::Tensor> & bias, c10::SymIntArrayRef padding, c10::SymIntArrayRef stride=c10::SymInt(1)) {
43
+ return at::_ops::_nnpack_spatial_convolution::call(input, weight, bias, padding, stride);
44
+ }
45
+ }
46
+
47
+ // aten::_nnpack_spatial_convolution.out(Tensor input, Tensor weight, Tensor? bias, SymInt[2] padding, SymInt[2] stride=1, *, Tensor(a!) out) -> Tensor(a!)
48
+ inline at::Tensor & _nnpack_spatial_convolution_out(at::Tensor & out, const at::Tensor & input, const at::Tensor & weight, const ::std::optional<at::Tensor> & bias, at::IntArrayRef padding, at::IntArrayRef stride=1) {
49
+ return at::_ops::_nnpack_spatial_convolution_out::call(input, weight, bias, c10::fromIntArrayRefSlow(padding), c10::fromIntArrayRefSlow(stride), out);
50
+ }
51
+ namespace symint {
52
+ template <typename T, typename = std::enable_if_t<std::is_same<T, int64_t>::value>>
53
+ at::Tensor & _nnpack_spatial_convolution_out(at::Tensor & out, const at::Tensor & input, const at::Tensor & weight, const ::std::optional<at::Tensor> & bias, at::IntArrayRef padding, at::IntArrayRef stride=1) {
54
+ return at::_ops::_nnpack_spatial_convolution_out::call(input, weight, bias, c10::fromIntArrayRefSlow(padding), c10::fromIntArrayRefSlow(stride), out);
55
+ }
56
+ }
57
+
58
+ // aten::_nnpack_spatial_convolution.out(Tensor input, Tensor weight, Tensor? bias, SymInt[2] padding, SymInt[2] stride=1, *, Tensor(a!) out) -> Tensor(a!)
59
+ inline at::Tensor & _nnpack_spatial_convolution_outf(const at::Tensor & input, const at::Tensor & weight, const ::std::optional<at::Tensor> & bias, at::IntArrayRef padding, at::IntArrayRef stride, at::Tensor & out) {
60
+ return at::_ops::_nnpack_spatial_convolution_out::call(input, weight, bias, c10::fromIntArrayRefSlow(padding), c10::fromIntArrayRefSlow(stride), out);
61
+ }
62
+ namespace symint {
63
+ template <typename T, typename = std::enable_if_t<std::is_same<T, int64_t>::value>>
64
+ at::Tensor & _nnpack_spatial_convolution_outf(const at::Tensor & input, const at::Tensor & weight, const ::std::optional<at::Tensor> & bias, at::IntArrayRef padding, at::IntArrayRef stride, at::Tensor & out) {
65
+ return at::_ops::_nnpack_spatial_convolution_out::call(input, weight, bias, c10::fromIntArrayRefSlow(padding), c10::fromIntArrayRefSlow(stride), out);
66
+ }
67
+ }
68
+
69
+ // aten::_nnpack_spatial_convolution.out(Tensor input, Tensor weight, Tensor? bias, SymInt[2] padding, SymInt[2] stride=1, *, Tensor(a!) out) -> Tensor(a!)
70
+ inline at::Tensor & _nnpack_spatial_convolution_symint_out(at::Tensor & out, const at::Tensor & input, const at::Tensor & weight, const ::std::optional<at::Tensor> & bias, c10::SymIntArrayRef padding, c10::SymIntArrayRef stride=c10::SymInt(1)) {
71
+ return at::_ops::_nnpack_spatial_convolution_out::call(input, weight, bias, padding, stride, out);
72
+ }
73
+ namespace symint {
74
+ template <typename T, typename = std::enable_if_t<std::is_same<T, c10::SymInt>::value>>
75
+ at::Tensor & _nnpack_spatial_convolution_out(at::Tensor & out, const at::Tensor & input, const at::Tensor & weight, const ::std::optional<at::Tensor> & bias, c10::SymIntArrayRef padding, c10::SymIntArrayRef stride=c10::SymInt(1)) {
76
+ return at::_ops::_nnpack_spatial_convolution_out::call(input, weight, bias, padding, stride, out);
77
+ }
78
+ }
79
+
80
+ // aten::_nnpack_spatial_convolution.out(Tensor input, Tensor weight, Tensor? bias, SymInt[2] padding, SymInt[2] stride=1, *, Tensor(a!) out) -> Tensor(a!)
81
+ inline at::Tensor & _nnpack_spatial_convolution_symint_outf(const at::Tensor & input, const at::Tensor & weight, const ::std::optional<at::Tensor> & bias, c10::SymIntArrayRef padding, c10::SymIntArrayRef stride, at::Tensor & out) {
82
+ return at::_ops::_nnpack_spatial_convolution_out::call(input, weight, bias, padding, stride, out);
83
+ }
84
+ namespace symint {
85
+ template <typename T, typename = std::enable_if_t<std::is_same<T, c10::SymInt>::value>>
86
+ at::Tensor & _nnpack_spatial_convolution_outf(const at::Tensor & input, const at::Tensor & weight, const ::std::optional<at::Tensor> & bias, c10::SymIntArrayRef padding, c10::SymIntArrayRef stride, at::Tensor & out) {
87
+ return at::_ops::_nnpack_spatial_convolution_out::call(input, weight, bias, padding, stride, out);
88
+ }
89
+ }
90
+
91
+ }
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_prelu_kernel_backward_native.h ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from NativeFunction.h
4
+
5
+ #include <c10/core/Scalar.h>
6
+ #include <c10/core/Storage.h>
7
+ #include <c10/core/TensorOptions.h>
8
+ #include <c10/util/Deprecated.h>
9
+ #include <c10/util/Optional.h>
10
+ #include <c10/core/QScheme.h>
11
+ #include <ATen/core/Reduction.h>
12
+ #include <ATen/core/Tensor.h>
13
+ #include <tuple>
14
+ #include <vector>
15
+
16
+
17
+ namespace at {
18
+ namespace native {
19
+ TORCH_API ::std::tuple<at::Tensor,at::Tensor> _prelu_kernel_backward(const at::Tensor & grad_output, const at::Tensor & self, const at::Tensor & weight);
20
+ TORCH_API ::std::tuple<at::Tensor,at::Tensor> mkldnn_prelu_backward(const at::Tensor & grad_output, const at::Tensor & self, const at::Tensor & weight);
21
+ } // namespace native
22
+ } // namespace at
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_sparse_semi_structured_apply_native.h ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from NativeFunction.h
4
+
5
+ #include <c10/core/Scalar.h>
6
+ #include <c10/core/Storage.h>
7
+ #include <c10/core/TensorOptions.h>
8
+ #include <c10/util/Deprecated.h>
9
+ #include <c10/util/Optional.h>
10
+ #include <c10/core/QScheme.h>
11
+ #include <ATen/core/Reduction.h>
12
+ #include <ATen/core/Tensor.h>
13
+ #include <tuple>
14
+ #include <vector>
15
+
16
+
17
+ namespace at {
18
+ namespace native {
19
+ TORCH_API ::std::tuple<at::Tensor,at::Tensor> _sparse_semi_structured_apply(const at::Tensor & input, const at::Tensor & thread_masks);
20
+ } // namespace native
21
+ } // namespace at
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_unsafe_index_put_compositeexplicitautograd_dispatch.h ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+ // @generated by torchgen/gen.py from DispatchKeyFunction.h
3
+
4
+ // NB: The implementing C++ file is RegisterDispatchKey.cpp
5
+
6
+ // The only #includes we need are for custom classes that have defaults in the C++ API
7
+ #include <c10/core/MemoryFormat.h>
8
+ #include <c10/core/Scalar.h>
9
+ #include <ATen/core/Reduction.h>
10
+
11
+ // Forward declarations of any types needed in the operator signatures.
12
+ // We can't directly include these classes because it will cause circular include dependencies.
13
+ // This file is included by TensorBody.h, which defines the Tensor class.
14
+ #include <ATen/core/ATen_fwd.h>
15
+
16
+ namespace at {
17
+
18
+ namespace compositeexplicitautograd {
19
+
20
+ TORCH_API at::Tensor _unsafe_index_put(const at::Tensor & self, const c10::List<::std::optional<at::Tensor>> & indices, const at::Tensor & values, bool accumulate=false);
21
+
22
+ } // namespace compositeexplicitautograd
23
+ } // namespace at
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_upsample_bicubic2d_aa_backward_meta_dispatch.h ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+ // @generated by torchgen/gen.py from DispatchKeyFunction.h
3
+
4
+ // NB: The implementing C++ file is RegisterDispatchKey.cpp
5
+
6
+ // The only #includes we need are for custom classes that have defaults in the C++ API
7
+ #include <c10/core/MemoryFormat.h>
8
+ #include <c10/core/Scalar.h>
9
+ #include <ATen/core/Reduction.h>
10
+
11
+ // Forward declarations of any types needed in the operator signatures.
12
+ // We can't directly include these classes because it will cause circular include dependencies.
13
+ // This file is included by TensorBody.h, which defines the Tensor class.
14
+ #include <ATen/core/ATen_fwd.h>
15
+
16
+ namespace at {
17
+
18
+ namespace meta {
19
+
20
+ TORCH_API at::Tensor _upsample_bicubic2d_aa_backward(const at::Tensor & grad_output, at::IntArrayRef output_size, at::IntArrayRef input_size, bool align_corners, ::std::optional<double> scales_h=::std::nullopt, ::std::optional<double> scales_w=::std::nullopt);
21
+ TORCH_API at::Tensor _upsample_bicubic2d_aa_backward_symint(const at::Tensor & grad_output, c10::SymIntArrayRef output_size, c10::SymIntArrayRef input_size, bool align_corners, ::std::optional<double> scales_h=::std::nullopt, ::std::optional<double> scales_w=::std::nullopt);
22
+ TORCH_API at::Tensor & _upsample_bicubic2d_aa_backward_out(at::Tensor & grad_input, const at::Tensor & grad_output, at::IntArrayRef output_size, at::IntArrayRef input_size, bool align_corners, ::std::optional<double> scales_h=::std::nullopt, ::std::optional<double> scales_w=::std::nullopt);
23
+ TORCH_API at::Tensor & _upsample_bicubic2d_aa_backward_outf(const at::Tensor & grad_output, at::IntArrayRef output_size, at::IntArrayRef input_size, bool align_corners, ::std::optional<double> scales_h, ::std::optional<double> scales_w, at::Tensor & grad_input);
24
+ TORCH_API at::Tensor & _upsample_bicubic2d_aa_backward_symint_out(at::Tensor & grad_input, const at::Tensor & grad_output, c10::SymIntArrayRef output_size, c10::SymIntArrayRef input_size, bool align_corners, ::std::optional<double> scales_h=::std::nullopt, ::std::optional<double> scales_w=::std::nullopt);
25
+ TORCH_API at::Tensor & _upsample_bicubic2d_aa_backward_symint_outf(const at::Tensor & grad_output, c10::SymIntArrayRef output_size, c10::SymIntArrayRef input_size, bool align_corners, ::std::optional<double> scales_h, ::std::optional<double> scales_w, at::Tensor & grad_input);
26
+
27
+ } // namespace meta
28
+ } // namespace at
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_upsample_bilinear2d_aa_meta.h ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from NativeMetaFunction.h
4
+
5
+ #include <c10/core/Scalar.h>
6
+ #include <c10/core/Storage.h>
7
+ #include <c10/core/TensorOptions.h>
8
+ #include <c10/util/Deprecated.h>
9
+ #include <c10/util/Optional.h>
10
+ #include <c10/core/QScheme.h>
11
+ #include <ATen/core/Reduction.h>
12
+ #include <ATen/TensorIterator.h>
13
+ #include <ATen/TensorMeta.h>
14
+ #include <tuple>
15
+ #include <vector>
16
+
17
+ namespace at {
18
+ namespace meta {
19
+
20
+ struct TORCH_API structured__upsample_bilinear2d_aa : public at::impl::MetaBase {
21
+
22
+
23
+ void meta(const at::Tensor & self, at::ArrayRef<int64_t> output_size, bool align_corners, ::std::optional<double> scales_h, ::std::optional<double> scales_w);
24
+ };
25
+
26
+ } // namespace native
27
+ } // namespace at
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/adaptive_avg_pool3d.h ADDED
@@ -0,0 +1,91 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from Function.h
4
+
5
+ #include <ATen/Context.h>
6
+ #include <ATen/DeviceGuard.h>
7
+ #include <ATen/TensorUtils.h>
8
+ #include <ATen/TracerMode.h>
9
+ #include <ATen/core/Generator.h>
10
+ #include <ATen/core/Reduction.h>
11
+ #include <ATen/core/Tensor.h>
12
+ #include <c10/core/Scalar.h>
13
+ #include <c10/core/Storage.h>
14
+ #include <c10/core/TensorOptions.h>
15
+ #include <c10/util/Deprecated.h>
16
+ #include <c10/util/Optional.h>
17
+
18
+
19
+
20
+ #include <ATen/ops/adaptive_avg_pool3d_ops.h>
21
+
22
+ namespace at {
23
+
24
+
25
+ // aten::adaptive_avg_pool3d.out(Tensor self, SymInt[3] output_size, *, Tensor(a!) out) -> Tensor(a!)
26
+ inline at::Tensor & adaptive_avg_pool3d_out(at::Tensor & out, const at::Tensor & self, at::IntArrayRef output_size) {
27
+ return at::_ops::adaptive_avg_pool3d_out::call(self, c10::fromIntArrayRefSlow(output_size), out);
28
+ }
29
+ namespace symint {
30
+ template <typename T, typename = std::enable_if_t<std::is_same<T, int64_t>::value>>
31
+ at::Tensor & adaptive_avg_pool3d_out(at::Tensor & out, const at::Tensor & self, at::IntArrayRef output_size) {
32
+ return at::_ops::adaptive_avg_pool3d_out::call(self, c10::fromIntArrayRefSlow(output_size), out);
33
+ }
34
+ }
35
+
36
+ // aten::adaptive_avg_pool3d.out(Tensor self, SymInt[3] output_size, *, Tensor(a!) out) -> Tensor(a!)
37
+ inline at::Tensor & adaptive_avg_pool3d_outf(const at::Tensor & self, at::IntArrayRef output_size, at::Tensor & out) {
38
+ return at::_ops::adaptive_avg_pool3d_out::call(self, c10::fromIntArrayRefSlow(output_size), out);
39
+ }
40
+ namespace symint {
41
+ template <typename T, typename = std::enable_if_t<std::is_same<T, int64_t>::value>>
42
+ at::Tensor & adaptive_avg_pool3d_outf(const at::Tensor & self, at::IntArrayRef output_size, at::Tensor & out) {
43
+ return at::_ops::adaptive_avg_pool3d_out::call(self, c10::fromIntArrayRefSlow(output_size), out);
44
+ }
45
+ }
46
+
47
+ // aten::adaptive_avg_pool3d.out(Tensor self, SymInt[3] output_size, *, Tensor(a!) out) -> Tensor(a!)
48
+ inline at::Tensor & adaptive_avg_pool3d_symint_out(at::Tensor & out, const at::Tensor & self, c10::SymIntArrayRef output_size) {
49
+ return at::_ops::adaptive_avg_pool3d_out::call(self, output_size, out);
50
+ }
51
+ namespace symint {
52
+ template <typename T, typename = std::enable_if_t<std::is_same<T, c10::SymInt>::value>>
53
+ at::Tensor & adaptive_avg_pool3d_out(at::Tensor & out, const at::Tensor & self, c10::SymIntArrayRef output_size) {
54
+ return at::_ops::adaptive_avg_pool3d_out::call(self, output_size, out);
55
+ }
56
+ }
57
+
58
+ // aten::adaptive_avg_pool3d.out(Tensor self, SymInt[3] output_size, *, Tensor(a!) out) -> Tensor(a!)
59
+ inline at::Tensor & adaptive_avg_pool3d_symint_outf(const at::Tensor & self, c10::SymIntArrayRef output_size, at::Tensor & out) {
60
+ return at::_ops::adaptive_avg_pool3d_out::call(self, output_size, out);
61
+ }
62
+ namespace symint {
63
+ template <typename T, typename = std::enable_if_t<std::is_same<T, c10::SymInt>::value>>
64
+ at::Tensor & adaptive_avg_pool3d_outf(const at::Tensor & self, c10::SymIntArrayRef output_size, at::Tensor & out) {
65
+ return at::_ops::adaptive_avg_pool3d_out::call(self, output_size, out);
66
+ }
67
+ }
68
+
69
+ // aten::adaptive_avg_pool3d(Tensor self, SymInt[3] output_size) -> Tensor
70
+ inline at::Tensor adaptive_avg_pool3d(const at::Tensor & self, at::IntArrayRef output_size) {
71
+ return at::_ops::adaptive_avg_pool3d::call(self, c10::fromIntArrayRefSlow(output_size));
72
+ }
73
+ namespace symint {
74
+ template <typename T, typename = std::enable_if_t<std::is_same<T, int64_t>::value>>
75
+ at::Tensor adaptive_avg_pool3d(const at::Tensor & self, at::IntArrayRef output_size) {
76
+ return at::_ops::adaptive_avg_pool3d::call(self, c10::fromIntArrayRefSlow(output_size));
77
+ }
78
+ }
79
+
80
+ // aten::adaptive_avg_pool3d(Tensor self, SymInt[3] output_size) -> Tensor
81
+ inline at::Tensor adaptive_avg_pool3d_symint(const at::Tensor & self, c10::SymIntArrayRef output_size) {
82
+ return at::_ops::adaptive_avg_pool3d::call(self, output_size);
83
+ }
84
+ namespace symint {
85
+ template <typename T, typename = std::enable_if_t<std::is_same<T, c10::SymInt>::value>>
86
+ at::Tensor adaptive_avg_pool3d(const at::Tensor & self, c10::SymIntArrayRef output_size) {
87
+ return at::_ops::adaptive_avg_pool3d::call(self, output_size);
88
+ }
89
+ }
90
+
91
+ }
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/adaptive_max_pool2d_backward_meta.h ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from NativeMetaFunction.h
4
+
5
+ #include <c10/core/Scalar.h>
6
+ #include <c10/core/Storage.h>
7
+ #include <c10/core/TensorOptions.h>
8
+ #include <c10/util/Deprecated.h>
9
+ #include <c10/util/Optional.h>
10
+ #include <c10/core/QScheme.h>
11
+ #include <ATen/core/Reduction.h>
12
+ #include <ATen/TensorIterator.h>
13
+ #include <ATen/TensorMeta.h>
14
+ #include <tuple>
15
+ #include <vector>
16
+
17
+ namespace at {
18
+ namespace meta {
19
+
20
+ struct TORCH_API structured_adaptive_max_pool2d_backward : public at::impl::MetaBase {
21
+
22
+
23
+ void meta(const at::Tensor & grad_output, const at::Tensor & self, const at::Tensor & indices);
24
+ };
25
+
26
+ } // namespace native
27
+ } // namespace at
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/atan2_meta_dispatch.h ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+ // @generated by torchgen/gen.py from DispatchKeyFunction.h
3
+
4
+ // NB: The implementing C++ file is RegisterDispatchKey.cpp
5
+
6
+ // The only #includes we need are for custom classes that have defaults in the C++ API
7
+ #include <c10/core/MemoryFormat.h>
8
+ #include <c10/core/Scalar.h>
9
+ #include <ATen/core/Reduction.h>
10
+
11
+ // Forward declarations of any types needed in the operator signatures.
12
+ // We can't directly include these classes because it will cause circular include dependencies.
13
+ // This file is included by TensorBody.h, which defines the Tensor class.
14
+ #include <ATen/core/ATen_fwd.h>
15
+
16
+ namespace at {
17
+
18
+ namespace meta {
19
+
20
+ TORCH_API at::Tensor atan2(const at::Tensor & self, const at::Tensor & other);
21
+ TORCH_API at::Tensor & atan2_out(at::Tensor & out, const at::Tensor & self, const at::Tensor & other);
22
+ TORCH_API at::Tensor & atan2_outf(const at::Tensor & self, const at::Tensor & other, at::Tensor & out);
23
+ TORCH_API at::Tensor & atan2_(at::Tensor & self, const at::Tensor & other);
24
+
25
+ } // namespace meta
26
+ } // namespace at
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/batch_norm_gather_stats_compositeexplicitautograd_dispatch.h ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+ // @generated by torchgen/gen.py from DispatchKeyFunction.h
3
+
4
+ // NB: The implementing C++ file is RegisterDispatchKey.cpp
5
+
6
+ // The only #includes we need are for custom classes that have defaults in the C++ API
7
+ #include <c10/core/MemoryFormat.h>
8
+ #include <c10/core/Scalar.h>
9
+ #include <ATen/core/Reduction.h>
10
+
11
+ // Forward declarations of any types needed in the operator signatures.
12
+ // We can't directly include these classes because it will cause circular include dependencies.
13
+ // This file is included by TensorBody.h, which defines the Tensor class.
14
+ #include <ATen/core/ATen_fwd.h>
15
+
16
+ namespace at {
17
+
18
+ namespace compositeexplicitautograd {
19
+
20
+ TORCH_API ::std::tuple<at::Tensor &,at::Tensor &> batch_norm_gather_stats_out(at::Tensor & out0, at::Tensor & out1, const at::Tensor & input, const at::Tensor & mean, const at::Tensor & invstd, const ::std::optional<at::Tensor> & running_mean, const ::std::optional<at::Tensor> & running_var, double momentum, double eps, int64_t count);
21
+ TORCH_API ::std::tuple<at::Tensor &,at::Tensor &> batch_norm_gather_stats_outf(const at::Tensor & input, const at::Tensor & mean, const at::Tensor & invstd, const ::std::optional<at::Tensor> & running_mean, const ::std::optional<at::Tensor> & running_var, double momentum, double eps, int64_t count, at::Tensor & out0, at::Tensor & out1);
22
+
23
+ } // namespace compositeexplicitautograd
24
+ } // namespace at
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/batch_norm_gather_stats_with_counts_compositeexplicitautograd_dispatch.h ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+ // @generated by torchgen/gen.py from DispatchKeyFunction.h
3
+
4
+ // NB: The implementing C++ file is RegisterDispatchKey.cpp
5
+
6
+ // The only #includes we need are for custom classes that have defaults in the C++ API
7
+ #include <c10/core/MemoryFormat.h>
8
+ #include <c10/core/Scalar.h>
9
+ #include <ATen/core/Reduction.h>
10
+
11
+ // Forward declarations of any types needed in the operator signatures.
12
+ // We can't directly include these classes because it will cause circular include dependencies.
13
+ // This file is included by TensorBody.h, which defines the Tensor class.
14
+ #include <ATen/core/ATen_fwd.h>
15
+
16
+ namespace at {
17
+
18
+ namespace compositeexplicitautograd {
19
+
20
+ TORCH_API ::std::tuple<at::Tensor &,at::Tensor &> batch_norm_gather_stats_with_counts_out(at::Tensor & out0, at::Tensor & out1, const at::Tensor & input, const at::Tensor & mean, const at::Tensor & invstd, const ::std::optional<at::Tensor> & running_mean, const ::std::optional<at::Tensor> & running_var, double momentum, double eps, const at::Tensor & counts);
21
+ TORCH_API ::std::tuple<at::Tensor &,at::Tensor &> batch_norm_gather_stats_with_counts_outf(const at::Tensor & input, const at::Tensor & mean, const at::Tensor & invstd, const ::std::optional<at::Tensor> & running_mean, const ::std::optional<at::Tensor> & running_var, double momentum, double eps, const at::Tensor & counts, at::Tensor & out0, at::Tensor & out1);
22
+
23
+ } // namespace compositeexplicitautograd
24
+ } // namespace at
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/complex_cuda_dispatch.h ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+ // @generated by torchgen/gen.py from DispatchKeyFunction.h
3
+
4
+ // NB: The implementing C++ file is RegisterDispatchKey.cpp
5
+
6
+ // The only #includes we need are for custom classes that have defaults in the C++ API
7
+ #include <c10/core/MemoryFormat.h>
8
+ #include <c10/core/Scalar.h>
9
+ #include <ATen/core/Reduction.h>
10
+
11
+ // Forward declarations of any types needed in the operator signatures.
12
+ // We can't directly include these classes because it will cause circular include dependencies.
13
+ // This file is included by TensorBody.h, which defines the Tensor class.
14
+ #include <ATen/core/ATen_fwd.h>
15
+
16
+ namespace at {
17
+
18
+ namespace cuda {
19
+
20
+ TORCH_API at::Tensor & complex_out(at::Tensor & out, const at::Tensor & real, const at::Tensor & imag);
21
+ TORCH_API at::Tensor & complex_outf(const at::Tensor & real, const at::Tensor & imag, at::Tensor & out);
22
+
23
+ } // namespace cuda
24
+ } // namespace at
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/conv1d_ops.h ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from Operator.h
4
+
5
+ #include <tuple>
6
+ #include <vector>
7
+
8
+ // Forward declarations of any types needed in the operator signatures.
9
+ // We can't directly include these classes because it will cause circular include dependencies.
10
+ // This file is included by TensorBody.h, which defines the Tensor class.
11
+ #include <ATen/core/ATen_fwd.h>
12
+
13
+ namespace at {
14
+ namespace _ops {
15
+
16
+
17
+ struct TORCH_API conv1d {
18
+ using schema = at::Tensor (const at::Tensor &, const at::Tensor &, const ::std::optional<at::Tensor> &, c10::SymIntArrayRef, c10::SymIntArrayRef, c10::SymIntArrayRef, c10::SymInt);
19
+ using ptr_schema = schema*;
20
+ // See Note [static constexpr char* members for windows NVCC]
21
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(name, "aten::conv1d")
22
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(overload_name, "")
23
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(schema_str, "conv1d(Tensor input, Tensor weight, Tensor? bias=None, SymInt[1] stride=1, SymInt[1] padding=0, SymInt[1] dilation=1, SymInt groups=1) -> Tensor")
24
+ static at::Tensor call(const at::Tensor & input, const at::Tensor & weight, const ::std::optional<at::Tensor> & bias, c10::SymIntArrayRef stride, c10::SymIntArrayRef padding, c10::SymIntArrayRef dilation, c10::SymInt groups);
25
+ static at::Tensor redispatch(c10::DispatchKeySet dispatchKeySet, const at::Tensor & input, const at::Tensor & weight, const ::std::optional<at::Tensor> & bias, c10::SymIntArrayRef stride, c10::SymIntArrayRef padding, c10::SymIntArrayRef dilation, c10::SymInt groups);
26
+ };
27
+
28
+ struct TORCH_API conv1d_padding {
29
+ using schema = at::Tensor (const at::Tensor &, const at::Tensor &, const ::std::optional<at::Tensor> &, c10::SymIntArrayRef, c10::string_view, c10::SymIntArrayRef, c10::SymInt);
30
+ using ptr_schema = schema*;
31
+ // See Note [static constexpr char* members for windows NVCC]
32
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(name, "aten::conv1d")
33
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(overload_name, "padding")
34
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(schema_str, "conv1d.padding(Tensor input, Tensor weight, Tensor? bias=None, SymInt[1] stride=1, str padding=\"valid\", SymInt[1] dilation=1, SymInt groups=1) -> Tensor")
35
+ static at::Tensor call(const at::Tensor & input, const at::Tensor & weight, const ::std::optional<at::Tensor> & bias, c10::SymIntArrayRef stride, c10::string_view padding, c10::SymIntArrayRef dilation, c10::SymInt groups);
36
+ static at::Tensor redispatch(c10::DispatchKeySet dispatchKeySet, const at::Tensor & input, const at::Tensor & weight, const ::std::optional<at::Tensor> & bias, c10::SymIntArrayRef stride, c10::string_view padding, c10::SymIntArrayRef dilation, c10::SymInt groups);
37
+ };
38
+
39
+ }} // namespace at::_ops
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/cudnn_grid_sampler_ops.h ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from Operator.h
4
+
5
+ #include <tuple>
6
+ #include <vector>
7
+
8
+ // Forward declarations of any types needed in the operator signatures.
9
+ // We can't directly include these classes because it will cause circular include dependencies.
10
+ // This file is included by TensorBody.h, which defines the Tensor class.
11
+ #include <ATen/core/ATen_fwd.h>
12
+
13
+ namespace at {
14
+ namespace _ops {
15
+
16
+
17
+ struct TORCH_API cudnn_grid_sampler {
18
+ using schema = at::Tensor (const at::Tensor &, const at::Tensor &);
19
+ using ptr_schema = schema*;
20
+ // See Note [static constexpr char* members for windows NVCC]
21
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(name, "aten::cudnn_grid_sampler")
22
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(overload_name, "")
23
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(schema_str, "cudnn_grid_sampler(Tensor self, Tensor grid) -> Tensor output")
24
+ static at::Tensor call(const at::Tensor & self, const at::Tensor & grid);
25
+ static at::Tensor redispatch(c10::DispatchKeySet dispatchKeySet, const at::Tensor & self, const at::Tensor & grid);
26
+ };
27
+
28
+ struct TORCH_API cudnn_grid_sampler_out {
29
+ using schema = at::Tensor & (const at::Tensor &, const at::Tensor &, at::Tensor &);
30
+ using ptr_schema = schema*;
31
+ // See Note [static constexpr char* members for windows NVCC]
32
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(name, "aten::cudnn_grid_sampler")
33
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(overload_name, "out")
34
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(schema_str, "cudnn_grid_sampler.out(Tensor self, Tensor grid, *, Tensor(a!) out) -> Tensor(a!)")
35
+ static at::Tensor & call(const at::Tensor & self, const at::Tensor & grid, at::Tensor & out);
36
+ static at::Tensor & redispatch(c10::DispatchKeySet dispatchKeySet, const at::Tensor & self, const at::Tensor & grid, at::Tensor & out);
37
+ };
38
+
39
+ }} // namespace at::_ops
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/cummax.h ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from Function.h
4
+
5
+ #include <ATen/Context.h>
6
+ #include <ATen/DeviceGuard.h>
7
+ #include <ATen/TensorUtils.h>
8
+ #include <ATen/TracerMode.h>
9
+ #include <ATen/core/Generator.h>
10
+ #include <ATen/core/Reduction.h>
11
+ #include <ATen/core/Tensor.h>
12
+ #include <c10/core/Scalar.h>
13
+ #include <c10/core/Storage.h>
14
+ #include <c10/core/TensorOptions.h>
15
+ #include <c10/util/Deprecated.h>
16
+ #include <c10/util/Optional.h>
17
+
18
+
19
+
20
+ #include <ATen/ops/cummax_ops.h>
21
+
22
+ namespace at {
23
+
24
+
25
+ // aten::cummax(Tensor self, int dim) -> (Tensor values, Tensor indices)
26
+ inline ::std::tuple<at::Tensor,at::Tensor> cummax(const at::Tensor & self, int64_t dim) {
27
+ return at::_ops::cummax::call(self, dim);
28
+ }
29
+
30
+ // aten::cummax.out(Tensor self, int dim, *, Tensor(a!) values, Tensor(b!) indices) -> (Tensor(a!) values, Tensor(b!) indices)
31
+ inline ::std::tuple<at::Tensor &,at::Tensor &> cummax_out(at::Tensor & values, at::Tensor & indices, const at::Tensor & self, int64_t dim) {
32
+ return at::_ops::cummax_out::call(self, dim, values, indices);
33
+ }
34
+ // aten::cummax.out(Tensor self, int dim, *, Tensor(a!) values, Tensor(b!) indices) -> (Tensor(a!) values, Tensor(b!) indices)
35
+ inline ::std::tuple<at::Tensor &,at::Tensor &> cummax_outf(const at::Tensor & self, int64_t dim, at::Tensor & values, at::Tensor & indices) {
36
+ return at::_ops::cummax_out::call(self, dim, values, indices);
37
+ }
38
+
39
+ // aten::cummax.dimname(Tensor self, Dimname dim) -> (Tensor values, Tensor indices)
40
+ inline ::std::tuple<at::Tensor,at::Tensor> cummax(const at::Tensor & self, at::Dimname dim) {
41
+ return at::_ops::cummax_dimname::call(self, dim);
42
+ }
43
+
44
+ // aten::cummax.dimname_out(Tensor self, Dimname dim, *, Tensor(a!) values, Tensor(b!) indices) -> (Tensor(a!) values, Tensor(b!) indices)
45
+ inline ::std::tuple<at::Tensor &,at::Tensor &> cummax_out(at::Tensor & values, at::Tensor & indices, const at::Tensor & self, at::Dimname dim) {
46
+ return at::_ops::cummax_dimname_out::call(self, dim, values, indices);
47
+ }
48
+ // aten::cummax.dimname_out(Tensor self, Dimname dim, *, Tensor(a!) values, Tensor(b!) indices) -> (Tensor(a!) values, Tensor(b!) indices)
49
+ inline ::std::tuple<at::Tensor &,at::Tensor &> cummax_outf(const at::Tensor & self, at::Dimname dim, at::Tensor & values, at::Tensor & indices) {
50
+ return at::_ops::cummax_dimname_out::call(self, dim, values, indices);
51
+ }
52
+
53
+ }
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/diagonal_native.h ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from NativeFunction.h
4
+
5
+ #include <c10/core/Scalar.h>
6
+ #include <c10/core/Storage.h>
7
+ #include <c10/core/TensorOptions.h>
8
+ #include <c10/util/Deprecated.h>
9
+ #include <c10/util/Optional.h>
10
+ #include <c10/core/QScheme.h>
11
+ #include <ATen/core/Reduction.h>
12
+ #include <ATen/core/Tensor.h>
13
+ #include <tuple>
14
+ #include <vector>
15
+
16
+
17
+ namespace at {
18
+ namespace native {
19
+ TORCH_API at::Tensor diagonal(const at::Tensor & self, int64_t offset=0, int64_t dim1=0, int64_t dim2=1);
20
+ TORCH_API at::Tensor diagonal(const at::Tensor & self, at::Dimname outdim, at::Dimname dim1, at::Dimname dim2, int64_t offset=0);
21
+ } // namespace native
22
+ } // namespace at
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/diff_compositeimplicitautograd_dispatch.h ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+ // @generated by torchgen/gen.py from DispatchKeyFunction.h
3
+
4
+ // NB: The implementing C++ file is RegisterDispatchKey.cpp
5
+
6
+ // The only #includes we need are for custom classes that have defaults in the C++ API
7
+ #include <c10/core/MemoryFormat.h>
8
+ #include <c10/core/Scalar.h>
9
+ #include <ATen/core/Reduction.h>
10
+
11
+ // Forward declarations of any types needed in the operator signatures.
12
+ // We can't directly include these classes because it will cause circular include dependencies.
13
+ // This file is included by TensorBody.h, which defines the Tensor class.
14
+ #include <ATen/core/ATen_fwd.h>
15
+
16
+ namespace at {
17
+
18
+ namespace compositeimplicitautograd {
19
+
20
+ TORCH_API at::Tensor diff(const at::Tensor & self, int64_t n=1, int64_t dim=-1, const ::std::optional<at::Tensor> & prepend={}, const ::std::optional<at::Tensor> & append={});
21
+ TORCH_API at::Tensor & diff_out(at::Tensor & out, const at::Tensor & self, int64_t n=1, int64_t dim=-1, const ::std::optional<at::Tensor> & prepend={}, const ::std::optional<at::Tensor> & append={});
22
+ TORCH_API at::Tensor & diff_outf(const at::Tensor & self, int64_t n, int64_t dim, const ::std::optional<at::Tensor> & prepend, const ::std::optional<at::Tensor> & append, at::Tensor & out);
23
+
24
+ } // namespace compositeimplicitautograd
25
+ } // namespace at
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/embedding_dense_backward.h ADDED
@@ -0,0 +1,91 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from Function.h
4
+
5
+ #include <ATen/Context.h>
6
+ #include <ATen/DeviceGuard.h>
7
+ #include <ATen/TensorUtils.h>
8
+ #include <ATen/TracerMode.h>
9
+ #include <ATen/core/Generator.h>
10
+ #include <ATen/core/Reduction.h>
11
+ #include <ATen/core/Tensor.h>
12
+ #include <c10/core/Scalar.h>
13
+ #include <c10/core/Storage.h>
14
+ #include <c10/core/TensorOptions.h>
15
+ #include <c10/util/Deprecated.h>
16
+ #include <c10/util/Optional.h>
17
+
18
+
19
+
20
+ #include <ATen/ops/embedding_dense_backward_ops.h>
21
+
22
+ namespace at {
23
+
24
+
25
+ // aten::embedding_dense_backward(Tensor grad_output, Tensor indices, SymInt num_weights, SymInt padding_idx, bool scale_grad_by_freq) -> Tensor
26
+ inline at::Tensor embedding_dense_backward(const at::Tensor & grad_output, const at::Tensor & indices, int64_t num_weights, int64_t padding_idx, bool scale_grad_by_freq) {
27
+ return at::_ops::embedding_dense_backward::call(grad_output, indices, num_weights, padding_idx, scale_grad_by_freq);
28
+ }
29
+ namespace symint {
30
+ template <typename T, typename = std::enable_if_t<std::is_same<T, int64_t>::value>>
31
+ at::Tensor embedding_dense_backward(const at::Tensor & grad_output, const at::Tensor & indices, int64_t num_weights, int64_t padding_idx, bool scale_grad_by_freq) {
32
+ return at::_ops::embedding_dense_backward::call(grad_output, indices, num_weights, padding_idx, scale_grad_by_freq);
33
+ }
34
+ }
35
+
36
+ // aten::embedding_dense_backward(Tensor grad_output, Tensor indices, SymInt num_weights, SymInt padding_idx, bool scale_grad_by_freq) -> Tensor
37
+ inline at::Tensor embedding_dense_backward_symint(const at::Tensor & grad_output, const at::Tensor & indices, c10::SymInt num_weights, c10::SymInt padding_idx, bool scale_grad_by_freq) {
38
+ return at::_ops::embedding_dense_backward::call(grad_output, indices, num_weights, padding_idx, scale_grad_by_freq);
39
+ }
40
+ namespace symint {
41
+ template <typename T, typename = std::enable_if_t<std::is_same<T, c10::SymInt>::value>>
42
+ at::Tensor embedding_dense_backward(const at::Tensor & grad_output, const at::Tensor & indices, c10::SymInt num_weights, c10::SymInt padding_idx, bool scale_grad_by_freq) {
43
+ return at::_ops::embedding_dense_backward::call(grad_output, indices, num_weights, padding_idx, scale_grad_by_freq);
44
+ }
45
+ }
46
+
47
+ // aten::embedding_dense_backward.out(Tensor grad_output, Tensor indices, SymInt num_weights, SymInt padding_idx, bool scale_grad_by_freq, *, Tensor(a!) out) -> Tensor(a!)
48
+ inline at::Tensor & embedding_dense_backward_out(at::Tensor & out, const at::Tensor & grad_output, const at::Tensor & indices, int64_t num_weights, int64_t padding_idx, bool scale_grad_by_freq) {
49
+ return at::_ops::embedding_dense_backward_out::call(grad_output, indices, num_weights, padding_idx, scale_grad_by_freq, out);
50
+ }
51
+ namespace symint {
52
+ template <typename T, typename = std::enable_if_t<std::is_same<T, int64_t>::value>>
53
+ at::Tensor & embedding_dense_backward_out(at::Tensor & out, const at::Tensor & grad_output, const at::Tensor & indices, int64_t num_weights, int64_t padding_idx, bool scale_grad_by_freq) {
54
+ return at::_ops::embedding_dense_backward_out::call(grad_output, indices, num_weights, padding_idx, scale_grad_by_freq, out);
55
+ }
56
+ }
57
+
58
+ // aten::embedding_dense_backward.out(Tensor grad_output, Tensor indices, SymInt num_weights, SymInt padding_idx, bool scale_grad_by_freq, *, Tensor(a!) out) -> Tensor(a!)
59
+ inline at::Tensor & embedding_dense_backward_outf(const at::Tensor & grad_output, const at::Tensor & indices, int64_t num_weights, int64_t padding_idx, bool scale_grad_by_freq, at::Tensor & out) {
60
+ return at::_ops::embedding_dense_backward_out::call(grad_output, indices, num_weights, padding_idx, scale_grad_by_freq, out);
61
+ }
62
+ namespace symint {
63
+ template <typename T, typename = std::enable_if_t<std::is_same<T, int64_t>::value>>
64
+ at::Tensor & embedding_dense_backward_outf(const at::Tensor & grad_output, const at::Tensor & indices, int64_t num_weights, int64_t padding_idx, bool scale_grad_by_freq, at::Tensor & out) {
65
+ return at::_ops::embedding_dense_backward_out::call(grad_output, indices, num_weights, padding_idx, scale_grad_by_freq, out);
66
+ }
67
+ }
68
+
69
+ // aten::embedding_dense_backward.out(Tensor grad_output, Tensor indices, SymInt num_weights, SymInt padding_idx, bool scale_grad_by_freq, *, Tensor(a!) out) -> Tensor(a!)
70
+ inline at::Tensor & embedding_dense_backward_symint_out(at::Tensor & out, const at::Tensor & grad_output, const at::Tensor & indices, c10::SymInt num_weights, c10::SymInt padding_idx, bool scale_grad_by_freq) {
71
+ return at::_ops::embedding_dense_backward_out::call(grad_output, indices, num_weights, padding_idx, scale_grad_by_freq, out);
72
+ }
73
+ namespace symint {
74
+ template <typename T, typename = std::enable_if_t<std::is_same<T, c10::SymInt>::value>>
75
+ at::Tensor & embedding_dense_backward_out(at::Tensor & out, const at::Tensor & grad_output, const at::Tensor & indices, c10::SymInt num_weights, c10::SymInt padding_idx, bool scale_grad_by_freq) {
76
+ return at::_ops::embedding_dense_backward_out::call(grad_output, indices, num_weights, padding_idx, scale_grad_by_freq, out);
77
+ }
78
+ }
79
+
80
+ // aten::embedding_dense_backward.out(Tensor grad_output, Tensor indices, SymInt num_weights, SymInt padding_idx, bool scale_grad_by_freq, *, Tensor(a!) out) -> Tensor(a!)
81
+ inline at::Tensor & embedding_dense_backward_symint_outf(const at::Tensor & grad_output, const at::Tensor & indices, c10::SymInt num_weights, c10::SymInt padding_idx, bool scale_grad_by_freq, at::Tensor & out) {
82
+ return at::_ops::embedding_dense_backward_out::call(grad_output, indices, num_weights, padding_idx, scale_grad_by_freq, out);
83
+ }
84
+ namespace symint {
85
+ template <typename T, typename = std::enable_if_t<std::is_same<T, c10::SymInt>::value>>
86
+ at::Tensor & embedding_dense_backward_outf(const at::Tensor & grad_output, const at::Tensor & indices, c10::SymInt num_weights, c10::SymInt padding_idx, bool scale_grad_by_freq, at::Tensor & out) {
87
+ return at::_ops::embedding_dense_backward_out::call(grad_output, indices, num_weights, padding_idx, scale_grad_by_freq, out);
88
+ }
89
+ }
90
+
91
+ }