Add files using upload-large-folder tool
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- .gitattributes +6 -0
- llava/lib/libasan.so.6.0.0 +3 -0
- llava/lib/libbz2.so +3 -0
- llava/lib/libbz2.so.1.0.8 +3 -0
- llava/lib/libncurses.so +3 -0
- llava/lib/libssl.so.3 +3 -0
- llava/lib/libtinfo.so.6 +3 -0
- llava/lib/tcl8.6/msgs/af_za.msg +6 -0
- llava/lib/tcl8.6/msgs/ar_sy.msg +39 -0
- llava/lib/tcl8.6/msgs/en_nz.msg +7 -0
- llava/lib/tcl8.6/msgs/eo.msg +54 -0
- llava/lib/tcl8.6/msgs/es.msg +52 -0
- llava/lib/tcl8.6/msgs/es_gt.msg +6 -0
- llava/lib/tcl8.6/msgs/es_hn.msg +6 -0
- llava/lib/tcl8.6/msgs/es_ni.msg +6 -0
- llava/lib/tcl8.6/msgs/es_pr.msg +6 -0
- llava/lib/tcl8.6/msgs/gl.msg +47 -0
- llava/lib/tcl8.6/msgs/kok_in.msg +6 -0
- llava/lib/tcl8.6/msgs/mr.msg +39 -0
- llava/lib/tcl8.6/msgs/nl.msg +50 -0
- llava/lib/tcl8.6/msgs/nl_be.msg +7 -0
- llava/lib/tcl8.6/msgs/nn.msg +52 -0
- llava/lib/tcl8.6/msgs/pt.msg +50 -0
- llava/lib/tcl8.6/msgs/ro.msg +52 -0
- llava/lib/tcl8.6/msgs/ru_ua.msg +6 -0
- llava/lib/tcl8.6/msgs/sw.msg +49 -0
- llava/lib/tcl8.6/msgs/th.msg +54 -0
- llava/lib/tcl8.6/msgs/zh_sg.msg +8 -0
- parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_foreach_floor_cuda_dispatch.h +24 -0
- parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_foreach_log2.h +44 -0
- parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_foreach_neg_cuda_dispatch.h +24 -0
- parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_lazy_clone_ops.h +28 -0
- parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_nnpack_spatial_convolution.h +91 -0
- parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_prelu_kernel_backward_native.h +22 -0
- parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_sparse_semi_structured_apply_native.h +21 -0
- parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_unsafe_index_put_compositeexplicitautograd_dispatch.h +23 -0
- parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_upsample_bicubic2d_aa_backward_meta_dispatch.h +28 -0
- parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_upsample_bilinear2d_aa_meta.h +27 -0
- parrot/lib/python3.10/site-packages/torch/include/ATen/ops/adaptive_avg_pool3d.h +91 -0
- parrot/lib/python3.10/site-packages/torch/include/ATen/ops/adaptive_max_pool2d_backward_meta.h +27 -0
- parrot/lib/python3.10/site-packages/torch/include/ATen/ops/atan2_meta_dispatch.h +26 -0
- parrot/lib/python3.10/site-packages/torch/include/ATen/ops/batch_norm_gather_stats_compositeexplicitautograd_dispatch.h +24 -0
- parrot/lib/python3.10/site-packages/torch/include/ATen/ops/batch_norm_gather_stats_with_counts_compositeexplicitautograd_dispatch.h +24 -0
- parrot/lib/python3.10/site-packages/torch/include/ATen/ops/complex_cuda_dispatch.h +24 -0
- parrot/lib/python3.10/site-packages/torch/include/ATen/ops/conv1d_ops.h +39 -0
- parrot/lib/python3.10/site-packages/torch/include/ATen/ops/cudnn_grid_sampler_ops.h +39 -0
- parrot/lib/python3.10/site-packages/torch/include/ATen/ops/cummax.h +53 -0
- parrot/lib/python3.10/site-packages/torch/include/ATen/ops/diagonal_native.h +22 -0
- parrot/lib/python3.10/site-packages/torch/include/ATen/ops/diff_compositeimplicitautograd_dispatch.h +25 -0
- parrot/lib/python3.10/site-packages/torch/include/ATen/ops/embedding_dense_backward.h +91 -0
.gitattributes
CHANGED
|
@@ -403,3 +403,9 @@ llava/lib/libquadmath.so filter=lfs diff=lfs merge=lfs -text
|
|
| 403 |
llava/lib/liblzma.so.5 filter=lfs diff=lfs merge=lfs -text
|
| 404 |
llava/lib/libncurses.so.6 filter=lfs diff=lfs merge=lfs -text
|
| 405 |
llava/lib/libtsan.so.0.0.0 filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 403 |
llava/lib/liblzma.so.5 filter=lfs diff=lfs merge=lfs -text
|
| 404 |
llava/lib/libncurses.so.6 filter=lfs diff=lfs merge=lfs -text
|
| 405 |
llava/lib/libtsan.so.0.0.0 filter=lfs diff=lfs merge=lfs -text
|
| 406 |
+
llava/lib/libncurses.so filter=lfs diff=lfs merge=lfs -text
|
| 407 |
+
llava/lib/libtinfo.so.6 filter=lfs diff=lfs merge=lfs -text
|
| 408 |
+
llava/lib/libbz2.so filter=lfs diff=lfs merge=lfs -text
|
| 409 |
+
llava/lib/libasan.so.6.0.0 filter=lfs diff=lfs merge=lfs -text
|
| 410 |
+
llava/lib/libbz2.so.1.0.8 filter=lfs diff=lfs merge=lfs -text
|
| 411 |
+
llava/lib/libssl.so.3 filter=lfs diff=lfs merge=lfs -text
|
llava/lib/libasan.so.6.0.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2a8a7995a4d84a8817af8d1604bef621e99d0622df4eda14f6fe5245735a952e
|
| 3 |
+
size 7575272
|
llava/lib/libbz2.so
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4979469ae49ac144f62202f75bbdd69b17197aedb879d633337c8cf7e4aba301
|
| 3 |
+
size 229016
|
llava/lib/libbz2.so.1.0.8
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4979469ae49ac144f62202f75bbdd69b17197aedb879d633337c8cf7e4aba301
|
| 3 |
+
size 229016
|
llava/lib/libncurses.so
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7fa4e5e93804d78660b0eef727cdb4211209e1742e4ad3669348022668d90962
|
| 3 |
+
size 271304
|
llava/lib/libssl.so.3
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3480c91df4e0c1a33514955568641405e37924f680e8ba42f494a209640516c6
|
| 3 |
+
size 775712
|
llava/lib/libtinfo.so.6
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f5679c9d7cc0ec2d6b08c6058198667efe71f657e89dcc0bd7adcf5d6cbc80c5
|
| 3 |
+
size 287080
|
llava/lib/tcl8.6/msgs/af_za.msg
ADDED
|
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# created by tools/loadICU.tcl -- do not edit
|
| 2 |
+
namespace eval ::tcl::clock {
|
| 3 |
+
::msgcat::mcset af_ZA DATE_FORMAT "%d %B %Y"
|
| 4 |
+
::msgcat::mcset af_ZA TIME_FORMAT_12 "%l:%M:%S %P"
|
| 5 |
+
::msgcat::mcset af_ZA DATE_TIME_FORMAT "%d %B %Y %l:%M:%S %P %z"
|
| 6 |
+
}
|
llava/lib/tcl8.6/msgs/ar_sy.msg
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# created by tools/loadICU.tcl -- do not edit
|
| 2 |
+
namespace eval ::tcl::clock {
|
| 3 |
+
::msgcat::mcset ar_SY DAYS_OF_WEEK_ABBREV [list \
|
| 4 |
+
"\u0627\u0644\u0623\u062d\u062f"\
|
| 5 |
+
"\u0627\u0644\u0627\u062b\u0646\u064a\u0646"\
|
| 6 |
+
"\u0627\u0644\u062b\u0644\u0627\u062b\u0627\u0621"\
|
| 7 |
+
"\u0627\u0644\u0623\u0631\u0628\u0639\u0627\u0621"\
|
| 8 |
+
"\u0627\u0644\u062e\u0645\u064a\u0633"\
|
| 9 |
+
"\u0627\u0644\u062c\u0645\u0639\u0629"\
|
| 10 |
+
"\u0627\u0644\u0633\u0628\u062a"]
|
| 11 |
+
::msgcat::mcset ar_SY MONTHS_ABBREV [list \
|
| 12 |
+
"\u0643\u0627\u0646\u0648\u0646 \u0627\u0644\u062b\u0627\u0646\u064a"\
|
| 13 |
+
"\u0634\u0628\u0627\u0637"\
|
| 14 |
+
"\u0622\u0630\u0627\u0631"\
|
| 15 |
+
"\u0646\u064a\u0633\u0627\u0646"\
|
| 16 |
+
"\u0646\u0648\u0627\u0631"\
|
| 17 |
+
"\u062d\u0632\u064a\u0631\u0627\u0646"\
|
| 18 |
+
"\u062a\u0645\u0648\u0632"\
|
| 19 |
+
"\u0622\u0628"\
|
| 20 |
+
"\u0623\u064a\u0644\u0648\u0644"\
|
| 21 |
+
"\u062a\u0634\u0631\u064a\u0646 \u0627\u0644\u0623\u0648\u0644"\
|
| 22 |
+
"\u062a\u0634\u0631\u064a\u0646 \u0627\u0644\u062b\u0627\u0646\u064a"\
|
| 23 |
+
"\u0643\u0627\u0646\u0648\u0646 \u0627\u0644\u0623\u0648\u0644"\
|
| 24 |
+
""]
|
| 25 |
+
::msgcat::mcset ar_SY MONTHS_FULL [list \
|
| 26 |
+
"\u0643\u0627\u0646\u0648\u0646 \u0627\u0644\u062b\u0627\u0646\u064a"\
|
| 27 |
+
"\u0634\u0628\u0627\u0637"\
|
| 28 |
+
"\u0622\u0630\u0627\u0631"\
|
| 29 |
+
"\u0646\u064a\u0633\u0627\u0646"\
|
| 30 |
+
"\u0646\u0648\u0627\u0631\u0627\u0646"\
|
| 31 |
+
"\u062d\u0632\u064a\u0631"\
|
| 32 |
+
"\u062a\u0645\u0648\u0632"\
|
| 33 |
+
"\u0622\u0628"\
|
| 34 |
+
"\u0623\u064a\u0644\u0648\u0644"\
|
| 35 |
+
"\u062a\u0634\u0631\u064a\u0646 \u0627\u0644\u0623\u0648\u0644"\
|
| 36 |
+
"\u062a\u0634\u0631\u064a\u0646 \u0627\u0644\u062b\u0627\u0646\u064a"\
|
| 37 |
+
"\u0643\u0627\u0646\u0648\u0646 \u0627\u0644\u0623\u0648\u0644"\
|
| 38 |
+
""]
|
| 39 |
+
}
|
llava/lib/tcl8.6/msgs/en_nz.msg
ADDED
|
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# created by tools/loadICU.tcl -- do not edit
|
| 2 |
+
namespace eval ::tcl::clock {
|
| 3 |
+
::msgcat::mcset en_NZ DATE_FORMAT "%e/%m/%Y"
|
| 4 |
+
::msgcat::mcset en_NZ TIME_FORMAT "%H:%M:%S"
|
| 5 |
+
::msgcat::mcset en_NZ TIME_FORMAT_12 "%I:%M:%S %P %z"
|
| 6 |
+
::msgcat::mcset en_NZ DATE_TIME_FORMAT "%e/%m/%Y %H:%M:%S %z"
|
| 7 |
+
}
|
llava/lib/tcl8.6/msgs/eo.msg
ADDED
|
@@ -0,0 +1,54 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# created by tools/loadICU.tcl -- do not edit
|
| 2 |
+
namespace eval ::tcl::clock {
|
| 3 |
+
::msgcat::mcset eo DAYS_OF_WEEK_ABBREV [list \
|
| 4 |
+
"di"\
|
| 5 |
+
"lu"\
|
| 6 |
+
"ma"\
|
| 7 |
+
"me"\
|
| 8 |
+
"\u0135a"\
|
| 9 |
+
"ve"\
|
| 10 |
+
"sa"]
|
| 11 |
+
::msgcat::mcset eo DAYS_OF_WEEK_FULL [list \
|
| 12 |
+
"diman\u0109o"\
|
| 13 |
+
"lundo"\
|
| 14 |
+
"mardo"\
|
| 15 |
+
"merkredo"\
|
| 16 |
+
"\u0135a\u016ddo"\
|
| 17 |
+
"vendredo"\
|
| 18 |
+
"sabato"]
|
| 19 |
+
::msgcat::mcset eo MONTHS_ABBREV [list \
|
| 20 |
+
"jan"\
|
| 21 |
+
"feb"\
|
| 22 |
+
"mar"\
|
| 23 |
+
"apr"\
|
| 24 |
+
"maj"\
|
| 25 |
+
"jun"\
|
| 26 |
+
"jul"\
|
| 27 |
+
"a\u016dg"\
|
| 28 |
+
"sep"\
|
| 29 |
+
"okt"\
|
| 30 |
+
"nov"\
|
| 31 |
+
"dec"\
|
| 32 |
+
""]
|
| 33 |
+
::msgcat::mcset eo MONTHS_FULL [list \
|
| 34 |
+
"januaro"\
|
| 35 |
+
"februaro"\
|
| 36 |
+
"marto"\
|
| 37 |
+
"aprilo"\
|
| 38 |
+
"majo"\
|
| 39 |
+
"junio"\
|
| 40 |
+
"julio"\
|
| 41 |
+
"a\u016dgusto"\
|
| 42 |
+
"septembro"\
|
| 43 |
+
"oktobro"\
|
| 44 |
+
"novembro"\
|
| 45 |
+
"decembro"\
|
| 46 |
+
""]
|
| 47 |
+
::msgcat::mcset eo BCE "aK"
|
| 48 |
+
::msgcat::mcset eo CE "pK"
|
| 49 |
+
::msgcat::mcset eo AM "atm"
|
| 50 |
+
::msgcat::mcset eo PM "ptm"
|
| 51 |
+
::msgcat::mcset eo DATE_FORMAT "%Y-%b-%d"
|
| 52 |
+
::msgcat::mcset eo TIME_FORMAT "%H:%M:%S"
|
| 53 |
+
::msgcat::mcset eo DATE_TIME_FORMAT "%Y-%b-%d %H:%M:%S %z"
|
| 54 |
+
}
|
llava/lib/tcl8.6/msgs/es.msg
ADDED
|
@@ -0,0 +1,52 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# created by tools/loadICU.tcl -- do not edit
|
| 2 |
+
namespace eval ::tcl::clock {
|
| 3 |
+
::msgcat::mcset es DAYS_OF_WEEK_ABBREV [list \
|
| 4 |
+
"dom"\
|
| 5 |
+
"lun"\
|
| 6 |
+
"mar"\
|
| 7 |
+
"mi\u00e9"\
|
| 8 |
+
"jue"\
|
| 9 |
+
"vie"\
|
| 10 |
+
"s\u00e1b"]
|
| 11 |
+
::msgcat::mcset es DAYS_OF_WEEK_FULL [list \
|
| 12 |
+
"domingo"\
|
| 13 |
+
"lunes"\
|
| 14 |
+
"martes"\
|
| 15 |
+
"mi\u00e9rcoles"\
|
| 16 |
+
"jueves"\
|
| 17 |
+
"viernes"\
|
| 18 |
+
"s\u00e1bado"]
|
| 19 |
+
::msgcat::mcset es MONTHS_ABBREV [list \
|
| 20 |
+
"ene"\
|
| 21 |
+
"feb"\
|
| 22 |
+
"mar"\
|
| 23 |
+
"abr"\
|
| 24 |
+
"may"\
|
| 25 |
+
"jun"\
|
| 26 |
+
"jul"\
|
| 27 |
+
"ago"\
|
| 28 |
+
"sep"\
|
| 29 |
+
"oct"\
|
| 30 |
+
"nov"\
|
| 31 |
+
"dic"\
|
| 32 |
+
""]
|
| 33 |
+
::msgcat::mcset es MONTHS_FULL [list \
|
| 34 |
+
"enero"\
|
| 35 |
+
"febrero"\
|
| 36 |
+
"marzo"\
|
| 37 |
+
"abril"\
|
| 38 |
+
"mayo"\
|
| 39 |
+
"junio"\
|
| 40 |
+
"julio"\
|
| 41 |
+
"agosto"\
|
| 42 |
+
"septiembre"\
|
| 43 |
+
"octubre"\
|
| 44 |
+
"noviembre"\
|
| 45 |
+
"diciembre"\
|
| 46 |
+
""]
|
| 47 |
+
::msgcat::mcset es BCE "a.C."
|
| 48 |
+
::msgcat::mcset es CE "d.C."
|
| 49 |
+
::msgcat::mcset es DATE_FORMAT "%e de %B de %Y"
|
| 50 |
+
::msgcat::mcset es TIME_FORMAT "%k:%M:%S"
|
| 51 |
+
::msgcat::mcset es DATE_TIME_FORMAT "%e de %B de %Y %k:%M:%S %z"
|
| 52 |
+
}
|
llava/lib/tcl8.6/msgs/es_gt.msg
ADDED
|
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# created by tools/loadICU.tcl -- do not edit
|
| 2 |
+
namespace eval ::tcl::clock {
|
| 3 |
+
::msgcat::mcset es_GT DATE_FORMAT "%e/%m/%Y"
|
| 4 |
+
::msgcat::mcset es_GT TIME_FORMAT_12 "%I:%M:%S %P"
|
| 5 |
+
::msgcat::mcset es_GT DATE_TIME_FORMAT "%e/%m/%Y %I:%M:%S %P %z"
|
| 6 |
+
}
|
llava/lib/tcl8.6/msgs/es_hn.msg
ADDED
|
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# created by tools/loadICU.tcl -- do not edit
|
| 2 |
+
namespace eval ::tcl::clock {
|
| 3 |
+
::msgcat::mcset es_HN DATE_FORMAT "%m-%d-%Y"
|
| 4 |
+
::msgcat::mcset es_HN TIME_FORMAT_12 "%I:%M:%S %P"
|
| 5 |
+
::msgcat::mcset es_HN DATE_TIME_FORMAT "%m-%d-%Y %I:%M:%S %P %z"
|
| 6 |
+
}
|
llava/lib/tcl8.6/msgs/es_ni.msg
ADDED
|
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# created by tools/loadICU.tcl -- do not edit
|
| 2 |
+
namespace eval ::tcl::clock {
|
| 3 |
+
::msgcat::mcset es_NI DATE_FORMAT "%m-%d-%Y"
|
| 4 |
+
::msgcat::mcset es_NI TIME_FORMAT_12 "%I:%M:%S %P"
|
| 5 |
+
::msgcat::mcset es_NI DATE_TIME_FORMAT "%m-%d-%Y %I:%M:%S %P %z"
|
| 6 |
+
}
|
llava/lib/tcl8.6/msgs/es_pr.msg
ADDED
|
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# created by tools/loadICU.tcl -- do not edit
|
| 2 |
+
namespace eval ::tcl::clock {
|
| 3 |
+
::msgcat::mcset es_PR DATE_FORMAT "%m-%d-%Y"
|
| 4 |
+
::msgcat::mcset es_PR TIME_FORMAT_12 "%I:%M:%S %P"
|
| 5 |
+
::msgcat::mcset es_PR DATE_TIME_FORMAT "%m-%d-%Y %I:%M:%S %P %z"
|
| 6 |
+
}
|
llava/lib/tcl8.6/msgs/gl.msg
ADDED
|
@@ -0,0 +1,47 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# created by tools/loadICU.tcl -- do not edit
|
| 2 |
+
namespace eval ::tcl::clock {
|
| 3 |
+
::msgcat::mcset gl DAYS_OF_WEEK_ABBREV [list \
|
| 4 |
+
"Dom"\
|
| 5 |
+
"Lun"\
|
| 6 |
+
"Mar"\
|
| 7 |
+
"M\u00e9r"\
|
| 8 |
+
"Xov"\
|
| 9 |
+
"Ven"\
|
| 10 |
+
"S\u00e1b"]
|
| 11 |
+
::msgcat::mcset gl DAYS_OF_WEEK_FULL [list \
|
| 12 |
+
"Domingo"\
|
| 13 |
+
"Luns"\
|
| 14 |
+
"Martes"\
|
| 15 |
+
"M\u00e9rcores"\
|
| 16 |
+
"Xoves"\
|
| 17 |
+
"Venres"\
|
| 18 |
+
"S\u00e1bado"]
|
| 19 |
+
::msgcat::mcset gl MONTHS_ABBREV [list \
|
| 20 |
+
"Xan"\
|
| 21 |
+
"Feb"\
|
| 22 |
+
"Mar"\
|
| 23 |
+
"Abr"\
|
| 24 |
+
"Mai"\
|
| 25 |
+
"Xu\u00f1"\
|
| 26 |
+
"Xul"\
|
| 27 |
+
"Ago"\
|
| 28 |
+
"Set"\
|
| 29 |
+
"Out"\
|
| 30 |
+
"Nov"\
|
| 31 |
+
"Dec"\
|
| 32 |
+
""]
|
| 33 |
+
::msgcat::mcset gl MONTHS_FULL [list \
|
| 34 |
+
"Xaneiro"\
|
| 35 |
+
"Febreiro"\
|
| 36 |
+
"Marzo"\
|
| 37 |
+
"Abril"\
|
| 38 |
+
"Maio"\
|
| 39 |
+
"Xu\u00f1o"\
|
| 40 |
+
"Xullo"\
|
| 41 |
+
"Agosto"\
|
| 42 |
+
"Setembro"\
|
| 43 |
+
"Outubro"\
|
| 44 |
+
"Novembro"\
|
| 45 |
+
"Decembro"\
|
| 46 |
+
""]
|
| 47 |
+
}
|
llava/lib/tcl8.6/msgs/kok_in.msg
ADDED
|
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# created by tools/loadICU.tcl -- do not edit
|
| 2 |
+
namespace eval ::tcl::clock {
|
| 3 |
+
::msgcat::mcset kok_IN DATE_FORMAT "%d %M %Y"
|
| 4 |
+
::msgcat::mcset kok_IN TIME_FORMAT_12 "%I:%M:%S %P"
|
| 5 |
+
::msgcat::mcset kok_IN DATE_TIME_FORMAT "%d %M %Y %I:%M:%S %P %z"
|
| 6 |
+
}
|
llava/lib/tcl8.6/msgs/mr.msg
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# created by tools/loadICU.tcl -- do not edit
|
| 2 |
+
namespace eval ::tcl::clock {
|
| 3 |
+
::msgcat::mcset mr DAYS_OF_WEEK_FULL [list \
|
| 4 |
+
"\u0930\u0935\u093f\u0935\u093e\u0930"\
|
| 5 |
+
"\u0938\u094b\u092e\u0935\u093e\u0930"\
|
| 6 |
+
"\u092e\u0902\u0917\u0933\u0935\u093e\u0930"\
|
| 7 |
+
"\u092e\u0902\u0917\u0933\u0935\u093e\u0930"\
|
| 8 |
+
"\u0917\u0941\u0930\u0941\u0935\u093e\u0930"\
|
| 9 |
+
"\u0936\u0941\u0915\u094d\u0930\u0935\u093e\u0930"\
|
| 10 |
+
"\u0936\u0928\u093f\u0935\u093e\u0930"]
|
| 11 |
+
::msgcat::mcset mr MONTHS_ABBREV [list \
|
| 12 |
+
"\u091c\u093e\u0928\u0947\u0935\u093e\u0930\u0940"\
|
| 13 |
+
"\u092b\u0947\u092c\u0943\u0935\u093e\u0930\u0940"\
|
| 14 |
+
"\u092e\u093e\u0930\u094d\u091a"\
|
| 15 |
+
"\u090f\u092a\u094d\u0930\u093f\u0932"\
|
| 16 |
+
"\u092e\u0947"\
|
| 17 |
+
"\u091c\u0942\u0928"\
|
| 18 |
+
"\u091c\u0941\u0932\u0948"\
|
| 19 |
+
"\u0913\u0917\u0938\u094d\u091f"\
|
| 20 |
+
"\u0938\u0947\u092a\u094d\u091f\u0947\u0902\u092c\u0930"\
|
| 21 |
+
"\u0913\u0915\u094d\u091f\u094b\u092c\u0930"\
|
| 22 |
+
"\u0928\u094b\u0935\u094d\u0939\u0947\u0902\u092c\u0930"\
|
| 23 |
+
"\u0921\u093f\u0938\u0947\u0902\u092c\u0930"]
|
| 24 |
+
::msgcat::mcset mr MONTHS_FULL [list \
|
| 25 |
+
"\u091c\u093e\u0928\u0947\u0935\u093e\u0930\u0940"\
|
| 26 |
+
"\u092b\u0947\u092c\u0943\u0935\u093e\u0930\u0940"\
|
| 27 |
+
"\u092e\u093e\u0930\u094d\u091a"\
|
| 28 |
+
"\u090f\u092a\u094d\u0930\u093f\u0932"\
|
| 29 |
+
"\u092e\u0947"\
|
| 30 |
+
"\u091c\u0942\u0928"\
|
| 31 |
+
"\u091c\u0941\u0932\u0948"\
|
| 32 |
+
"\u0913\u0917\u0938\u094d\u091f"\
|
| 33 |
+
"\u0938\u0947\u092a\u094d\u091f\u0947\u0902\u092c\u0930"\
|
| 34 |
+
"\u0913\u0915\u094d\u091f\u094b\u092c\u0930"\
|
| 35 |
+
"\u0928\u094b\u0935\u094d\u0939\u0947\u0902\u092c\u0930"\
|
| 36 |
+
"\u0921\u093f\u0938\u0947\u0902\u092c\u0930"]
|
| 37 |
+
::msgcat::mcset mr AM "BC"
|
| 38 |
+
::msgcat::mcset mr PM "AD"
|
| 39 |
+
}
|
llava/lib/tcl8.6/msgs/nl.msg
ADDED
|
@@ -0,0 +1,50 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# created by tools/loadICU.tcl -- do not edit
|
| 2 |
+
namespace eval ::tcl::clock {
|
| 3 |
+
::msgcat::mcset nl DAYS_OF_WEEK_ABBREV [list \
|
| 4 |
+
"zo"\
|
| 5 |
+
"ma"\
|
| 6 |
+
"di"\
|
| 7 |
+
"wo"\
|
| 8 |
+
"do"\
|
| 9 |
+
"vr"\
|
| 10 |
+
"za"]
|
| 11 |
+
::msgcat::mcset nl DAYS_OF_WEEK_FULL [list \
|
| 12 |
+
"zondag"\
|
| 13 |
+
"maandag"\
|
| 14 |
+
"dinsdag"\
|
| 15 |
+
"woensdag"\
|
| 16 |
+
"donderdag"\
|
| 17 |
+
"vrijdag"\
|
| 18 |
+
"zaterdag"]
|
| 19 |
+
::msgcat::mcset nl MONTHS_ABBREV [list \
|
| 20 |
+
"jan"\
|
| 21 |
+
"feb"\
|
| 22 |
+
"mrt"\
|
| 23 |
+
"apr"\
|
| 24 |
+
"mei"\
|
| 25 |
+
"jun"\
|
| 26 |
+
"jul"\
|
| 27 |
+
"aug"\
|
| 28 |
+
"sep"\
|
| 29 |
+
"okt"\
|
| 30 |
+
"nov"\
|
| 31 |
+
"dec"\
|
| 32 |
+
""]
|
| 33 |
+
::msgcat::mcset nl MONTHS_FULL [list \
|
| 34 |
+
"januari"\
|
| 35 |
+
"februari"\
|
| 36 |
+
"maart"\
|
| 37 |
+
"april"\
|
| 38 |
+
"mei"\
|
| 39 |
+
"juni"\
|
| 40 |
+
"juli"\
|
| 41 |
+
"augustus"\
|
| 42 |
+
"september"\
|
| 43 |
+
"oktober"\
|
| 44 |
+
"november"\
|
| 45 |
+
"december"\
|
| 46 |
+
""]
|
| 47 |
+
::msgcat::mcset nl DATE_FORMAT "%e %B %Y"
|
| 48 |
+
::msgcat::mcset nl TIME_FORMAT "%k:%M:%S"
|
| 49 |
+
::msgcat::mcset nl DATE_TIME_FORMAT "%e %B %Y %k:%M:%S %z"
|
| 50 |
+
}
|
llava/lib/tcl8.6/msgs/nl_be.msg
ADDED
|
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# created by tools/loadICU.tcl -- do not edit
|
| 2 |
+
namespace eval ::tcl::clock {
|
| 3 |
+
::msgcat::mcset nl_BE DATE_FORMAT "%d-%m-%y"
|
| 4 |
+
::msgcat::mcset nl_BE TIME_FORMAT "%T"
|
| 5 |
+
::msgcat::mcset nl_BE TIME_FORMAT_12 "%T"
|
| 6 |
+
::msgcat::mcset nl_BE DATE_TIME_FORMAT "%a %d %b %Y %T %z"
|
| 7 |
+
}
|
llava/lib/tcl8.6/msgs/nn.msg
ADDED
|
@@ -0,0 +1,52 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# created by tools/loadICU.tcl -- do not edit
|
| 2 |
+
namespace eval ::tcl::clock {
|
| 3 |
+
::msgcat::mcset nn DAYS_OF_WEEK_ABBREV [list \
|
| 4 |
+
"su"\
|
| 5 |
+
"m\u00e5"\
|
| 6 |
+
"ty"\
|
| 7 |
+
"on"\
|
| 8 |
+
"to"\
|
| 9 |
+
"fr"\
|
| 10 |
+
"lau"]
|
| 11 |
+
::msgcat::mcset nn DAYS_OF_WEEK_FULL [list \
|
| 12 |
+
"sundag"\
|
| 13 |
+
"m\u00e5ndag"\
|
| 14 |
+
"tysdag"\
|
| 15 |
+
"onsdag"\
|
| 16 |
+
"torsdag"\
|
| 17 |
+
"fredag"\
|
| 18 |
+
"laurdag"]
|
| 19 |
+
::msgcat::mcset nn MONTHS_ABBREV [list \
|
| 20 |
+
"jan"\
|
| 21 |
+
"feb"\
|
| 22 |
+
"mar"\
|
| 23 |
+
"apr"\
|
| 24 |
+
"mai"\
|
| 25 |
+
"jun"\
|
| 26 |
+
"jul"\
|
| 27 |
+
"aug"\
|
| 28 |
+
"sep"\
|
| 29 |
+
"okt"\
|
| 30 |
+
"nov"\
|
| 31 |
+
"des"\
|
| 32 |
+
""]
|
| 33 |
+
::msgcat::mcset nn MONTHS_FULL [list \
|
| 34 |
+
"januar"\
|
| 35 |
+
"februar"\
|
| 36 |
+
"mars"\
|
| 37 |
+
"april"\
|
| 38 |
+
"mai"\
|
| 39 |
+
"juni"\
|
| 40 |
+
"juli"\
|
| 41 |
+
"august"\
|
| 42 |
+
"september"\
|
| 43 |
+
"oktober"\
|
| 44 |
+
"november"\
|
| 45 |
+
"desember"\
|
| 46 |
+
""]
|
| 47 |
+
::msgcat::mcset nn BCE "f.Kr."
|
| 48 |
+
::msgcat::mcset nn CE "e.Kr."
|
| 49 |
+
::msgcat::mcset nn DATE_FORMAT "%e. %B %Y"
|
| 50 |
+
::msgcat::mcset nn TIME_FORMAT "%H:%M:%S"
|
| 51 |
+
::msgcat::mcset nn DATE_TIME_FORMAT "%e. %B %Y %H:%M:%S %z"
|
| 52 |
+
}
|
llava/lib/tcl8.6/msgs/pt.msg
ADDED
|
@@ -0,0 +1,50 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# created by tools/loadICU.tcl -- do not edit
|
| 2 |
+
namespace eval ::tcl::clock {
|
| 3 |
+
::msgcat::mcset pt DAYS_OF_WEEK_ABBREV [list \
|
| 4 |
+
"Dom"\
|
| 5 |
+
"Seg"\
|
| 6 |
+
"Ter"\
|
| 7 |
+
"Qua"\
|
| 8 |
+
"Qui"\
|
| 9 |
+
"Sex"\
|
| 10 |
+
"S\u00e1b"]
|
| 11 |
+
::msgcat::mcset pt DAYS_OF_WEEK_FULL [list \
|
| 12 |
+
"Domingo"\
|
| 13 |
+
"Segunda-feira"\
|
| 14 |
+
"Ter\u00e7a-feira"\
|
| 15 |
+
"Quarta-feira"\
|
| 16 |
+
"Quinta-feira"\
|
| 17 |
+
"Sexta-feira"\
|
| 18 |
+
"S\u00e1bado"]
|
| 19 |
+
::msgcat::mcset pt MONTHS_ABBREV [list \
|
| 20 |
+
"Jan"\
|
| 21 |
+
"Fev"\
|
| 22 |
+
"Mar"\
|
| 23 |
+
"Abr"\
|
| 24 |
+
"Mai"\
|
| 25 |
+
"Jun"\
|
| 26 |
+
"Jul"\
|
| 27 |
+
"Ago"\
|
| 28 |
+
"Set"\
|
| 29 |
+
"Out"\
|
| 30 |
+
"Nov"\
|
| 31 |
+
"Dez"\
|
| 32 |
+
""]
|
| 33 |
+
::msgcat::mcset pt MONTHS_FULL [list \
|
| 34 |
+
"Janeiro"\
|
| 35 |
+
"Fevereiro"\
|
| 36 |
+
"Mar\u00e7o"\
|
| 37 |
+
"Abril"\
|
| 38 |
+
"Maio"\
|
| 39 |
+
"Junho"\
|
| 40 |
+
"Julho"\
|
| 41 |
+
"Agosto"\
|
| 42 |
+
"Setembro"\
|
| 43 |
+
"Outubro"\
|
| 44 |
+
"Novembro"\
|
| 45 |
+
"Dezembro"\
|
| 46 |
+
""]
|
| 47 |
+
::msgcat::mcset pt DATE_FORMAT "%d-%m-%Y"
|
| 48 |
+
::msgcat::mcset pt TIME_FORMAT "%k:%M:%S"
|
| 49 |
+
::msgcat::mcset pt DATE_TIME_FORMAT "%d-%m-%Y %k:%M:%S %z"
|
| 50 |
+
}
|
llava/lib/tcl8.6/msgs/ro.msg
ADDED
|
@@ -0,0 +1,52 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# created by tools/loadICU.tcl -- do not edit
|
| 2 |
+
namespace eval ::tcl::clock {
|
| 3 |
+
::msgcat::mcset ro DAYS_OF_WEEK_ABBREV [list \
|
| 4 |
+
"D"\
|
| 5 |
+
"L"\
|
| 6 |
+
"Ma"\
|
| 7 |
+
"Mi"\
|
| 8 |
+
"J"\
|
| 9 |
+
"V"\
|
| 10 |
+
"S"]
|
| 11 |
+
::msgcat::mcset ro DAYS_OF_WEEK_FULL [list \
|
| 12 |
+
"duminic\u0103"\
|
| 13 |
+
"luni"\
|
| 14 |
+
"mar\u0163i"\
|
| 15 |
+
"miercuri"\
|
| 16 |
+
"joi"\
|
| 17 |
+
"vineri"\
|
| 18 |
+
"s\u00eemb\u0103t\u0103"]
|
| 19 |
+
::msgcat::mcset ro MONTHS_ABBREV [list \
|
| 20 |
+
"Ian"\
|
| 21 |
+
"Feb"\
|
| 22 |
+
"Mar"\
|
| 23 |
+
"Apr"\
|
| 24 |
+
"Mai"\
|
| 25 |
+
"Iun"\
|
| 26 |
+
"Iul"\
|
| 27 |
+
"Aug"\
|
| 28 |
+
"Sep"\
|
| 29 |
+
"Oct"\
|
| 30 |
+
"Nov"\
|
| 31 |
+
"Dec"\
|
| 32 |
+
""]
|
| 33 |
+
::msgcat::mcset ro MONTHS_FULL [list \
|
| 34 |
+
"ianuarie"\
|
| 35 |
+
"februarie"\
|
| 36 |
+
"martie"\
|
| 37 |
+
"aprilie"\
|
| 38 |
+
"mai"\
|
| 39 |
+
"iunie"\
|
| 40 |
+
"iulie"\
|
| 41 |
+
"august"\
|
| 42 |
+
"septembrie"\
|
| 43 |
+
"octombrie"\
|
| 44 |
+
"noiembrie"\
|
| 45 |
+
"decembrie"\
|
| 46 |
+
""]
|
| 47 |
+
::msgcat::mcset ro BCE "d.C."
|
| 48 |
+
::msgcat::mcset ro CE "\u00ee.d.C."
|
| 49 |
+
::msgcat::mcset ro DATE_FORMAT "%d.%m.%Y"
|
| 50 |
+
::msgcat::mcset ro TIME_FORMAT "%H:%M:%S"
|
| 51 |
+
::msgcat::mcset ro DATE_TIME_FORMAT "%d.%m.%Y %H:%M:%S %z"
|
| 52 |
+
}
|
llava/lib/tcl8.6/msgs/ru_ua.msg
ADDED
|
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# created by tools/loadICU.tcl -- do not edit
|
| 2 |
+
namespace eval ::tcl::clock {
|
| 3 |
+
::msgcat::mcset ru_UA DATE_FORMAT "%d.%m.%Y"
|
| 4 |
+
::msgcat::mcset ru_UA TIME_FORMAT "%k:%M:%S"
|
| 5 |
+
::msgcat::mcset ru_UA DATE_TIME_FORMAT "%d.%m.%Y %k:%M:%S %z"
|
| 6 |
+
}
|
llava/lib/tcl8.6/msgs/sw.msg
ADDED
|
@@ -0,0 +1,49 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# created by tools/loadICU.tcl -- do not edit
|
| 2 |
+
namespace eval ::tcl::clock {
|
| 3 |
+
::msgcat::mcset sw DAYS_OF_WEEK_ABBREV [list \
|
| 4 |
+
"Jpi"\
|
| 5 |
+
"Jtt"\
|
| 6 |
+
"Jnn"\
|
| 7 |
+
"Jtn"\
|
| 8 |
+
"Alh"\
|
| 9 |
+
"Iju"\
|
| 10 |
+
"Jmo"]
|
| 11 |
+
::msgcat::mcset sw DAYS_OF_WEEK_FULL [list \
|
| 12 |
+
"Jumapili"\
|
| 13 |
+
"Jumatatu"\
|
| 14 |
+
"Jumanne"\
|
| 15 |
+
"Jumatano"\
|
| 16 |
+
"Alhamisi"\
|
| 17 |
+
"Ijumaa"\
|
| 18 |
+
"Jumamosi"]
|
| 19 |
+
::msgcat::mcset sw MONTHS_ABBREV [list \
|
| 20 |
+
"Jan"\
|
| 21 |
+
"Feb"\
|
| 22 |
+
"Mar"\
|
| 23 |
+
"Apr"\
|
| 24 |
+
"Mei"\
|
| 25 |
+
"Jun"\
|
| 26 |
+
"Jul"\
|
| 27 |
+
"Ago"\
|
| 28 |
+
"Sep"\
|
| 29 |
+
"Okt"\
|
| 30 |
+
"Nov"\
|
| 31 |
+
"Des"\
|
| 32 |
+
""]
|
| 33 |
+
::msgcat::mcset sw MONTHS_FULL [list \
|
| 34 |
+
"Januari"\
|
| 35 |
+
"Februari"\
|
| 36 |
+
"Machi"\
|
| 37 |
+
"Aprili"\
|
| 38 |
+
"Mei"\
|
| 39 |
+
"Juni"\
|
| 40 |
+
"Julai"\
|
| 41 |
+
"Agosti"\
|
| 42 |
+
"Septemba"\
|
| 43 |
+
"Oktoba"\
|
| 44 |
+
"Novemba"\
|
| 45 |
+
"Desemba"\
|
| 46 |
+
""]
|
| 47 |
+
::msgcat::mcset sw BCE "KK"
|
| 48 |
+
::msgcat::mcset sw CE "BK"
|
| 49 |
+
}
|
llava/lib/tcl8.6/msgs/th.msg
ADDED
|
@@ -0,0 +1,54 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# created by tools/loadICU.tcl -- do not edit
|
| 2 |
+
namespace eval ::tcl::clock {
|
| 3 |
+
::msgcat::mcset th DAYS_OF_WEEK_ABBREV [list \
|
| 4 |
+
"\u0e2d\u0e32."\
|
| 5 |
+
"\u0e08."\
|
| 6 |
+
"\u0e2d."\
|
| 7 |
+
"\u0e1e."\
|
| 8 |
+
"\u0e1e\u0e24."\
|
| 9 |
+
"\u0e28."\
|
| 10 |
+
"\u0e2a."]
|
| 11 |
+
::msgcat::mcset th DAYS_OF_WEEK_FULL [list \
|
| 12 |
+
"\u0e27\u0e31\u0e19\u0e2d\u0e32\u0e17\u0e34\u0e15\u0e22\u0e4c"\
|
| 13 |
+
"\u0e27\u0e31\u0e19\u0e08\u0e31\u0e19\u0e17\u0e23\u0e4c"\
|
| 14 |
+
"\u0e27\u0e31\u0e19\u0e2d\u0e31\u0e07\u0e04\u0e32\u0e23"\
|
| 15 |
+
"\u0e27\u0e31\u0e19\u0e1e\u0e38\u0e18"\
|
| 16 |
+
"\u0e27\u0e31\u0e19\u0e1e\u0e24\u0e2b\u0e31\u0e2a\u0e1a\u0e14\u0e35"\
|
| 17 |
+
"\u0e27\u0e31\u0e19\u0e28\u0e38\u0e01\u0e23\u0e4c"\
|
| 18 |
+
"\u0e27\u0e31\u0e19\u0e40\u0e2a\u0e32\u0e23\u0e4c"]
|
| 19 |
+
::msgcat::mcset th MONTHS_ABBREV [list \
|
| 20 |
+
"\u0e21.\u0e04."\
|
| 21 |
+
"\u0e01.\u0e1e."\
|
| 22 |
+
"\u0e21\u0e35.\u0e04."\
|
| 23 |
+
"\u0e40\u0e21.\u0e22."\
|
| 24 |
+
"\u0e1e.\u0e04."\
|
| 25 |
+
"\u0e21\u0e34.\u0e22."\
|
| 26 |
+
"\u0e01.\u0e04."\
|
| 27 |
+
"\u0e2a.\u0e04."\
|
| 28 |
+
"\u0e01.\u0e22."\
|
| 29 |
+
"\u0e15.\u0e04."\
|
| 30 |
+
"\u0e1e.\u0e22."\
|
| 31 |
+
"\u0e18.\u0e04."\
|
| 32 |
+
""]
|
| 33 |
+
::msgcat::mcset th MONTHS_FULL [list \
|
| 34 |
+
"\u0e21\u0e01\u0e23\u0e32\u0e04\u0e21"\
|
| 35 |
+
"\u0e01\u0e38\u0e21\u0e20\u0e32\u0e1e\u0e31\u0e19\u0e18\u0e4c"\
|
| 36 |
+
"\u0e21\u0e35\u0e19\u0e32\u0e04\u0e21"\
|
| 37 |
+
"\u0e40\u0e21\u0e29\u0e32\u0e22\u0e19"\
|
| 38 |
+
"\u0e1e\u0e24\u0e29\u0e20\u0e32\u0e04\u0e21"\
|
| 39 |
+
"\u0e21\u0e34\u0e16\u0e38\u0e19\u0e32\u0e22\u0e19"\
|
| 40 |
+
"\u0e01\u0e23\u0e01\u0e0e\u0e32\u0e04\u0e21"\
|
| 41 |
+
"\u0e2a\u0e34\u0e07\u0e2b\u0e32\u0e04\u0e21"\
|
| 42 |
+
"\u0e01\u0e31\u0e19\u0e22\u0e32\u0e22\u0e19"\
|
| 43 |
+
"\u0e15\u0e38\u0e25\u0e32\u0e04\u0e21"\
|
| 44 |
+
"\u0e1e\u0e24\u0e28\u0e08\u0e34\u0e01\u0e32\u0e22\u0e19"\
|
| 45 |
+
"\u0e18\u0e31\u0e19\u0e27\u0e32\u0e04\u0e21"\
|
| 46 |
+
""]
|
| 47 |
+
::msgcat::mcset th BCE "\u0e25\u0e17\u0e35\u0e48"
|
| 48 |
+
::msgcat::mcset th CE "\u0e04.\u0e28."
|
| 49 |
+
::msgcat::mcset th AM "\u0e01\u0e48\u0e2d\u0e19\u0e40\u0e17\u0e35\u0e48\u0e22\u0e07"
|
| 50 |
+
::msgcat::mcset th PM "\u0e2b\u0e25\u0e31\u0e07\u0e40\u0e17\u0e35\u0e48\u0e22\u0e07"
|
| 51 |
+
::msgcat::mcset th DATE_FORMAT "%e/%m/%Y"
|
| 52 |
+
::msgcat::mcset th TIME_FORMAT "%k:%M:%S"
|
| 53 |
+
::msgcat::mcset th DATE_TIME_FORMAT "%e/%m/%Y %k:%M:%S %z"
|
| 54 |
+
}
|
llava/lib/tcl8.6/msgs/zh_sg.msg
ADDED
|
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# created by tools/loadICU.tcl -- do not edit
|
| 2 |
+
namespace eval ::tcl::clock {
|
| 3 |
+
::msgcat::mcset zh_SG AM "\u4e0a\u5348"
|
| 4 |
+
::msgcat::mcset zh_SG PM "\u4e2d\u5348"
|
| 5 |
+
::msgcat::mcset zh_SG DATE_FORMAT "%d %B %Y"
|
| 6 |
+
::msgcat::mcset zh_SG TIME_FORMAT_12 "%P %I:%M:%S"
|
| 7 |
+
::msgcat::mcset zh_SG DATE_TIME_FORMAT "%d %B %Y %P %I:%M:%S %z"
|
| 8 |
+
}
|
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_foreach_floor_cuda_dispatch.h
ADDED
|
@@ -0,0 +1,24 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
#pragma once
|
| 2 |
+
// @generated by torchgen/gen.py from DispatchKeyFunction.h
|
| 3 |
+
|
| 4 |
+
// NB: The implementing C++ file is RegisterDispatchKey.cpp
|
| 5 |
+
|
| 6 |
+
// The only #includes we need are for custom classes that have defaults in the C++ API
|
| 7 |
+
#include <c10/core/MemoryFormat.h>
|
| 8 |
+
#include <c10/core/Scalar.h>
|
| 9 |
+
#include <ATen/core/Reduction.h>
|
| 10 |
+
|
| 11 |
+
// Forward declarations of any types needed in the operator signatures.
|
| 12 |
+
// We can't directly include these classes because it will cause circular include dependencies.
|
| 13 |
+
// This file is included by TensorBody.h, which defines the Tensor class.
|
| 14 |
+
#include <ATen/core/ATen_fwd.h>
|
| 15 |
+
|
| 16 |
+
namespace at {
|
| 17 |
+
|
| 18 |
+
namespace cuda {
|
| 19 |
+
|
| 20 |
+
TORCH_API ::std::vector<at::Tensor> _foreach_floor(at::TensorList self);
|
| 21 |
+
TORCH_API void _foreach_floor_(at::TensorList self);
|
| 22 |
+
|
| 23 |
+
} // namespace cuda
|
| 24 |
+
} // namespace at
|
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_foreach_log2.h
ADDED
|
@@ -0,0 +1,44 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
#pragma once
|
| 2 |
+
|
| 3 |
+
// @generated by torchgen/gen.py from Function.h
|
| 4 |
+
|
| 5 |
+
#include <ATen/Context.h>
|
| 6 |
+
#include <ATen/DeviceGuard.h>
|
| 7 |
+
#include <ATen/TensorUtils.h>
|
| 8 |
+
#include <ATen/TracerMode.h>
|
| 9 |
+
#include <ATen/core/Generator.h>
|
| 10 |
+
#include <ATen/core/Reduction.h>
|
| 11 |
+
#include <ATen/core/Tensor.h>
|
| 12 |
+
#include <c10/core/Scalar.h>
|
| 13 |
+
#include <c10/core/Storage.h>
|
| 14 |
+
#include <c10/core/TensorOptions.h>
|
| 15 |
+
#include <c10/util/Deprecated.h>
|
| 16 |
+
#include <c10/util/Optional.h>
|
| 17 |
+
|
| 18 |
+
|
| 19 |
+
|
| 20 |
+
#include <ATen/ops/_foreach_log2_ops.h>
|
| 21 |
+
|
| 22 |
+
namespace at {
|
| 23 |
+
|
| 24 |
+
|
| 25 |
+
// aten::_foreach_log2(Tensor[] self) -> Tensor[]
|
| 26 |
+
inline ::std::vector<at::Tensor> _foreach_log2(at::TensorList self) {
|
| 27 |
+
return at::_ops::_foreach_log2::call(self);
|
| 28 |
+
}
|
| 29 |
+
|
| 30 |
+
// aten::_foreach_log2_(Tensor(a!)[] self) -> ()
|
| 31 |
+
inline void _foreach_log2_(at::TensorList self) {
|
| 32 |
+
return at::_ops::_foreach_log2_::call(self);
|
| 33 |
+
}
|
| 34 |
+
|
| 35 |
+
// aten::_foreach_log2.out(Tensor[] self, *, Tensor(a!)[] out) -> ()
|
| 36 |
+
inline void _foreach_log2_out(at::TensorList out, at::TensorList self) {
|
| 37 |
+
return at::_ops::_foreach_log2_out::call(self, out);
|
| 38 |
+
}
|
| 39 |
+
// aten::_foreach_log2.out(Tensor[] self, *, Tensor(a!)[] out) -> ()
|
| 40 |
+
inline void _foreach_log2_outf(at::TensorList self, at::TensorList out) {
|
| 41 |
+
return at::_ops::_foreach_log2_out::call(self, out);
|
| 42 |
+
}
|
| 43 |
+
|
| 44 |
+
}
|
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_foreach_neg_cuda_dispatch.h
ADDED
|
@@ -0,0 +1,24 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
#pragma once
|
| 2 |
+
// @generated by torchgen/gen.py from DispatchKeyFunction.h
|
| 3 |
+
|
| 4 |
+
// NB: The implementing C++ file is RegisterDispatchKey.cpp
|
| 5 |
+
|
| 6 |
+
// The only #includes we need are for custom classes that have defaults in the C++ API
|
| 7 |
+
#include <c10/core/MemoryFormat.h>
|
| 8 |
+
#include <c10/core/Scalar.h>
|
| 9 |
+
#include <ATen/core/Reduction.h>
|
| 10 |
+
|
| 11 |
+
// Forward declarations of any types needed in the operator signatures.
|
| 12 |
+
// We can't directly include these classes because it will cause circular include dependencies.
|
| 13 |
+
// This file is included by TensorBody.h, which defines the Tensor class.
|
| 14 |
+
#include <ATen/core/ATen_fwd.h>
|
| 15 |
+
|
| 16 |
+
namespace at {
|
| 17 |
+
|
| 18 |
+
namespace cuda {
|
| 19 |
+
|
| 20 |
+
TORCH_API ::std::vector<at::Tensor> _foreach_neg(at::TensorList self);
|
| 21 |
+
TORCH_API void _foreach_neg_(at::TensorList self);
|
| 22 |
+
|
| 23 |
+
} // namespace cuda
|
| 24 |
+
} // namespace at
|
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_lazy_clone_ops.h
ADDED
|
@@ -0,0 +1,28 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
#pragma once
|
| 2 |
+
|
| 3 |
+
// @generated by torchgen/gen.py from Operator.h
|
| 4 |
+
|
| 5 |
+
#include <tuple>
|
| 6 |
+
#include <vector>
|
| 7 |
+
|
| 8 |
+
// Forward declarations of any types needed in the operator signatures.
|
| 9 |
+
// We can't directly include these classes because it will cause circular include dependencies.
|
| 10 |
+
// This file is included by TensorBody.h, which defines the Tensor class.
|
| 11 |
+
#include <ATen/core/ATen_fwd.h>
|
| 12 |
+
|
| 13 |
+
namespace at {
|
| 14 |
+
namespace _ops {
|
| 15 |
+
|
| 16 |
+
|
| 17 |
+
struct TORCH_API _lazy_clone {
|
| 18 |
+
using schema = at::Tensor (const at::Tensor &);
|
| 19 |
+
using ptr_schema = schema*;
|
| 20 |
+
// See Note [static constexpr char* members for windows NVCC]
|
| 21 |
+
STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(name, "aten::_lazy_clone")
|
| 22 |
+
STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(overload_name, "")
|
| 23 |
+
STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(schema_str, "_lazy_clone(Tensor self) -> Tensor")
|
| 24 |
+
static at::Tensor call(const at::Tensor & self);
|
| 25 |
+
static at::Tensor redispatch(c10::DispatchKeySet dispatchKeySet, const at::Tensor & self);
|
| 26 |
+
};
|
| 27 |
+
|
| 28 |
+
}} // namespace at::_ops
|
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_nnpack_spatial_convolution.h
ADDED
|
@@ -0,0 +1,91 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
#pragma once
|
| 2 |
+
|
| 3 |
+
// @generated by torchgen/gen.py from Function.h
|
| 4 |
+
|
| 5 |
+
#include <ATen/Context.h>
|
| 6 |
+
#include <ATen/DeviceGuard.h>
|
| 7 |
+
#include <ATen/TensorUtils.h>
|
| 8 |
+
#include <ATen/TracerMode.h>
|
| 9 |
+
#include <ATen/core/Generator.h>
|
| 10 |
+
#include <ATen/core/Reduction.h>
|
| 11 |
+
#include <ATen/core/Tensor.h>
|
| 12 |
+
#include <c10/core/Scalar.h>
|
| 13 |
+
#include <c10/core/Storage.h>
|
| 14 |
+
#include <c10/core/TensorOptions.h>
|
| 15 |
+
#include <c10/util/Deprecated.h>
|
| 16 |
+
#include <c10/util/Optional.h>
|
| 17 |
+
|
| 18 |
+
|
| 19 |
+
|
| 20 |
+
#include <ATen/ops/_nnpack_spatial_convolution_ops.h>
|
| 21 |
+
|
| 22 |
+
namespace at {
|
| 23 |
+
|
| 24 |
+
|
| 25 |
+
// aten::_nnpack_spatial_convolution(Tensor input, Tensor weight, Tensor? bias, SymInt[2] padding, SymInt[2] stride=1) -> Tensor
|
| 26 |
+
inline at::Tensor _nnpack_spatial_convolution(const at::Tensor & input, const at::Tensor & weight, const ::std::optional<at::Tensor> & bias, at::IntArrayRef padding, at::IntArrayRef stride=1) {
|
| 27 |
+
return at::_ops::_nnpack_spatial_convolution::call(input, weight, bias, c10::fromIntArrayRefSlow(padding), c10::fromIntArrayRefSlow(stride));
|
| 28 |
+
}
|
| 29 |
+
namespace symint {
|
| 30 |
+
template <typename T, typename = std::enable_if_t<std::is_same<T, int64_t>::value>>
|
| 31 |
+
at::Tensor _nnpack_spatial_convolution(const at::Tensor & input, const at::Tensor & weight, const ::std::optional<at::Tensor> & bias, at::IntArrayRef padding, at::IntArrayRef stride=1) {
|
| 32 |
+
return at::_ops::_nnpack_spatial_convolution::call(input, weight, bias, c10::fromIntArrayRefSlow(padding), c10::fromIntArrayRefSlow(stride));
|
| 33 |
+
}
|
| 34 |
+
}
|
| 35 |
+
|
| 36 |
+
// aten::_nnpack_spatial_convolution(Tensor input, Tensor weight, Tensor? bias, SymInt[2] padding, SymInt[2] stride=1) -> Tensor
|
| 37 |
+
inline at::Tensor _nnpack_spatial_convolution_symint(const at::Tensor & input, const at::Tensor & weight, const ::std::optional<at::Tensor> & bias, c10::SymIntArrayRef padding, c10::SymIntArrayRef stride=c10::SymInt(1)) {
|
| 38 |
+
return at::_ops::_nnpack_spatial_convolution::call(input, weight, bias, padding, stride);
|
| 39 |
+
}
|
| 40 |
+
namespace symint {
|
| 41 |
+
template <typename T, typename = std::enable_if_t<std::is_same<T, c10::SymInt>::value>>
|
| 42 |
+
at::Tensor _nnpack_spatial_convolution(const at::Tensor & input, const at::Tensor & weight, const ::std::optional<at::Tensor> & bias, c10::SymIntArrayRef padding, c10::SymIntArrayRef stride=c10::SymInt(1)) {
|
| 43 |
+
return at::_ops::_nnpack_spatial_convolution::call(input, weight, bias, padding, stride);
|
| 44 |
+
}
|
| 45 |
+
}
|
| 46 |
+
|
| 47 |
+
// aten::_nnpack_spatial_convolution.out(Tensor input, Tensor weight, Tensor? bias, SymInt[2] padding, SymInt[2] stride=1, *, Tensor(a!) out) -> Tensor(a!)
|
| 48 |
+
inline at::Tensor & _nnpack_spatial_convolution_out(at::Tensor & out, const at::Tensor & input, const at::Tensor & weight, const ::std::optional<at::Tensor> & bias, at::IntArrayRef padding, at::IntArrayRef stride=1) {
|
| 49 |
+
return at::_ops::_nnpack_spatial_convolution_out::call(input, weight, bias, c10::fromIntArrayRefSlow(padding), c10::fromIntArrayRefSlow(stride), out);
|
| 50 |
+
}
|
| 51 |
+
namespace symint {
|
| 52 |
+
template <typename T, typename = std::enable_if_t<std::is_same<T, int64_t>::value>>
|
| 53 |
+
at::Tensor & _nnpack_spatial_convolution_out(at::Tensor & out, const at::Tensor & input, const at::Tensor & weight, const ::std::optional<at::Tensor> & bias, at::IntArrayRef padding, at::IntArrayRef stride=1) {
|
| 54 |
+
return at::_ops::_nnpack_spatial_convolution_out::call(input, weight, bias, c10::fromIntArrayRefSlow(padding), c10::fromIntArrayRefSlow(stride), out);
|
| 55 |
+
}
|
| 56 |
+
}
|
| 57 |
+
|
| 58 |
+
// aten::_nnpack_spatial_convolution.out(Tensor input, Tensor weight, Tensor? bias, SymInt[2] padding, SymInt[2] stride=1, *, Tensor(a!) out) -> Tensor(a!)
|
| 59 |
+
inline at::Tensor & _nnpack_spatial_convolution_outf(const at::Tensor & input, const at::Tensor & weight, const ::std::optional<at::Tensor> & bias, at::IntArrayRef padding, at::IntArrayRef stride, at::Tensor & out) {
|
| 60 |
+
return at::_ops::_nnpack_spatial_convolution_out::call(input, weight, bias, c10::fromIntArrayRefSlow(padding), c10::fromIntArrayRefSlow(stride), out);
|
| 61 |
+
}
|
| 62 |
+
namespace symint {
|
| 63 |
+
template <typename T, typename = std::enable_if_t<std::is_same<T, int64_t>::value>>
|
| 64 |
+
at::Tensor & _nnpack_spatial_convolution_outf(const at::Tensor & input, const at::Tensor & weight, const ::std::optional<at::Tensor> & bias, at::IntArrayRef padding, at::IntArrayRef stride, at::Tensor & out) {
|
| 65 |
+
return at::_ops::_nnpack_spatial_convolution_out::call(input, weight, bias, c10::fromIntArrayRefSlow(padding), c10::fromIntArrayRefSlow(stride), out);
|
| 66 |
+
}
|
| 67 |
+
}
|
| 68 |
+
|
| 69 |
+
// aten::_nnpack_spatial_convolution.out(Tensor input, Tensor weight, Tensor? bias, SymInt[2] padding, SymInt[2] stride=1, *, Tensor(a!) out) -> Tensor(a!)
|
| 70 |
+
inline at::Tensor & _nnpack_spatial_convolution_symint_out(at::Tensor & out, const at::Tensor & input, const at::Tensor & weight, const ::std::optional<at::Tensor> & bias, c10::SymIntArrayRef padding, c10::SymIntArrayRef stride=c10::SymInt(1)) {
|
| 71 |
+
return at::_ops::_nnpack_spatial_convolution_out::call(input, weight, bias, padding, stride, out);
|
| 72 |
+
}
|
| 73 |
+
namespace symint {
|
| 74 |
+
template <typename T, typename = std::enable_if_t<std::is_same<T, c10::SymInt>::value>>
|
| 75 |
+
at::Tensor & _nnpack_spatial_convolution_out(at::Tensor & out, const at::Tensor & input, const at::Tensor & weight, const ::std::optional<at::Tensor> & bias, c10::SymIntArrayRef padding, c10::SymIntArrayRef stride=c10::SymInt(1)) {
|
| 76 |
+
return at::_ops::_nnpack_spatial_convolution_out::call(input, weight, bias, padding, stride, out);
|
| 77 |
+
}
|
| 78 |
+
}
|
| 79 |
+
|
| 80 |
+
// aten::_nnpack_spatial_convolution.out(Tensor input, Tensor weight, Tensor? bias, SymInt[2] padding, SymInt[2] stride=1, *, Tensor(a!) out) -> Tensor(a!)
|
| 81 |
+
inline at::Tensor & _nnpack_spatial_convolution_symint_outf(const at::Tensor & input, const at::Tensor & weight, const ::std::optional<at::Tensor> & bias, c10::SymIntArrayRef padding, c10::SymIntArrayRef stride, at::Tensor & out) {
|
| 82 |
+
return at::_ops::_nnpack_spatial_convolution_out::call(input, weight, bias, padding, stride, out);
|
| 83 |
+
}
|
| 84 |
+
namespace symint {
|
| 85 |
+
template <typename T, typename = std::enable_if_t<std::is_same<T, c10::SymInt>::value>>
|
| 86 |
+
at::Tensor & _nnpack_spatial_convolution_outf(const at::Tensor & input, const at::Tensor & weight, const ::std::optional<at::Tensor> & bias, c10::SymIntArrayRef padding, c10::SymIntArrayRef stride, at::Tensor & out) {
|
| 87 |
+
return at::_ops::_nnpack_spatial_convolution_out::call(input, weight, bias, padding, stride, out);
|
| 88 |
+
}
|
| 89 |
+
}
|
| 90 |
+
|
| 91 |
+
}
|
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_prelu_kernel_backward_native.h
ADDED
|
@@ -0,0 +1,22 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
#pragma once
|
| 2 |
+
|
| 3 |
+
// @generated by torchgen/gen.py from NativeFunction.h
|
| 4 |
+
|
| 5 |
+
#include <c10/core/Scalar.h>
|
| 6 |
+
#include <c10/core/Storage.h>
|
| 7 |
+
#include <c10/core/TensorOptions.h>
|
| 8 |
+
#include <c10/util/Deprecated.h>
|
| 9 |
+
#include <c10/util/Optional.h>
|
| 10 |
+
#include <c10/core/QScheme.h>
|
| 11 |
+
#include <ATen/core/Reduction.h>
|
| 12 |
+
#include <ATen/core/Tensor.h>
|
| 13 |
+
#include <tuple>
|
| 14 |
+
#include <vector>
|
| 15 |
+
|
| 16 |
+
|
| 17 |
+
namespace at {
|
| 18 |
+
namespace native {
|
| 19 |
+
TORCH_API ::std::tuple<at::Tensor,at::Tensor> _prelu_kernel_backward(const at::Tensor & grad_output, const at::Tensor & self, const at::Tensor & weight);
|
| 20 |
+
TORCH_API ::std::tuple<at::Tensor,at::Tensor> mkldnn_prelu_backward(const at::Tensor & grad_output, const at::Tensor & self, const at::Tensor & weight);
|
| 21 |
+
} // namespace native
|
| 22 |
+
} // namespace at
|
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_sparse_semi_structured_apply_native.h
ADDED
|
@@ -0,0 +1,21 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
#pragma once
|
| 2 |
+
|
| 3 |
+
// @generated by torchgen/gen.py from NativeFunction.h
|
| 4 |
+
|
| 5 |
+
#include <c10/core/Scalar.h>
|
| 6 |
+
#include <c10/core/Storage.h>
|
| 7 |
+
#include <c10/core/TensorOptions.h>
|
| 8 |
+
#include <c10/util/Deprecated.h>
|
| 9 |
+
#include <c10/util/Optional.h>
|
| 10 |
+
#include <c10/core/QScheme.h>
|
| 11 |
+
#include <ATen/core/Reduction.h>
|
| 12 |
+
#include <ATen/core/Tensor.h>
|
| 13 |
+
#include <tuple>
|
| 14 |
+
#include <vector>
|
| 15 |
+
|
| 16 |
+
|
| 17 |
+
namespace at {
|
| 18 |
+
namespace native {
|
| 19 |
+
TORCH_API ::std::tuple<at::Tensor,at::Tensor> _sparse_semi_structured_apply(const at::Tensor & input, const at::Tensor & thread_masks);
|
| 20 |
+
} // namespace native
|
| 21 |
+
} // namespace at
|
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_unsafe_index_put_compositeexplicitautograd_dispatch.h
ADDED
|
@@ -0,0 +1,23 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
#pragma once
|
| 2 |
+
// @generated by torchgen/gen.py from DispatchKeyFunction.h
|
| 3 |
+
|
| 4 |
+
// NB: The implementing C++ file is RegisterDispatchKey.cpp
|
| 5 |
+
|
| 6 |
+
// The only #includes we need are for custom classes that have defaults in the C++ API
|
| 7 |
+
#include <c10/core/MemoryFormat.h>
|
| 8 |
+
#include <c10/core/Scalar.h>
|
| 9 |
+
#include <ATen/core/Reduction.h>
|
| 10 |
+
|
| 11 |
+
// Forward declarations of any types needed in the operator signatures.
|
| 12 |
+
// We can't directly include these classes because it will cause circular include dependencies.
|
| 13 |
+
// This file is included by TensorBody.h, which defines the Tensor class.
|
| 14 |
+
#include <ATen/core/ATen_fwd.h>
|
| 15 |
+
|
| 16 |
+
namespace at {
|
| 17 |
+
|
| 18 |
+
namespace compositeexplicitautograd {
|
| 19 |
+
|
| 20 |
+
TORCH_API at::Tensor _unsafe_index_put(const at::Tensor & self, const c10::List<::std::optional<at::Tensor>> & indices, const at::Tensor & values, bool accumulate=false);
|
| 21 |
+
|
| 22 |
+
} // namespace compositeexplicitautograd
|
| 23 |
+
} // namespace at
|
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_upsample_bicubic2d_aa_backward_meta_dispatch.h
ADDED
|
@@ -0,0 +1,28 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
#pragma once
|
| 2 |
+
// @generated by torchgen/gen.py from DispatchKeyFunction.h
|
| 3 |
+
|
| 4 |
+
// NB: The implementing C++ file is RegisterDispatchKey.cpp
|
| 5 |
+
|
| 6 |
+
// The only #includes we need are for custom classes that have defaults in the C++ API
|
| 7 |
+
#include <c10/core/MemoryFormat.h>
|
| 8 |
+
#include <c10/core/Scalar.h>
|
| 9 |
+
#include <ATen/core/Reduction.h>
|
| 10 |
+
|
| 11 |
+
// Forward declarations of any types needed in the operator signatures.
|
| 12 |
+
// We can't directly include these classes because it will cause circular include dependencies.
|
| 13 |
+
// This file is included by TensorBody.h, which defines the Tensor class.
|
| 14 |
+
#include <ATen/core/ATen_fwd.h>
|
| 15 |
+
|
| 16 |
+
namespace at {
|
| 17 |
+
|
| 18 |
+
namespace meta {
|
| 19 |
+
|
| 20 |
+
TORCH_API at::Tensor _upsample_bicubic2d_aa_backward(const at::Tensor & grad_output, at::IntArrayRef output_size, at::IntArrayRef input_size, bool align_corners, ::std::optional<double> scales_h=::std::nullopt, ::std::optional<double> scales_w=::std::nullopt);
|
| 21 |
+
TORCH_API at::Tensor _upsample_bicubic2d_aa_backward_symint(const at::Tensor & grad_output, c10::SymIntArrayRef output_size, c10::SymIntArrayRef input_size, bool align_corners, ::std::optional<double> scales_h=::std::nullopt, ::std::optional<double> scales_w=::std::nullopt);
|
| 22 |
+
TORCH_API at::Tensor & _upsample_bicubic2d_aa_backward_out(at::Tensor & grad_input, const at::Tensor & grad_output, at::IntArrayRef output_size, at::IntArrayRef input_size, bool align_corners, ::std::optional<double> scales_h=::std::nullopt, ::std::optional<double> scales_w=::std::nullopt);
|
| 23 |
+
TORCH_API at::Tensor & _upsample_bicubic2d_aa_backward_outf(const at::Tensor & grad_output, at::IntArrayRef output_size, at::IntArrayRef input_size, bool align_corners, ::std::optional<double> scales_h, ::std::optional<double> scales_w, at::Tensor & grad_input);
|
| 24 |
+
TORCH_API at::Tensor & _upsample_bicubic2d_aa_backward_symint_out(at::Tensor & grad_input, const at::Tensor & grad_output, c10::SymIntArrayRef output_size, c10::SymIntArrayRef input_size, bool align_corners, ::std::optional<double> scales_h=::std::nullopt, ::std::optional<double> scales_w=::std::nullopt);
|
| 25 |
+
TORCH_API at::Tensor & _upsample_bicubic2d_aa_backward_symint_outf(const at::Tensor & grad_output, c10::SymIntArrayRef output_size, c10::SymIntArrayRef input_size, bool align_corners, ::std::optional<double> scales_h, ::std::optional<double> scales_w, at::Tensor & grad_input);
|
| 26 |
+
|
| 27 |
+
} // namespace meta
|
| 28 |
+
} // namespace at
|
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/_upsample_bilinear2d_aa_meta.h
ADDED
|
@@ -0,0 +1,27 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
#pragma once
|
| 2 |
+
|
| 3 |
+
// @generated by torchgen/gen.py from NativeMetaFunction.h
|
| 4 |
+
|
| 5 |
+
#include <c10/core/Scalar.h>
|
| 6 |
+
#include <c10/core/Storage.h>
|
| 7 |
+
#include <c10/core/TensorOptions.h>
|
| 8 |
+
#include <c10/util/Deprecated.h>
|
| 9 |
+
#include <c10/util/Optional.h>
|
| 10 |
+
#include <c10/core/QScheme.h>
|
| 11 |
+
#include <ATen/core/Reduction.h>
|
| 12 |
+
#include <ATen/TensorIterator.h>
|
| 13 |
+
#include <ATen/TensorMeta.h>
|
| 14 |
+
#include <tuple>
|
| 15 |
+
#include <vector>
|
| 16 |
+
|
| 17 |
+
namespace at {
|
| 18 |
+
namespace meta {
|
| 19 |
+
|
| 20 |
+
struct TORCH_API structured__upsample_bilinear2d_aa : public at::impl::MetaBase {
|
| 21 |
+
|
| 22 |
+
|
| 23 |
+
void meta(const at::Tensor & self, at::ArrayRef<int64_t> output_size, bool align_corners, ::std::optional<double> scales_h, ::std::optional<double> scales_w);
|
| 24 |
+
};
|
| 25 |
+
|
| 26 |
+
} // namespace native
|
| 27 |
+
} // namespace at
|
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/adaptive_avg_pool3d.h
ADDED
|
@@ -0,0 +1,91 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
#pragma once
|
| 2 |
+
|
| 3 |
+
// @generated by torchgen/gen.py from Function.h
|
| 4 |
+
|
| 5 |
+
#include <ATen/Context.h>
|
| 6 |
+
#include <ATen/DeviceGuard.h>
|
| 7 |
+
#include <ATen/TensorUtils.h>
|
| 8 |
+
#include <ATen/TracerMode.h>
|
| 9 |
+
#include <ATen/core/Generator.h>
|
| 10 |
+
#include <ATen/core/Reduction.h>
|
| 11 |
+
#include <ATen/core/Tensor.h>
|
| 12 |
+
#include <c10/core/Scalar.h>
|
| 13 |
+
#include <c10/core/Storage.h>
|
| 14 |
+
#include <c10/core/TensorOptions.h>
|
| 15 |
+
#include <c10/util/Deprecated.h>
|
| 16 |
+
#include <c10/util/Optional.h>
|
| 17 |
+
|
| 18 |
+
|
| 19 |
+
|
| 20 |
+
#include <ATen/ops/adaptive_avg_pool3d_ops.h>
|
| 21 |
+
|
| 22 |
+
namespace at {
|
| 23 |
+
|
| 24 |
+
|
| 25 |
+
// aten::adaptive_avg_pool3d.out(Tensor self, SymInt[3] output_size, *, Tensor(a!) out) -> Tensor(a!)
|
| 26 |
+
inline at::Tensor & adaptive_avg_pool3d_out(at::Tensor & out, const at::Tensor & self, at::IntArrayRef output_size) {
|
| 27 |
+
return at::_ops::adaptive_avg_pool3d_out::call(self, c10::fromIntArrayRefSlow(output_size), out);
|
| 28 |
+
}
|
| 29 |
+
namespace symint {
|
| 30 |
+
template <typename T, typename = std::enable_if_t<std::is_same<T, int64_t>::value>>
|
| 31 |
+
at::Tensor & adaptive_avg_pool3d_out(at::Tensor & out, const at::Tensor & self, at::IntArrayRef output_size) {
|
| 32 |
+
return at::_ops::adaptive_avg_pool3d_out::call(self, c10::fromIntArrayRefSlow(output_size), out);
|
| 33 |
+
}
|
| 34 |
+
}
|
| 35 |
+
|
| 36 |
+
// aten::adaptive_avg_pool3d.out(Tensor self, SymInt[3] output_size, *, Tensor(a!) out) -> Tensor(a!)
|
| 37 |
+
inline at::Tensor & adaptive_avg_pool3d_outf(const at::Tensor & self, at::IntArrayRef output_size, at::Tensor & out) {
|
| 38 |
+
return at::_ops::adaptive_avg_pool3d_out::call(self, c10::fromIntArrayRefSlow(output_size), out);
|
| 39 |
+
}
|
| 40 |
+
namespace symint {
|
| 41 |
+
template <typename T, typename = std::enable_if_t<std::is_same<T, int64_t>::value>>
|
| 42 |
+
at::Tensor & adaptive_avg_pool3d_outf(const at::Tensor & self, at::IntArrayRef output_size, at::Tensor & out) {
|
| 43 |
+
return at::_ops::adaptive_avg_pool3d_out::call(self, c10::fromIntArrayRefSlow(output_size), out);
|
| 44 |
+
}
|
| 45 |
+
}
|
| 46 |
+
|
| 47 |
+
// aten::adaptive_avg_pool3d.out(Tensor self, SymInt[3] output_size, *, Tensor(a!) out) -> Tensor(a!)
|
| 48 |
+
inline at::Tensor & adaptive_avg_pool3d_symint_out(at::Tensor & out, const at::Tensor & self, c10::SymIntArrayRef output_size) {
|
| 49 |
+
return at::_ops::adaptive_avg_pool3d_out::call(self, output_size, out);
|
| 50 |
+
}
|
| 51 |
+
namespace symint {
|
| 52 |
+
template <typename T, typename = std::enable_if_t<std::is_same<T, c10::SymInt>::value>>
|
| 53 |
+
at::Tensor & adaptive_avg_pool3d_out(at::Tensor & out, const at::Tensor & self, c10::SymIntArrayRef output_size) {
|
| 54 |
+
return at::_ops::adaptive_avg_pool3d_out::call(self, output_size, out);
|
| 55 |
+
}
|
| 56 |
+
}
|
| 57 |
+
|
| 58 |
+
// aten::adaptive_avg_pool3d.out(Tensor self, SymInt[3] output_size, *, Tensor(a!) out) -> Tensor(a!)
|
| 59 |
+
inline at::Tensor & adaptive_avg_pool3d_symint_outf(const at::Tensor & self, c10::SymIntArrayRef output_size, at::Tensor & out) {
|
| 60 |
+
return at::_ops::adaptive_avg_pool3d_out::call(self, output_size, out);
|
| 61 |
+
}
|
| 62 |
+
namespace symint {
|
| 63 |
+
template <typename T, typename = std::enable_if_t<std::is_same<T, c10::SymInt>::value>>
|
| 64 |
+
at::Tensor & adaptive_avg_pool3d_outf(const at::Tensor & self, c10::SymIntArrayRef output_size, at::Tensor & out) {
|
| 65 |
+
return at::_ops::adaptive_avg_pool3d_out::call(self, output_size, out);
|
| 66 |
+
}
|
| 67 |
+
}
|
| 68 |
+
|
| 69 |
+
// aten::adaptive_avg_pool3d(Tensor self, SymInt[3] output_size) -> Tensor
|
| 70 |
+
inline at::Tensor adaptive_avg_pool3d(const at::Tensor & self, at::IntArrayRef output_size) {
|
| 71 |
+
return at::_ops::adaptive_avg_pool3d::call(self, c10::fromIntArrayRefSlow(output_size));
|
| 72 |
+
}
|
| 73 |
+
namespace symint {
|
| 74 |
+
template <typename T, typename = std::enable_if_t<std::is_same<T, int64_t>::value>>
|
| 75 |
+
at::Tensor adaptive_avg_pool3d(const at::Tensor & self, at::IntArrayRef output_size) {
|
| 76 |
+
return at::_ops::adaptive_avg_pool3d::call(self, c10::fromIntArrayRefSlow(output_size));
|
| 77 |
+
}
|
| 78 |
+
}
|
| 79 |
+
|
| 80 |
+
// aten::adaptive_avg_pool3d(Tensor self, SymInt[3] output_size) -> Tensor
|
| 81 |
+
inline at::Tensor adaptive_avg_pool3d_symint(const at::Tensor & self, c10::SymIntArrayRef output_size) {
|
| 82 |
+
return at::_ops::adaptive_avg_pool3d::call(self, output_size);
|
| 83 |
+
}
|
| 84 |
+
namespace symint {
|
| 85 |
+
template <typename T, typename = std::enable_if_t<std::is_same<T, c10::SymInt>::value>>
|
| 86 |
+
at::Tensor adaptive_avg_pool3d(const at::Tensor & self, c10::SymIntArrayRef output_size) {
|
| 87 |
+
return at::_ops::adaptive_avg_pool3d::call(self, output_size);
|
| 88 |
+
}
|
| 89 |
+
}
|
| 90 |
+
|
| 91 |
+
}
|
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/adaptive_max_pool2d_backward_meta.h
ADDED
|
@@ -0,0 +1,27 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
#pragma once
|
| 2 |
+
|
| 3 |
+
// @generated by torchgen/gen.py from NativeMetaFunction.h
|
| 4 |
+
|
| 5 |
+
#include <c10/core/Scalar.h>
|
| 6 |
+
#include <c10/core/Storage.h>
|
| 7 |
+
#include <c10/core/TensorOptions.h>
|
| 8 |
+
#include <c10/util/Deprecated.h>
|
| 9 |
+
#include <c10/util/Optional.h>
|
| 10 |
+
#include <c10/core/QScheme.h>
|
| 11 |
+
#include <ATen/core/Reduction.h>
|
| 12 |
+
#include <ATen/TensorIterator.h>
|
| 13 |
+
#include <ATen/TensorMeta.h>
|
| 14 |
+
#include <tuple>
|
| 15 |
+
#include <vector>
|
| 16 |
+
|
| 17 |
+
namespace at {
|
| 18 |
+
namespace meta {
|
| 19 |
+
|
| 20 |
+
struct TORCH_API structured_adaptive_max_pool2d_backward : public at::impl::MetaBase {
|
| 21 |
+
|
| 22 |
+
|
| 23 |
+
void meta(const at::Tensor & grad_output, const at::Tensor & self, const at::Tensor & indices);
|
| 24 |
+
};
|
| 25 |
+
|
| 26 |
+
} // namespace native
|
| 27 |
+
} // namespace at
|
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/atan2_meta_dispatch.h
ADDED
|
@@ -0,0 +1,26 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
#pragma once
|
| 2 |
+
// @generated by torchgen/gen.py from DispatchKeyFunction.h
|
| 3 |
+
|
| 4 |
+
// NB: The implementing C++ file is RegisterDispatchKey.cpp
|
| 5 |
+
|
| 6 |
+
// The only #includes we need are for custom classes that have defaults in the C++ API
|
| 7 |
+
#include <c10/core/MemoryFormat.h>
|
| 8 |
+
#include <c10/core/Scalar.h>
|
| 9 |
+
#include <ATen/core/Reduction.h>
|
| 10 |
+
|
| 11 |
+
// Forward declarations of any types needed in the operator signatures.
|
| 12 |
+
// We can't directly include these classes because it will cause circular include dependencies.
|
| 13 |
+
// This file is included by TensorBody.h, which defines the Tensor class.
|
| 14 |
+
#include <ATen/core/ATen_fwd.h>
|
| 15 |
+
|
| 16 |
+
namespace at {
|
| 17 |
+
|
| 18 |
+
namespace meta {
|
| 19 |
+
|
| 20 |
+
TORCH_API at::Tensor atan2(const at::Tensor & self, const at::Tensor & other);
|
| 21 |
+
TORCH_API at::Tensor & atan2_out(at::Tensor & out, const at::Tensor & self, const at::Tensor & other);
|
| 22 |
+
TORCH_API at::Tensor & atan2_outf(const at::Tensor & self, const at::Tensor & other, at::Tensor & out);
|
| 23 |
+
TORCH_API at::Tensor & atan2_(at::Tensor & self, const at::Tensor & other);
|
| 24 |
+
|
| 25 |
+
} // namespace meta
|
| 26 |
+
} // namespace at
|
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/batch_norm_gather_stats_compositeexplicitautograd_dispatch.h
ADDED
|
@@ -0,0 +1,24 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
#pragma once
|
| 2 |
+
// @generated by torchgen/gen.py from DispatchKeyFunction.h
|
| 3 |
+
|
| 4 |
+
// NB: The implementing C++ file is RegisterDispatchKey.cpp
|
| 5 |
+
|
| 6 |
+
// The only #includes we need are for custom classes that have defaults in the C++ API
|
| 7 |
+
#include <c10/core/MemoryFormat.h>
|
| 8 |
+
#include <c10/core/Scalar.h>
|
| 9 |
+
#include <ATen/core/Reduction.h>
|
| 10 |
+
|
| 11 |
+
// Forward declarations of any types needed in the operator signatures.
|
| 12 |
+
// We can't directly include these classes because it will cause circular include dependencies.
|
| 13 |
+
// This file is included by TensorBody.h, which defines the Tensor class.
|
| 14 |
+
#include <ATen/core/ATen_fwd.h>
|
| 15 |
+
|
| 16 |
+
namespace at {
|
| 17 |
+
|
| 18 |
+
namespace compositeexplicitautograd {
|
| 19 |
+
|
| 20 |
+
TORCH_API ::std::tuple<at::Tensor &,at::Tensor &> batch_norm_gather_stats_out(at::Tensor & out0, at::Tensor & out1, const at::Tensor & input, const at::Tensor & mean, const at::Tensor & invstd, const ::std::optional<at::Tensor> & running_mean, const ::std::optional<at::Tensor> & running_var, double momentum, double eps, int64_t count);
|
| 21 |
+
TORCH_API ::std::tuple<at::Tensor &,at::Tensor &> batch_norm_gather_stats_outf(const at::Tensor & input, const at::Tensor & mean, const at::Tensor & invstd, const ::std::optional<at::Tensor> & running_mean, const ::std::optional<at::Tensor> & running_var, double momentum, double eps, int64_t count, at::Tensor & out0, at::Tensor & out1);
|
| 22 |
+
|
| 23 |
+
} // namespace compositeexplicitautograd
|
| 24 |
+
} // namespace at
|
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/batch_norm_gather_stats_with_counts_compositeexplicitautograd_dispatch.h
ADDED
|
@@ -0,0 +1,24 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
#pragma once
|
| 2 |
+
// @generated by torchgen/gen.py from DispatchKeyFunction.h
|
| 3 |
+
|
| 4 |
+
// NB: The implementing C++ file is RegisterDispatchKey.cpp
|
| 5 |
+
|
| 6 |
+
// The only #includes we need are for custom classes that have defaults in the C++ API
|
| 7 |
+
#include <c10/core/MemoryFormat.h>
|
| 8 |
+
#include <c10/core/Scalar.h>
|
| 9 |
+
#include <ATen/core/Reduction.h>
|
| 10 |
+
|
| 11 |
+
// Forward declarations of any types needed in the operator signatures.
|
| 12 |
+
// We can't directly include these classes because it will cause circular include dependencies.
|
| 13 |
+
// This file is included by TensorBody.h, which defines the Tensor class.
|
| 14 |
+
#include <ATen/core/ATen_fwd.h>
|
| 15 |
+
|
| 16 |
+
namespace at {
|
| 17 |
+
|
| 18 |
+
namespace compositeexplicitautograd {
|
| 19 |
+
|
| 20 |
+
TORCH_API ::std::tuple<at::Tensor &,at::Tensor &> batch_norm_gather_stats_with_counts_out(at::Tensor & out0, at::Tensor & out1, const at::Tensor & input, const at::Tensor & mean, const at::Tensor & invstd, const ::std::optional<at::Tensor> & running_mean, const ::std::optional<at::Tensor> & running_var, double momentum, double eps, const at::Tensor & counts);
|
| 21 |
+
TORCH_API ::std::tuple<at::Tensor &,at::Tensor &> batch_norm_gather_stats_with_counts_outf(const at::Tensor & input, const at::Tensor & mean, const at::Tensor & invstd, const ::std::optional<at::Tensor> & running_mean, const ::std::optional<at::Tensor> & running_var, double momentum, double eps, const at::Tensor & counts, at::Tensor & out0, at::Tensor & out1);
|
| 22 |
+
|
| 23 |
+
} // namespace compositeexplicitautograd
|
| 24 |
+
} // namespace at
|
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/complex_cuda_dispatch.h
ADDED
|
@@ -0,0 +1,24 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
#pragma once
|
| 2 |
+
// @generated by torchgen/gen.py from DispatchKeyFunction.h
|
| 3 |
+
|
| 4 |
+
// NB: The implementing C++ file is RegisterDispatchKey.cpp
|
| 5 |
+
|
| 6 |
+
// The only #includes we need are for custom classes that have defaults in the C++ API
|
| 7 |
+
#include <c10/core/MemoryFormat.h>
|
| 8 |
+
#include <c10/core/Scalar.h>
|
| 9 |
+
#include <ATen/core/Reduction.h>
|
| 10 |
+
|
| 11 |
+
// Forward declarations of any types needed in the operator signatures.
|
| 12 |
+
// We can't directly include these classes because it will cause circular include dependencies.
|
| 13 |
+
// This file is included by TensorBody.h, which defines the Tensor class.
|
| 14 |
+
#include <ATen/core/ATen_fwd.h>
|
| 15 |
+
|
| 16 |
+
namespace at {
|
| 17 |
+
|
| 18 |
+
namespace cuda {
|
| 19 |
+
|
| 20 |
+
TORCH_API at::Tensor & complex_out(at::Tensor & out, const at::Tensor & real, const at::Tensor & imag);
|
| 21 |
+
TORCH_API at::Tensor & complex_outf(const at::Tensor & real, const at::Tensor & imag, at::Tensor & out);
|
| 22 |
+
|
| 23 |
+
} // namespace cuda
|
| 24 |
+
} // namespace at
|
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/conv1d_ops.h
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
#pragma once
|
| 2 |
+
|
| 3 |
+
// @generated by torchgen/gen.py from Operator.h
|
| 4 |
+
|
| 5 |
+
#include <tuple>
|
| 6 |
+
#include <vector>
|
| 7 |
+
|
| 8 |
+
// Forward declarations of any types needed in the operator signatures.
|
| 9 |
+
// We can't directly include these classes because it will cause circular include dependencies.
|
| 10 |
+
// This file is included by TensorBody.h, which defines the Tensor class.
|
| 11 |
+
#include <ATen/core/ATen_fwd.h>
|
| 12 |
+
|
| 13 |
+
namespace at {
|
| 14 |
+
namespace _ops {
|
| 15 |
+
|
| 16 |
+
|
| 17 |
+
struct TORCH_API conv1d {
|
| 18 |
+
using schema = at::Tensor (const at::Tensor &, const at::Tensor &, const ::std::optional<at::Tensor> &, c10::SymIntArrayRef, c10::SymIntArrayRef, c10::SymIntArrayRef, c10::SymInt);
|
| 19 |
+
using ptr_schema = schema*;
|
| 20 |
+
// See Note [static constexpr char* members for windows NVCC]
|
| 21 |
+
STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(name, "aten::conv1d")
|
| 22 |
+
STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(overload_name, "")
|
| 23 |
+
STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(schema_str, "conv1d(Tensor input, Tensor weight, Tensor? bias=None, SymInt[1] stride=1, SymInt[1] padding=0, SymInt[1] dilation=1, SymInt groups=1) -> Tensor")
|
| 24 |
+
static at::Tensor call(const at::Tensor & input, const at::Tensor & weight, const ::std::optional<at::Tensor> & bias, c10::SymIntArrayRef stride, c10::SymIntArrayRef padding, c10::SymIntArrayRef dilation, c10::SymInt groups);
|
| 25 |
+
static at::Tensor redispatch(c10::DispatchKeySet dispatchKeySet, const at::Tensor & input, const at::Tensor & weight, const ::std::optional<at::Tensor> & bias, c10::SymIntArrayRef stride, c10::SymIntArrayRef padding, c10::SymIntArrayRef dilation, c10::SymInt groups);
|
| 26 |
+
};
|
| 27 |
+
|
| 28 |
+
struct TORCH_API conv1d_padding {
|
| 29 |
+
using schema = at::Tensor (const at::Tensor &, const at::Tensor &, const ::std::optional<at::Tensor> &, c10::SymIntArrayRef, c10::string_view, c10::SymIntArrayRef, c10::SymInt);
|
| 30 |
+
using ptr_schema = schema*;
|
| 31 |
+
// See Note [static constexpr char* members for windows NVCC]
|
| 32 |
+
STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(name, "aten::conv1d")
|
| 33 |
+
STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(overload_name, "padding")
|
| 34 |
+
STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(schema_str, "conv1d.padding(Tensor input, Tensor weight, Tensor? bias=None, SymInt[1] stride=1, str padding=\"valid\", SymInt[1] dilation=1, SymInt groups=1) -> Tensor")
|
| 35 |
+
static at::Tensor call(const at::Tensor & input, const at::Tensor & weight, const ::std::optional<at::Tensor> & bias, c10::SymIntArrayRef stride, c10::string_view padding, c10::SymIntArrayRef dilation, c10::SymInt groups);
|
| 36 |
+
static at::Tensor redispatch(c10::DispatchKeySet dispatchKeySet, const at::Tensor & input, const at::Tensor & weight, const ::std::optional<at::Tensor> & bias, c10::SymIntArrayRef stride, c10::string_view padding, c10::SymIntArrayRef dilation, c10::SymInt groups);
|
| 37 |
+
};
|
| 38 |
+
|
| 39 |
+
}} // namespace at::_ops
|
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/cudnn_grid_sampler_ops.h
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
#pragma once
|
| 2 |
+
|
| 3 |
+
// @generated by torchgen/gen.py from Operator.h
|
| 4 |
+
|
| 5 |
+
#include <tuple>
|
| 6 |
+
#include <vector>
|
| 7 |
+
|
| 8 |
+
// Forward declarations of any types needed in the operator signatures.
|
| 9 |
+
// We can't directly include these classes because it will cause circular include dependencies.
|
| 10 |
+
// This file is included by TensorBody.h, which defines the Tensor class.
|
| 11 |
+
#include <ATen/core/ATen_fwd.h>
|
| 12 |
+
|
| 13 |
+
namespace at {
|
| 14 |
+
namespace _ops {
|
| 15 |
+
|
| 16 |
+
|
| 17 |
+
struct TORCH_API cudnn_grid_sampler {
|
| 18 |
+
using schema = at::Tensor (const at::Tensor &, const at::Tensor &);
|
| 19 |
+
using ptr_schema = schema*;
|
| 20 |
+
// See Note [static constexpr char* members for windows NVCC]
|
| 21 |
+
STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(name, "aten::cudnn_grid_sampler")
|
| 22 |
+
STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(overload_name, "")
|
| 23 |
+
STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(schema_str, "cudnn_grid_sampler(Tensor self, Tensor grid) -> Tensor output")
|
| 24 |
+
static at::Tensor call(const at::Tensor & self, const at::Tensor & grid);
|
| 25 |
+
static at::Tensor redispatch(c10::DispatchKeySet dispatchKeySet, const at::Tensor & self, const at::Tensor & grid);
|
| 26 |
+
};
|
| 27 |
+
|
| 28 |
+
struct TORCH_API cudnn_grid_sampler_out {
|
| 29 |
+
using schema = at::Tensor & (const at::Tensor &, const at::Tensor &, at::Tensor &);
|
| 30 |
+
using ptr_schema = schema*;
|
| 31 |
+
// See Note [static constexpr char* members for windows NVCC]
|
| 32 |
+
STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(name, "aten::cudnn_grid_sampler")
|
| 33 |
+
STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(overload_name, "out")
|
| 34 |
+
STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(schema_str, "cudnn_grid_sampler.out(Tensor self, Tensor grid, *, Tensor(a!) out) -> Tensor(a!)")
|
| 35 |
+
static at::Tensor & call(const at::Tensor & self, const at::Tensor & grid, at::Tensor & out);
|
| 36 |
+
static at::Tensor & redispatch(c10::DispatchKeySet dispatchKeySet, const at::Tensor & self, const at::Tensor & grid, at::Tensor & out);
|
| 37 |
+
};
|
| 38 |
+
|
| 39 |
+
}} // namespace at::_ops
|
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/cummax.h
ADDED
|
@@ -0,0 +1,53 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
#pragma once
|
| 2 |
+
|
| 3 |
+
// @generated by torchgen/gen.py from Function.h
|
| 4 |
+
|
| 5 |
+
#include <ATen/Context.h>
|
| 6 |
+
#include <ATen/DeviceGuard.h>
|
| 7 |
+
#include <ATen/TensorUtils.h>
|
| 8 |
+
#include <ATen/TracerMode.h>
|
| 9 |
+
#include <ATen/core/Generator.h>
|
| 10 |
+
#include <ATen/core/Reduction.h>
|
| 11 |
+
#include <ATen/core/Tensor.h>
|
| 12 |
+
#include <c10/core/Scalar.h>
|
| 13 |
+
#include <c10/core/Storage.h>
|
| 14 |
+
#include <c10/core/TensorOptions.h>
|
| 15 |
+
#include <c10/util/Deprecated.h>
|
| 16 |
+
#include <c10/util/Optional.h>
|
| 17 |
+
|
| 18 |
+
|
| 19 |
+
|
| 20 |
+
#include <ATen/ops/cummax_ops.h>
|
| 21 |
+
|
| 22 |
+
namespace at {
|
| 23 |
+
|
| 24 |
+
|
| 25 |
+
// aten::cummax(Tensor self, int dim) -> (Tensor values, Tensor indices)
|
| 26 |
+
inline ::std::tuple<at::Tensor,at::Tensor> cummax(const at::Tensor & self, int64_t dim) {
|
| 27 |
+
return at::_ops::cummax::call(self, dim);
|
| 28 |
+
}
|
| 29 |
+
|
| 30 |
+
// aten::cummax.out(Tensor self, int dim, *, Tensor(a!) values, Tensor(b!) indices) -> (Tensor(a!) values, Tensor(b!) indices)
|
| 31 |
+
inline ::std::tuple<at::Tensor &,at::Tensor &> cummax_out(at::Tensor & values, at::Tensor & indices, const at::Tensor & self, int64_t dim) {
|
| 32 |
+
return at::_ops::cummax_out::call(self, dim, values, indices);
|
| 33 |
+
}
|
| 34 |
+
// aten::cummax.out(Tensor self, int dim, *, Tensor(a!) values, Tensor(b!) indices) -> (Tensor(a!) values, Tensor(b!) indices)
|
| 35 |
+
inline ::std::tuple<at::Tensor &,at::Tensor &> cummax_outf(const at::Tensor & self, int64_t dim, at::Tensor & values, at::Tensor & indices) {
|
| 36 |
+
return at::_ops::cummax_out::call(self, dim, values, indices);
|
| 37 |
+
}
|
| 38 |
+
|
| 39 |
+
// aten::cummax.dimname(Tensor self, Dimname dim) -> (Tensor values, Tensor indices)
|
| 40 |
+
inline ::std::tuple<at::Tensor,at::Tensor> cummax(const at::Tensor & self, at::Dimname dim) {
|
| 41 |
+
return at::_ops::cummax_dimname::call(self, dim);
|
| 42 |
+
}
|
| 43 |
+
|
| 44 |
+
// aten::cummax.dimname_out(Tensor self, Dimname dim, *, Tensor(a!) values, Tensor(b!) indices) -> (Tensor(a!) values, Tensor(b!) indices)
|
| 45 |
+
inline ::std::tuple<at::Tensor &,at::Tensor &> cummax_out(at::Tensor & values, at::Tensor & indices, const at::Tensor & self, at::Dimname dim) {
|
| 46 |
+
return at::_ops::cummax_dimname_out::call(self, dim, values, indices);
|
| 47 |
+
}
|
| 48 |
+
// aten::cummax.dimname_out(Tensor self, Dimname dim, *, Tensor(a!) values, Tensor(b!) indices) -> (Tensor(a!) values, Tensor(b!) indices)
|
| 49 |
+
inline ::std::tuple<at::Tensor &,at::Tensor &> cummax_outf(const at::Tensor & self, at::Dimname dim, at::Tensor & values, at::Tensor & indices) {
|
| 50 |
+
return at::_ops::cummax_dimname_out::call(self, dim, values, indices);
|
| 51 |
+
}
|
| 52 |
+
|
| 53 |
+
}
|
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/diagonal_native.h
ADDED
|
@@ -0,0 +1,22 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
#pragma once
|
| 2 |
+
|
| 3 |
+
// @generated by torchgen/gen.py from NativeFunction.h
|
| 4 |
+
|
| 5 |
+
#include <c10/core/Scalar.h>
|
| 6 |
+
#include <c10/core/Storage.h>
|
| 7 |
+
#include <c10/core/TensorOptions.h>
|
| 8 |
+
#include <c10/util/Deprecated.h>
|
| 9 |
+
#include <c10/util/Optional.h>
|
| 10 |
+
#include <c10/core/QScheme.h>
|
| 11 |
+
#include <ATen/core/Reduction.h>
|
| 12 |
+
#include <ATen/core/Tensor.h>
|
| 13 |
+
#include <tuple>
|
| 14 |
+
#include <vector>
|
| 15 |
+
|
| 16 |
+
|
| 17 |
+
namespace at {
|
| 18 |
+
namespace native {
|
| 19 |
+
TORCH_API at::Tensor diagonal(const at::Tensor & self, int64_t offset=0, int64_t dim1=0, int64_t dim2=1);
|
| 20 |
+
TORCH_API at::Tensor diagonal(const at::Tensor & self, at::Dimname outdim, at::Dimname dim1, at::Dimname dim2, int64_t offset=0);
|
| 21 |
+
} // namespace native
|
| 22 |
+
} // namespace at
|
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/diff_compositeimplicitautograd_dispatch.h
ADDED
|
@@ -0,0 +1,25 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
#pragma once
|
| 2 |
+
// @generated by torchgen/gen.py from DispatchKeyFunction.h
|
| 3 |
+
|
| 4 |
+
// NB: The implementing C++ file is RegisterDispatchKey.cpp
|
| 5 |
+
|
| 6 |
+
// The only #includes we need are for custom classes that have defaults in the C++ API
|
| 7 |
+
#include <c10/core/MemoryFormat.h>
|
| 8 |
+
#include <c10/core/Scalar.h>
|
| 9 |
+
#include <ATen/core/Reduction.h>
|
| 10 |
+
|
| 11 |
+
// Forward declarations of any types needed in the operator signatures.
|
| 12 |
+
// We can't directly include these classes because it will cause circular include dependencies.
|
| 13 |
+
// This file is included by TensorBody.h, which defines the Tensor class.
|
| 14 |
+
#include <ATen/core/ATen_fwd.h>
|
| 15 |
+
|
| 16 |
+
namespace at {
|
| 17 |
+
|
| 18 |
+
namespace compositeimplicitautograd {
|
| 19 |
+
|
| 20 |
+
TORCH_API at::Tensor diff(const at::Tensor & self, int64_t n=1, int64_t dim=-1, const ::std::optional<at::Tensor> & prepend={}, const ::std::optional<at::Tensor> & append={});
|
| 21 |
+
TORCH_API at::Tensor & diff_out(at::Tensor & out, const at::Tensor & self, int64_t n=1, int64_t dim=-1, const ::std::optional<at::Tensor> & prepend={}, const ::std::optional<at::Tensor> & append={});
|
| 22 |
+
TORCH_API at::Tensor & diff_outf(const at::Tensor & self, int64_t n, int64_t dim, const ::std::optional<at::Tensor> & prepend, const ::std::optional<at::Tensor> & append, at::Tensor & out);
|
| 23 |
+
|
| 24 |
+
} // namespace compositeimplicitautograd
|
| 25 |
+
} // namespace at
|
parrot/lib/python3.10/site-packages/torch/include/ATen/ops/embedding_dense_backward.h
ADDED
|
@@ -0,0 +1,91 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
#pragma once
|
| 2 |
+
|
| 3 |
+
// @generated by torchgen/gen.py from Function.h
|
| 4 |
+
|
| 5 |
+
#include <ATen/Context.h>
|
| 6 |
+
#include <ATen/DeviceGuard.h>
|
| 7 |
+
#include <ATen/TensorUtils.h>
|
| 8 |
+
#include <ATen/TracerMode.h>
|
| 9 |
+
#include <ATen/core/Generator.h>
|
| 10 |
+
#include <ATen/core/Reduction.h>
|
| 11 |
+
#include <ATen/core/Tensor.h>
|
| 12 |
+
#include <c10/core/Scalar.h>
|
| 13 |
+
#include <c10/core/Storage.h>
|
| 14 |
+
#include <c10/core/TensorOptions.h>
|
| 15 |
+
#include <c10/util/Deprecated.h>
|
| 16 |
+
#include <c10/util/Optional.h>
|
| 17 |
+
|
| 18 |
+
|
| 19 |
+
|
| 20 |
+
#include <ATen/ops/embedding_dense_backward_ops.h>
|
| 21 |
+
|
| 22 |
+
namespace at {
|
| 23 |
+
|
| 24 |
+
|
| 25 |
+
// aten::embedding_dense_backward(Tensor grad_output, Tensor indices, SymInt num_weights, SymInt padding_idx, bool scale_grad_by_freq) -> Tensor
|
| 26 |
+
inline at::Tensor embedding_dense_backward(const at::Tensor & grad_output, const at::Tensor & indices, int64_t num_weights, int64_t padding_idx, bool scale_grad_by_freq) {
|
| 27 |
+
return at::_ops::embedding_dense_backward::call(grad_output, indices, num_weights, padding_idx, scale_grad_by_freq);
|
| 28 |
+
}
|
| 29 |
+
namespace symint {
|
| 30 |
+
template <typename T, typename = std::enable_if_t<std::is_same<T, int64_t>::value>>
|
| 31 |
+
at::Tensor embedding_dense_backward(const at::Tensor & grad_output, const at::Tensor & indices, int64_t num_weights, int64_t padding_idx, bool scale_grad_by_freq) {
|
| 32 |
+
return at::_ops::embedding_dense_backward::call(grad_output, indices, num_weights, padding_idx, scale_grad_by_freq);
|
| 33 |
+
}
|
| 34 |
+
}
|
| 35 |
+
|
| 36 |
+
// aten::embedding_dense_backward(Tensor grad_output, Tensor indices, SymInt num_weights, SymInt padding_idx, bool scale_grad_by_freq) -> Tensor
|
| 37 |
+
inline at::Tensor embedding_dense_backward_symint(const at::Tensor & grad_output, const at::Tensor & indices, c10::SymInt num_weights, c10::SymInt padding_idx, bool scale_grad_by_freq) {
|
| 38 |
+
return at::_ops::embedding_dense_backward::call(grad_output, indices, num_weights, padding_idx, scale_grad_by_freq);
|
| 39 |
+
}
|
| 40 |
+
namespace symint {
|
| 41 |
+
template <typename T, typename = std::enable_if_t<std::is_same<T, c10::SymInt>::value>>
|
| 42 |
+
at::Tensor embedding_dense_backward(const at::Tensor & grad_output, const at::Tensor & indices, c10::SymInt num_weights, c10::SymInt padding_idx, bool scale_grad_by_freq) {
|
| 43 |
+
return at::_ops::embedding_dense_backward::call(grad_output, indices, num_weights, padding_idx, scale_grad_by_freq);
|
| 44 |
+
}
|
| 45 |
+
}
|
| 46 |
+
|
| 47 |
+
// aten::embedding_dense_backward.out(Tensor grad_output, Tensor indices, SymInt num_weights, SymInt padding_idx, bool scale_grad_by_freq, *, Tensor(a!) out) -> Tensor(a!)
|
| 48 |
+
inline at::Tensor & embedding_dense_backward_out(at::Tensor & out, const at::Tensor & grad_output, const at::Tensor & indices, int64_t num_weights, int64_t padding_idx, bool scale_grad_by_freq) {
|
| 49 |
+
return at::_ops::embedding_dense_backward_out::call(grad_output, indices, num_weights, padding_idx, scale_grad_by_freq, out);
|
| 50 |
+
}
|
| 51 |
+
namespace symint {
|
| 52 |
+
template <typename T, typename = std::enable_if_t<std::is_same<T, int64_t>::value>>
|
| 53 |
+
at::Tensor & embedding_dense_backward_out(at::Tensor & out, const at::Tensor & grad_output, const at::Tensor & indices, int64_t num_weights, int64_t padding_idx, bool scale_grad_by_freq) {
|
| 54 |
+
return at::_ops::embedding_dense_backward_out::call(grad_output, indices, num_weights, padding_idx, scale_grad_by_freq, out);
|
| 55 |
+
}
|
| 56 |
+
}
|
| 57 |
+
|
| 58 |
+
// aten::embedding_dense_backward.out(Tensor grad_output, Tensor indices, SymInt num_weights, SymInt padding_idx, bool scale_grad_by_freq, *, Tensor(a!) out) -> Tensor(a!)
|
| 59 |
+
inline at::Tensor & embedding_dense_backward_outf(const at::Tensor & grad_output, const at::Tensor & indices, int64_t num_weights, int64_t padding_idx, bool scale_grad_by_freq, at::Tensor & out) {
|
| 60 |
+
return at::_ops::embedding_dense_backward_out::call(grad_output, indices, num_weights, padding_idx, scale_grad_by_freq, out);
|
| 61 |
+
}
|
| 62 |
+
namespace symint {
|
| 63 |
+
template <typename T, typename = std::enable_if_t<std::is_same<T, int64_t>::value>>
|
| 64 |
+
at::Tensor & embedding_dense_backward_outf(const at::Tensor & grad_output, const at::Tensor & indices, int64_t num_weights, int64_t padding_idx, bool scale_grad_by_freq, at::Tensor & out) {
|
| 65 |
+
return at::_ops::embedding_dense_backward_out::call(grad_output, indices, num_weights, padding_idx, scale_grad_by_freq, out);
|
| 66 |
+
}
|
| 67 |
+
}
|
| 68 |
+
|
| 69 |
+
// aten::embedding_dense_backward.out(Tensor grad_output, Tensor indices, SymInt num_weights, SymInt padding_idx, bool scale_grad_by_freq, *, Tensor(a!) out) -> Tensor(a!)
|
| 70 |
+
inline at::Tensor & embedding_dense_backward_symint_out(at::Tensor & out, const at::Tensor & grad_output, const at::Tensor & indices, c10::SymInt num_weights, c10::SymInt padding_idx, bool scale_grad_by_freq) {
|
| 71 |
+
return at::_ops::embedding_dense_backward_out::call(grad_output, indices, num_weights, padding_idx, scale_grad_by_freq, out);
|
| 72 |
+
}
|
| 73 |
+
namespace symint {
|
| 74 |
+
template <typename T, typename = std::enable_if_t<std::is_same<T, c10::SymInt>::value>>
|
| 75 |
+
at::Tensor & embedding_dense_backward_out(at::Tensor & out, const at::Tensor & grad_output, const at::Tensor & indices, c10::SymInt num_weights, c10::SymInt padding_idx, bool scale_grad_by_freq) {
|
| 76 |
+
return at::_ops::embedding_dense_backward_out::call(grad_output, indices, num_weights, padding_idx, scale_grad_by_freq, out);
|
| 77 |
+
}
|
| 78 |
+
}
|
| 79 |
+
|
| 80 |
+
// aten::embedding_dense_backward.out(Tensor grad_output, Tensor indices, SymInt num_weights, SymInt padding_idx, bool scale_grad_by_freq, *, Tensor(a!) out) -> Tensor(a!)
|
| 81 |
+
inline at::Tensor & embedding_dense_backward_symint_outf(const at::Tensor & grad_output, const at::Tensor & indices, c10::SymInt num_weights, c10::SymInt padding_idx, bool scale_grad_by_freq, at::Tensor & out) {
|
| 82 |
+
return at::_ops::embedding_dense_backward_out::call(grad_output, indices, num_weights, padding_idx, scale_grad_by_freq, out);
|
| 83 |
+
}
|
| 84 |
+
namespace symint {
|
| 85 |
+
template <typename T, typename = std::enable_if_t<std::is_same<T, c10::SymInt>::value>>
|
| 86 |
+
at::Tensor & embedding_dense_backward_outf(const at::Tensor & grad_output, const at::Tensor & indices, c10::SymInt num_weights, c10::SymInt padding_idx, bool scale_grad_by_freq, at::Tensor & out) {
|
| 87 |
+
return at::_ops::embedding_dense_backward_out::call(grad_output, indices, num_weights, padding_idx, scale_grad_by_freq, out);
|
| 88 |
+
}
|
| 89 |
+
}
|
| 90 |
+
|
| 91 |
+
}
|