ZTWHHH commited on
Commit
d886d14
·
verified ·
1 Parent(s): 10c06f3

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +1 -0
  2. mantis_evalkit/lib/python3.10/site-packages/pip/_internal/exceptions.py +809 -0
  3. moondream/lib/python3.10/site-packages/torch/__pycache__/_meta_registrations.cpython-310.pyc +3 -0
  4. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/_efficient_attention_forward_cuda_dispatch.h +23 -0
  5. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/_embedding_bag_backward_compositeimplicitautograd_dispatch.h +24 -0
  6. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/_foreach_log1p_native.h +25 -0
  7. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/_foreach_log_ops.h +50 -0
  8. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/_foreach_round_native.h +25 -0
  9. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/_fused_adamw_compositeexplicitautograd_dispatch.h +28 -0
  10. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/_nested_view_from_buffer_copy_native.h +22 -0
  11. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/_shape_as_tensor_native.h +21 -0
  12. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/_sparse_broadcast_to_native.h +21 -0
  13. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/_to_sparse_csc_native.h +24 -0
  14. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/adaptive_avg_pool2d_ops.h +39 -0
  15. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/adaptive_max_pool2d_native.h +26 -0
  16. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/angle_cuda_dispatch.h +25 -0
  17. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/argsort_native.h +24 -0
  18. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/cdist_native.h +21 -0
  19. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/col2im_ops.h +39 -0
  20. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/copy_native.h +27 -0
  21. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/cudnn_grid_sampler_cuda_dispatch.h +23 -0
  22. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/digamma_native.h +23 -0
  23. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/dstack_ops.h +39 -0
  24. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/embedding_backward_ops.h +28 -0
  25. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/exp_compositeexplicitautogradnonfunctional_dispatch.h +24 -0
  26. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/feature_alpha_dropout_compositeimplicitautograd_dispatch.h +24 -0
  27. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/flatten_dense_tensors.h +30 -0
  28. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/flip_cpu_dispatch.h +23 -0
  29. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/glu_ops.h +39 -0
  30. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/grid_sampler_2d_compositeexplicitautograd_dispatch.h +24 -0
  31. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/hardswish_cuda_dispatch.h +26 -0
  32. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/hardtanh_backward_native.h +22 -0
  33. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/histogramdd_compositeimplicitautograd_dispatch.h +25 -0
  34. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/huber_loss_cuda_dispatch.h +25 -0
  35. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/igammac_meta_dispatch.h +26 -0
  36. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/instance_norm_native.h +21 -0
  37. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/int_repr_native.h +23 -0
  38. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/is_distributed_ops.h +28 -0
  39. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/isnan_cpu_dispatch.h +23 -0
  40. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/isneginf_native.h +27 -0
  41. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/isposinf_meta_dispatch.h +25 -0
  42. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/le_compositeexplicitautogradnonfunctional_dispatch.h +26 -0
  43. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/linalg_cholesky_ex_native.h +23 -0
  44. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/linalg_eig_cuda_dispatch.h +25 -0
  45. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/linalg_eigh_native.h +22 -0
  46. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/linalg_lstsq.h +39 -0
  47. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/logaddexp_native.h +23 -0
  48. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/logspace_native.h +29 -0
  49. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/margin_ranking_loss_ops.h +28 -0
  50. moondream/lib/python3.10/site-packages/torch/include/ATen/ops/mode_ops.h +61 -0
.gitattributes CHANGED
@@ -518,3 +518,4 @@ mantis_evalkit/lib/python3.10/site-packages/scipy.libs/libquadmath-96973f99-934c
518
  mantis_evalkit/lib/python3.10/site-packages/scipy.libs/libgfortran-040039e1-0352e75f.so.5.0.0 filter=lfs diff=lfs merge=lfs -text
519
  mantis_evalkit/lib/python3.10/site-packages/scipy.libs/libquadmath-96973f99.so.0.0.0 filter=lfs diff=lfs merge=lfs -text
520
  mantis_evalkit/lib/python3.10/site-packages/kiwisolver/_cext.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
 
 
518
  mantis_evalkit/lib/python3.10/site-packages/scipy.libs/libgfortran-040039e1-0352e75f.so.5.0.0 filter=lfs diff=lfs merge=lfs -text
519
  mantis_evalkit/lib/python3.10/site-packages/scipy.libs/libquadmath-96973f99.so.0.0.0 filter=lfs diff=lfs merge=lfs -text
520
  mantis_evalkit/lib/python3.10/site-packages/kiwisolver/_cext.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
521
+ moondream/lib/python3.10/site-packages/torch/__pycache__/_meta_registrations.cpython-310.pyc filter=lfs diff=lfs merge=lfs -text
mantis_evalkit/lib/python3.10/site-packages/pip/_internal/exceptions.py ADDED
@@ -0,0 +1,809 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """Exceptions used throughout package.
2
+
3
+ This module MUST NOT try to import from anything within `pip._internal` to
4
+ operate. This is expected to be importable from any/all files within the
5
+ subpackage and, thus, should not depend on them.
6
+ """
7
+
8
+ import configparser
9
+ import contextlib
10
+ import locale
11
+ import logging
12
+ import pathlib
13
+ import re
14
+ import sys
15
+ from itertools import chain, groupby, repeat
16
+ from typing import TYPE_CHECKING, Dict, Iterator, List, Literal, Optional, Union
17
+
18
+ from pip._vendor.packaging.requirements import InvalidRequirement
19
+ from pip._vendor.packaging.version import InvalidVersion
20
+ from pip._vendor.rich.console import Console, ConsoleOptions, RenderResult
21
+ from pip._vendor.rich.markup import escape
22
+ from pip._vendor.rich.text import Text
23
+
24
+ if TYPE_CHECKING:
25
+ from hashlib import _Hash
26
+
27
+ from pip._vendor.requests.models import Request, Response
28
+
29
+ from pip._internal.metadata import BaseDistribution
30
+ from pip._internal.req.req_install import InstallRequirement
31
+
32
+ logger = logging.getLogger(__name__)
33
+
34
+
35
+ #
36
+ # Scaffolding
37
+ #
38
+ def _is_kebab_case(s: str) -> bool:
39
+ return re.match(r"^[a-z]+(-[a-z]+)*$", s) is not None
40
+
41
+
42
+ def _prefix_with_indent(
43
+ s: Union[Text, str],
44
+ console: Console,
45
+ *,
46
+ prefix: str,
47
+ indent: str,
48
+ ) -> Text:
49
+ if isinstance(s, Text):
50
+ text = s
51
+ else:
52
+ text = console.render_str(s)
53
+
54
+ return console.render_str(prefix, overflow="ignore") + console.render_str(
55
+ f"\n{indent}", overflow="ignore"
56
+ ).join(text.split(allow_blank=True))
57
+
58
+
59
+ class PipError(Exception):
60
+ """The base pip error."""
61
+
62
+
63
+ class DiagnosticPipError(PipError):
64
+ """An error, that presents diagnostic information to the user.
65
+
66
+ This contains a bunch of logic, to enable pretty presentation of our error
67
+ messages. Each error gets a unique reference. Each error can also include
68
+ additional context, a hint and/or a note -- which are presented with the
69
+ main error message in a consistent style.
70
+
71
+ This is adapted from the error output styling in `sphinx-theme-builder`.
72
+ """
73
+
74
+ reference: str
75
+
76
+ def __init__(
77
+ self,
78
+ *,
79
+ kind: 'Literal["error", "warning"]' = "error",
80
+ reference: Optional[str] = None,
81
+ message: Union[str, Text],
82
+ context: Optional[Union[str, Text]],
83
+ hint_stmt: Optional[Union[str, Text]],
84
+ note_stmt: Optional[Union[str, Text]] = None,
85
+ link: Optional[str] = None,
86
+ ) -> None:
87
+ # Ensure a proper reference is provided.
88
+ if reference is None:
89
+ assert hasattr(self, "reference"), "error reference not provided!"
90
+ reference = self.reference
91
+ assert _is_kebab_case(reference), "error reference must be kebab-case!"
92
+
93
+ self.kind = kind
94
+ self.reference = reference
95
+
96
+ self.message = message
97
+ self.context = context
98
+
99
+ self.note_stmt = note_stmt
100
+ self.hint_stmt = hint_stmt
101
+
102
+ self.link = link
103
+
104
+ super().__init__(f"<{self.__class__.__name__}: {self.reference}>")
105
+
106
+ def __repr__(self) -> str:
107
+ return (
108
+ f"<{self.__class__.__name__}("
109
+ f"reference={self.reference!r}, "
110
+ f"message={self.message!r}, "
111
+ f"context={self.context!r}, "
112
+ f"note_stmt={self.note_stmt!r}, "
113
+ f"hint_stmt={self.hint_stmt!r}"
114
+ ")>"
115
+ )
116
+
117
+ def __rich_console__(
118
+ self,
119
+ console: Console,
120
+ options: ConsoleOptions,
121
+ ) -> RenderResult:
122
+ colour = "red" if self.kind == "error" else "yellow"
123
+
124
+ yield f"[{colour} bold]{self.kind}[/]: [bold]{self.reference}[/]"
125
+ yield ""
126
+
127
+ if not options.ascii_only:
128
+ # Present the main message, with relevant context indented.
129
+ if self.context is not None:
130
+ yield _prefix_with_indent(
131
+ self.message,
132
+ console,
133
+ prefix=f"[{colour}]×[/] ",
134
+ indent=f"[{colour}]│[/] ",
135
+ )
136
+ yield _prefix_with_indent(
137
+ self.context,
138
+ console,
139
+ prefix=f"[{colour}]╰─>[/] ",
140
+ indent=f"[{colour}] [/] ",
141
+ )
142
+ else:
143
+ yield _prefix_with_indent(
144
+ self.message,
145
+ console,
146
+ prefix="[red]×[/] ",
147
+ indent=" ",
148
+ )
149
+ else:
150
+ yield self.message
151
+ if self.context is not None:
152
+ yield ""
153
+ yield self.context
154
+
155
+ if self.note_stmt is not None or self.hint_stmt is not None:
156
+ yield ""
157
+
158
+ if self.note_stmt is not None:
159
+ yield _prefix_with_indent(
160
+ self.note_stmt,
161
+ console,
162
+ prefix="[magenta bold]note[/]: ",
163
+ indent=" ",
164
+ )
165
+ if self.hint_stmt is not None:
166
+ yield _prefix_with_indent(
167
+ self.hint_stmt,
168
+ console,
169
+ prefix="[cyan bold]hint[/]: ",
170
+ indent=" ",
171
+ )
172
+
173
+ if self.link is not None:
174
+ yield ""
175
+ yield f"Link: {self.link}"
176
+
177
+
178
+ #
179
+ # Actual Errors
180
+ #
181
+ class ConfigurationError(PipError):
182
+ """General exception in configuration"""
183
+
184
+
185
+ class InstallationError(PipError):
186
+ """General exception during installation"""
187
+
188
+
189
+ class MissingPyProjectBuildRequires(DiagnosticPipError):
190
+ """Raised when pyproject.toml has `build-system`, but no `build-system.requires`."""
191
+
192
+ reference = "missing-pyproject-build-system-requires"
193
+
194
+ def __init__(self, *, package: str) -> None:
195
+ super().__init__(
196
+ message=f"Can not process {escape(package)}",
197
+ context=Text(
198
+ "This package has an invalid pyproject.toml file.\n"
199
+ "The [build-system] table is missing the mandatory `requires` key."
200
+ ),
201
+ note_stmt="This is an issue with the package mentioned above, not pip.",
202
+ hint_stmt=Text("See PEP 518 for the detailed specification."),
203
+ )
204
+
205
+
206
+ class InvalidPyProjectBuildRequires(DiagnosticPipError):
207
+ """Raised when pyproject.toml an invalid `build-system.requires`."""
208
+
209
+ reference = "invalid-pyproject-build-system-requires"
210
+
211
+ def __init__(self, *, package: str, reason: str) -> None:
212
+ super().__init__(
213
+ message=f"Can not process {escape(package)}",
214
+ context=Text(
215
+ "This package has an invalid `build-system.requires` key in "
216
+ f"pyproject.toml.\n{reason}"
217
+ ),
218
+ note_stmt="This is an issue with the package mentioned above, not pip.",
219
+ hint_stmt=Text("See PEP 518 for the detailed specification."),
220
+ )
221
+
222
+
223
+ class NoneMetadataError(PipError):
224
+ """Raised when accessing a Distribution's "METADATA" or "PKG-INFO".
225
+
226
+ This signifies an inconsistency, when the Distribution claims to have
227
+ the metadata file (if not, raise ``FileNotFoundError`` instead), but is
228
+ not actually able to produce its content. This may be due to permission
229
+ errors.
230
+ """
231
+
232
+ def __init__(
233
+ self,
234
+ dist: "BaseDistribution",
235
+ metadata_name: str,
236
+ ) -> None:
237
+ """
238
+ :param dist: A Distribution object.
239
+ :param metadata_name: The name of the metadata being accessed
240
+ (can be "METADATA" or "PKG-INFO").
241
+ """
242
+ self.dist = dist
243
+ self.metadata_name = metadata_name
244
+
245
+ def __str__(self) -> str:
246
+ # Use `dist` in the error message because its stringification
247
+ # includes more information, like the version and location.
248
+ return f"None {self.metadata_name} metadata found for distribution: {self.dist}"
249
+
250
+
251
+ class UserInstallationInvalid(InstallationError):
252
+ """A --user install is requested on an environment without user site."""
253
+
254
+ def __str__(self) -> str:
255
+ return "User base directory is not specified"
256
+
257
+
258
+ class InvalidSchemeCombination(InstallationError):
259
+ def __str__(self) -> str:
260
+ before = ", ".join(str(a) for a in self.args[:-1])
261
+ return f"Cannot set {before} and {self.args[-1]} together"
262
+
263
+
264
+ class DistributionNotFound(InstallationError):
265
+ """Raised when a distribution cannot be found to satisfy a requirement"""
266
+
267
+
268
+ class RequirementsFileParseError(InstallationError):
269
+ """Raised when a general error occurs parsing a requirements file line."""
270
+
271
+
272
+ class BestVersionAlreadyInstalled(PipError):
273
+ """Raised when the most up-to-date version of a package is already
274
+ installed."""
275
+
276
+
277
+ class BadCommand(PipError):
278
+ """Raised when virtualenv or a command is not found"""
279
+
280
+
281
+ class CommandError(PipError):
282
+ """Raised when there is an error in command-line arguments"""
283
+
284
+
285
+ class PreviousBuildDirError(PipError):
286
+ """Raised when there's a previous conflicting build directory"""
287
+
288
+
289
+ class NetworkConnectionError(PipError):
290
+ """HTTP connection error"""
291
+
292
+ def __init__(
293
+ self,
294
+ error_msg: str,
295
+ response: Optional["Response"] = None,
296
+ request: Optional["Request"] = None,
297
+ ) -> None:
298
+ """
299
+ Initialize NetworkConnectionError with `request` and `response`
300
+ objects.
301
+ """
302
+ self.response = response
303
+ self.request = request
304
+ self.error_msg = error_msg
305
+ if (
306
+ self.response is not None
307
+ and not self.request
308
+ and hasattr(response, "request")
309
+ ):
310
+ self.request = self.response.request
311
+ super().__init__(error_msg, response, request)
312
+
313
+ def __str__(self) -> str:
314
+ return str(self.error_msg)
315
+
316
+
317
+ class InvalidWheelFilename(InstallationError):
318
+ """Invalid wheel filename."""
319
+
320
+
321
+ class UnsupportedWheel(InstallationError):
322
+ """Unsupported wheel."""
323
+
324
+
325
+ class InvalidWheel(InstallationError):
326
+ """Invalid (e.g. corrupt) wheel."""
327
+
328
+ def __init__(self, location: str, name: str):
329
+ self.location = location
330
+ self.name = name
331
+
332
+ def __str__(self) -> str:
333
+ return f"Wheel '{self.name}' located at {self.location} is invalid."
334
+
335
+
336
+ class MetadataInconsistent(InstallationError):
337
+ """Built metadata contains inconsistent information.
338
+
339
+ This is raised when the metadata contains values (e.g. name and version)
340
+ that do not match the information previously obtained from sdist filename,
341
+ user-supplied ``#egg=`` value, or an install requirement name.
342
+ """
343
+
344
+ def __init__(
345
+ self, ireq: "InstallRequirement", field: str, f_val: str, m_val: str
346
+ ) -> None:
347
+ self.ireq = ireq
348
+ self.field = field
349
+ self.f_val = f_val
350
+ self.m_val = m_val
351
+
352
+ def __str__(self) -> str:
353
+ return (
354
+ f"Requested {self.ireq} has inconsistent {self.field}: "
355
+ f"expected {self.f_val!r}, but metadata has {self.m_val!r}"
356
+ )
357
+
358
+
359
+ class MetadataInvalid(InstallationError):
360
+ """Metadata is invalid."""
361
+
362
+ def __init__(self, ireq: "InstallRequirement", error: str) -> None:
363
+ self.ireq = ireq
364
+ self.error = error
365
+
366
+ def __str__(self) -> str:
367
+ return f"Requested {self.ireq} has invalid metadata: {self.error}"
368
+
369
+
370
+ class InstallationSubprocessError(DiagnosticPipError, InstallationError):
371
+ """A subprocess call failed."""
372
+
373
+ reference = "subprocess-exited-with-error"
374
+
375
+ def __init__(
376
+ self,
377
+ *,
378
+ command_description: str,
379
+ exit_code: int,
380
+ output_lines: Optional[List[str]],
381
+ ) -> None:
382
+ if output_lines is None:
383
+ output_prompt = Text("See above for output.")
384
+ else:
385
+ output_prompt = (
386
+ Text.from_markup(f"[red][{len(output_lines)} lines of output][/]\n")
387
+ + Text("".join(output_lines))
388
+ + Text.from_markup(R"[red]\[end of output][/]")
389
+ )
390
+
391
+ super().__init__(
392
+ message=(
393
+ f"[green]{escape(command_description)}[/] did not run successfully.\n"
394
+ f"exit code: {exit_code}"
395
+ ),
396
+ context=output_prompt,
397
+ hint_stmt=None,
398
+ note_stmt=(
399
+ "This error originates from a subprocess, and is likely not a "
400
+ "problem with pip."
401
+ ),
402
+ )
403
+
404
+ self.command_description = command_description
405
+ self.exit_code = exit_code
406
+
407
+ def __str__(self) -> str:
408
+ return f"{self.command_description} exited with {self.exit_code}"
409
+
410
+
411
+ class MetadataGenerationFailed(InstallationSubprocessError, InstallationError):
412
+ reference = "metadata-generation-failed"
413
+
414
+ def __init__(
415
+ self,
416
+ *,
417
+ package_details: str,
418
+ ) -> None:
419
+ super(InstallationSubprocessError, self).__init__(
420
+ message="Encountered error while generating package metadata.",
421
+ context=escape(package_details),
422
+ hint_stmt="See above for details.",
423
+ note_stmt="This is an issue with the package mentioned above, not pip.",
424
+ )
425
+
426
+ def __str__(self) -> str:
427
+ return "metadata generation failed"
428
+
429
+
430
+ class HashErrors(InstallationError):
431
+ """Multiple HashError instances rolled into one for reporting"""
432
+
433
+ def __init__(self) -> None:
434
+ self.errors: List[HashError] = []
435
+
436
+ def append(self, error: "HashError") -> None:
437
+ self.errors.append(error)
438
+
439
+ def __str__(self) -> str:
440
+ lines = []
441
+ self.errors.sort(key=lambda e: e.order)
442
+ for cls, errors_of_cls in groupby(self.errors, lambda e: e.__class__):
443
+ lines.append(cls.head)
444
+ lines.extend(e.body() for e in errors_of_cls)
445
+ if lines:
446
+ return "\n".join(lines)
447
+ return ""
448
+
449
+ def __bool__(self) -> bool:
450
+ return bool(self.errors)
451
+
452
+
453
+ class HashError(InstallationError):
454
+ """
455
+ A failure to verify a package against known-good hashes
456
+
457
+ :cvar order: An int sorting hash exception classes by difficulty of
458
+ recovery (lower being harder), so the user doesn't bother fretting
459
+ about unpinned packages when he has deeper issues, like VCS
460
+ dependencies, to deal with. Also keeps error reports in a
461
+ deterministic order.
462
+ :cvar head: A section heading for display above potentially many
463
+ exceptions of this kind
464
+ :ivar req: The InstallRequirement that triggered this error. This is
465
+ pasted on after the exception is instantiated, because it's not
466
+ typically available earlier.
467
+
468
+ """
469
+
470
+ req: Optional["InstallRequirement"] = None
471
+ head = ""
472
+ order: int = -1
473
+
474
+ def body(self) -> str:
475
+ """Return a summary of me for display under the heading.
476
+
477
+ This default implementation simply prints a description of the
478
+ triggering requirement.
479
+
480
+ :param req: The InstallRequirement that provoked this error, with
481
+ its link already populated by the resolver's _populate_link().
482
+
483
+ """
484
+ return f" {self._requirement_name()}"
485
+
486
+ def __str__(self) -> str:
487
+ return f"{self.head}\n{self.body()}"
488
+
489
+ def _requirement_name(self) -> str:
490
+ """Return a description of the requirement that triggered me.
491
+
492
+ This default implementation returns long description of the req, with
493
+ line numbers
494
+
495
+ """
496
+ return str(self.req) if self.req else "unknown package"
497
+
498
+
499
+ class VcsHashUnsupported(HashError):
500
+ """A hash was provided for a version-control-system-based requirement, but
501
+ we don't have a method for hashing those."""
502
+
503
+ order = 0
504
+ head = (
505
+ "Can't verify hashes for these requirements because we don't "
506
+ "have a way to hash version control repositories:"
507
+ )
508
+
509
+
510
+ class DirectoryUrlHashUnsupported(HashError):
511
+ """A hash was provided for a version-control-system-based requirement, but
512
+ we don't have a method for hashing those."""
513
+
514
+ order = 1
515
+ head = (
516
+ "Can't verify hashes for these file:// requirements because they "
517
+ "point to directories:"
518
+ )
519
+
520
+
521
+ class HashMissing(HashError):
522
+ """A hash was needed for a requirement but is absent."""
523
+
524
+ order = 2
525
+ head = (
526
+ "Hashes are required in --require-hashes mode, but they are "
527
+ "missing from some requirements. Here is a list of those "
528
+ "requirements along with the hashes their downloaded archives "
529
+ "actually had. Add lines like these to your requirements files to "
530
+ "prevent tampering. (If you did not enable --require-hashes "
531
+ "manually, note that it turns on automatically when any package "
532
+ "has a hash.)"
533
+ )
534
+
535
+ def __init__(self, gotten_hash: str) -> None:
536
+ """
537
+ :param gotten_hash: The hash of the (possibly malicious) archive we
538
+ just downloaded
539
+ """
540
+ self.gotten_hash = gotten_hash
541
+
542
+ def body(self) -> str:
543
+ # Dodge circular import.
544
+ from pip._internal.utils.hashes import FAVORITE_HASH
545
+
546
+ package = None
547
+ if self.req:
548
+ # In the case of URL-based requirements, display the original URL
549
+ # seen in the requirements file rather than the package name,
550
+ # so the output can be directly copied into the requirements file.
551
+ package = (
552
+ self.req.original_link
553
+ if self.req.is_direct
554
+ # In case someone feeds something downright stupid
555
+ # to InstallRequirement's constructor.
556
+ else getattr(self.req, "req", None)
557
+ )
558
+ return " {} --hash={}:{}".format(
559
+ package or "unknown package", FAVORITE_HASH, self.gotten_hash
560
+ )
561
+
562
+
563
+ class HashUnpinned(HashError):
564
+ """A requirement had a hash specified but was not pinned to a specific
565
+ version."""
566
+
567
+ order = 3
568
+ head = (
569
+ "In --require-hashes mode, all requirements must have their "
570
+ "versions pinned with ==. These do not:"
571
+ )
572
+
573
+
574
+ class HashMismatch(HashError):
575
+ """
576
+ Distribution file hash values don't match.
577
+
578
+ :ivar package_name: The name of the package that triggered the hash
579
+ mismatch. Feel free to write to this after the exception is raise to
580
+ improve its error message.
581
+
582
+ """
583
+
584
+ order = 4
585
+ head = (
586
+ "THESE PACKAGES DO NOT MATCH THE HASHES FROM THE REQUIREMENTS "
587
+ "FILE. If you have updated the package versions, please update "
588
+ "the hashes. Otherwise, examine the package contents carefully; "
589
+ "someone may have tampered with them."
590
+ )
591
+
592
+ def __init__(self, allowed: Dict[str, List[str]], gots: Dict[str, "_Hash"]) -> None:
593
+ """
594
+ :param allowed: A dict of algorithm names pointing to lists of allowed
595
+ hex digests
596
+ :param gots: A dict of algorithm names pointing to hashes we
597
+ actually got from the files under suspicion
598
+ """
599
+ self.allowed = allowed
600
+ self.gots = gots
601
+
602
+ def body(self) -> str:
603
+ return f" {self._requirement_name()}:\n{self._hash_comparison()}"
604
+
605
+ def _hash_comparison(self) -> str:
606
+ """
607
+ Return a comparison of actual and expected hash values.
608
+
609
+ Example::
610
+
611
+ Expected sha256 abcdeabcdeabcdeabcdeabcdeabcdeabcdeabcdeabcde
612
+ or 123451234512345123451234512345123451234512345
613
+ Got bcdefbcdefbcdefbcdefbcdefbcdefbcdefbcdefbcdef
614
+
615
+ """
616
+
617
+ def hash_then_or(hash_name: str) -> "chain[str]":
618
+ # For now, all the decent hashes have 6-char names, so we can get
619
+ # away with hard-coding space literals.
620
+ return chain([hash_name], repeat(" or"))
621
+
622
+ lines: List[str] = []
623
+ for hash_name, expecteds in self.allowed.items():
624
+ prefix = hash_then_or(hash_name)
625
+ lines.extend((f" Expected {next(prefix)} {e}") for e in expecteds)
626
+ lines.append(
627
+ f" Got {self.gots[hash_name].hexdigest()}\n"
628
+ )
629
+ return "\n".join(lines)
630
+
631
+
632
+ class UnsupportedPythonVersion(InstallationError):
633
+ """Unsupported python version according to Requires-Python package
634
+ metadata."""
635
+
636
+
637
+ class ConfigurationFileCouldNotBeLoaded(ConfigurationError):
638
+ """When there are errors while loading a configuration file"""
639
+
640
+ def __init__(
641
+ self,
642
+ reason: str = "could not be loaded",
643
+ fname: Optional[str] = None,
644
+ error: Optional[configparser.Error] = None,
645
+ ) -> None:
646
+ super().__init__(error)
647
+ self.reason = reason
648
+ self.fname = fname
649
+ self.error = error
650
+
651
+ def __str__(self) -> str:
652
+ if self.fname is not None:
653
+ message_part = f" in {self.fname}."
654
+ else:
655
+ assert self.error is not None
656
+ message_part = f".\n{self.error}\n"
657
+ return f"Configuration file {self.reason}{message_part}"
658
+
659
+
660
+ _DEFAULT_EXTERNALLY_MANAGED_ERROR = f"""\
661
+ The Python environment under {sys.prefix} is managed externally, and may not be
662
+ manipulated by the user. Please use specific tooling from the distributor of
663
+ the Python installation to interact with this environment instead.
664
+ """
665
+
666
+
667
+ class ExternallyManagedEnvironment(DiagnosticPipError):
668
+ """The current environment is externally managed.
669
+
670
+ This is raised when the current environment is externally managed, as
671
+ defined by `PEP 668`_. The ``EXTERNALLY-MANAGED`` configuration is checked
672
+ and displayed when the error is bubbled up to the user.
673
+
674
+ :param error: The error message read from ``EXTERNALLY-MANAGED``.
675
+ """
676
+
677
+ reference = "externally-managed-environment"
678
+
679
+ def __init__(self, error: Optional[str]) -> None:
680
+ if error is None:
681
+ context = Text(_DEFAULT_EXTERNALLY_MANAGED_ERROR)
682
+ else:
683
+ context = Text(error)
684
+ super().__init__(
685
+ message="This environment is externally managed",
686
+ context=context,
687
+ note_stmt=(
688
+ "If you believe this is a mistake, please contact your "
689
+ "Python installation or OS distribution provider. "
690
+ "You can override this, at the risk of breaking your Python "
691
+ "installation or OS, by passing --break-system-packages."
692
+ ),
693
+ hint_stmt=Text("See PEP 668 for the detailed specification."),
694
+ )
695
+
696
+ @staticmethod
697
+ def _iter_externally_managed_error_keys() -> Iterator[str]:
698
+ # LC_MESSAGES is in POSIX, but not the C standard. The most common
699
+ # platform that does not implement this category is Windows, where
700
+ # using other categories for console message localization is equally
701
+ # unreliable, so we fall back to the locale-less vendor message. This
702
+ # can always be re-evaluated when a vendor proposes a new alternative.
703
+ try:
704
+ category = locale.LC_MESSAGES
705
+ except AttributeError:
706
+ lang: Optional[str] = None
707
+ else:
708
+ lang, _ = locale.getlocale(category)
709
+ if lang is not None:
710
+ yield f"Error-{lang}"
711
+ for sep in ("-", "_"):
712
+ before, found, _ = lang.partition(sep)
713
+ if not found:
714
+ continue
715
+ yield f"Error-{before}"
716
+ yield "Error"
717
+
718
+ @classmethod
719
+ def from_config(
720
+ cls,
721
+ config: Union[pathlib.Path, str],
722
+ ) -> "ExternallyManagedEnvironment":
723
+ parser = configparser.ConfigParser(interpolation=None)
724
+ try:
725
+ parser.read(config, encoding="utf-8")
726
+ section = parser["externally-managed"]
727
+ for key in cls._iter_externally_managed_error_keys():
728
+ with contextlib.suppress(KeyError):
729
+ return cls(section[key])
730
+ except KeyError:
731
+ pass
732
+ except (OSError, UnicodeDecodeError, configparser.ParsingError):
733
+ from pip._internal.utils._log import VERBOSE
734
+
735
+ exc_info = logger.isEnabledFor(VERBOSE)
736
+ logger.warning("Failed to read %s", config, exc_info=exc_info)
737
+ return cls(None)
738
+
739
+
740
+ class UninstallMissingRecord(DiagnosticPipError):
741
+ reference = "uninstall-no-record-file"
742
+
743
+ def __init__(self, *, distribution: "BaseDistribution") -> None:
744
+ installer = distribution.installer
745
+ if not installer or installer == "pip":
746
+ dep = f"{distribution.raw_name}=={distribution.version}"
747
+ hint = Text.assemble(
748
+ "You might be able to recover from this via: ",
749
+ (f"pip install --force-reinstall --no-deps {dep}", "green"),
750
+ )
751
+ else:
752
+ hint = Text(
753
+ f"The package was installed by {installer}. "
754
+ "You should check if it can uninstall the package."
755
+ )
756
+
757
+ super().__init__(
758
+ message=Text(f"Cannot uninstall {distribution}"),
759
+ context=(
760
+ "The package's contents are unknown: "
761
+ f"no RECORD file was found for {distribution.raw_name}."
762
+ ),
763
+ hint_stmt=hint,
764
+ )
765
+
766
+
767
+ class LegacyDistutilsInstall(DiagnosticPipError):
768
+ reference = "uninstall-distutils-installed-package"
769
+
770
+ def __init__(self, *, distribution: "BaseDistribution") -> None:
771
+ super().__init__(
772
+ message=Text(f"Cannot uninstall {distribution}"),
773
+ context=(
774
+ "It is a distutils installed project and thus we cannot accurately "
775
+ "determine which files belong to it which would lead to only a partial "
776
+ "uninstall."
777
+ ),
778
+ hint_stmt=None,
779
+ )
780
+
781
+
782
+ class InvalidInstalledPackage(DiagnosticPipError):
783
+ reference = "invalid-installed-package"
784
+
785
+ def __init__(
786
+ self,
787
+ *,
788
+ dist: "BaseDistribution",
789
+ invalid_exc: Union[InvalidRequirement, InvalidVersion],
790
+ ) -> None:
791
+ installed_location = dist.installed_location
792
+
793
+ if isinstance(invalid_exc, InvalidRequirement):
794
+ invalid_type = "requirement"
795
+ else:
796
+ invalid_type = "version"
797
+
798
+ super().__init__(
799
+ message=Text(
800
+ f"Cannot process installed package {dist} "
801
+ + (f"in {installed_location!r} " if installed_location else "")
802
+ + f"because it has an invalid {invalid_type}:\n{invalid_exc.args[0]}"
803
+ ),
804
+ context=(
805
+ "Starting with pip 24.1, packages with invalid "
806
+ f"{invalid_type}s can not be processed."
807
+ ),
808
+ hint_stmt="To proceed this package must be uninstalled.",
809
+ )
moondream/lib/python3.10/site-packages/torch/__pycache__/_meta_registrations.cpython-310.pyc ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:628d7b3c3dc8df2b6184f400d285bbe9476cf07f5fe2890c7d5da0f2f9822ecb
3
+ size 165904
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/_efficient_attention_forward_cuda_dispatch.h ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+ // @generated by torchgen/gen.py from DispatchKeyFunction.h
3
+
4
+ // NB: The implementing C++ file is RegisterDispatchKey.cpp
5
+
6
+ // The only #includes we need are for custom classes that have defaults in the C++ API
7
+ #include <c10/core/MemoryFormat.h>
8
+ #include <c10/core/Scalar.h>
9
+ #include <ATen/core/Reduction.h>
10
+
11
+ // Forward declarations of any types needed in the operator signatures.
12
+ // We can't directly include these classes because it will cause circular include dependencies.
13
+ // This file is included by TensorBody.h, which defines the Tensor class.
14
+ #include <ATen/core/ATen_fwd.h>
15
+
16
+ namespace at {
17
+
18
+ namespace cuda {
19
+
20
+ TORCH_API ::std::tuple<at::Tensor,at::Tensor,at::Tensor,at::Tensor,c10::SymInt,c10::SymInt> _efficient_attention_forward(const at::Tensor & query, const at::Tensor & key, const at::Tensor & value, const c10::optional<at::Tensor> & bias, const c10::optional<at::Tensor> & cu_seqlens_q, const c10::optional<at::Tensor> & cu_seqlens_k, c10::optional<int64_t> max_seqlen_q, c10::optional<int64_t> max_seqlen_k, double dropout_p, int64_t custom_mask_type, bool compute_log_sumexp=false, c10::optional<double> scale=c10::nullopt, const c10::optional<at::Tensor> & causal_diagonal={}, const c10::optional<at::Tensor> & seqlen_k={});
21
+
22
+ } // namespace cuda
23
+ } // namespace at
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/_embedding_bag_backward_compositeimplicitautograd_dispatch.h ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+ // @generated by torchgen/gen.py from DispatchKeyFunction.h
3
+
4
+ // NB: The implementing C++ file is RegisterDispatchKey.cpp
5
+
6
+ // The only #includes we need are for custom classes that have defaults in the C++ API
7
+ #include <c10/core/MemoryFormat.h>
8
+ #include <c10/core/Scalar.h>
9
+ #include <ATen/core/Reduction.h>
10
+
11
+ // Forward declarations of any types needed in the operator signatures.
12
+ // We can't directly include these classes because it will cause circular include dependencies.
13
+ // This file is included by TensorBody.h, which defines the Tensor class.
14
+ #include <ATen/core/ATen_fwd.h>
15
+
16
+ namespace at {
17
+
18
+ namespace compositeimplicitautograd {
19
+
20
+ TORCH_API at::Tensor _embedding_bag_backward(const at::Tensor & grad, const at::Tensor & indices, const at::Tensor & offsets, const at::Tensor & offset2bag, const at::Tensor & bag_size, const at::Tensor & maximum_indices, int64_t num_weights, bool scale_grad_by_freq, int64_t mode, bool sparse, const c10::optional<at::Tensor> & per_sample_weights, int64_t padding_idx=-1);
21
+ TORCH_API at::Tensor _embedding_bag_backward_symint(const at::Tensor & grad, const at::Tensor & indices, const at::Tensor & offsets, const at::Tensor & offset2bag, const at::Tensor & bag_size, const at::Tensor & maximum_indices, c10::SymInt num_weights, bool scale_grad_by_freq, int64_t mode, bool sparse, const c10::optional<at::Tensor> & per_sample_weights, int64_t padding_idx=-1);
22
+
23
+ } // namespace compositeimplicitautograd
24
+ } // namespace at
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/_foreach_log1p_native.h ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from NativeFunction.h
4
+
5
+ #include <c10/core/Scalar.h>
6
+ #include <c10/core/Storage.h>
7
+ #include <c10/core/TensorOptions.h>
8
+ #include <c10/util/Deprecated.h>
9
+ #include <c10/util/Optional.h>
10
+ #include <c10/core/QScheme.h>
11
+ #include <ATen/core/Reduction.h>
12
+ #include <ATen/core/Tensor.h>
13
+ #include <tuple>
14
+ #include <vector>
15
+
16
+
17
+ namespace at {
18
+ namespace native {
19
+ TORCH_API void _foreach_log1p_out(at::TensorList self, at::TensorList out);
20
+ TORCH_API ::std::vector<at::Tensor> foreach_tensor_log1p_slow(at::TensorList self);
21
+ TORCH_API void foreach_tensor_log1p_slow_(at::TensorList self);
22
+ TORCH_API ::std::vector<at::Tensor> foreach_tensor_log1p_cuda(at::TensorList self);
23
+ TORCH_API void foreach_tensor_log1p_cuda_(at::TensorList self);
24
+ } // namespace native
25
+ } // namespace at
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/_foreach_log_ops.h ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from Operator.h
4
+
5
+ #include <tuple>
6
+ #include <vector>
7
+
8
+ // Forward declarations of any types needed in the operator signatures.
9
+ // We can't directly include these classes because it will cause circular include dependencies.
10
+ // This file is included by TensorBody.h, which defines the Tensor class.
11
+ #include <ATen/core/ATen_fwd.h>
12
+
13
+ namespace at {
14
+ namespace _ops {
15
+
16
+
17
+ struct TORCH_API _foreach_log {
18
+ using schema = ::std::vector<at::Tensor> (at::TensorList);
19
+ using ptr_schema = schema*;
20
+ // See Note [static constexpr char* members for windows NVCC]
21
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(name, "aten::_foreach_log")
22
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(overload_name, "")
23
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(schema_str, "_foreach_log(Tensor[] self) -> Tensor[]")
24
+ static ::std::vector<at::Tensor> call(at::TensorList self);
25
+ static ::std::vector<at::Tensor> redispatch(c10::DispatchKeySet dispatchKeySet, at::TensorList self);
26
+ };
27
+
28
+ struct TORCH_API _foreach_log_ {
29
+ using schema = void (at::TensorList);
30
+ using ptr_schema = schema*;
31
+ // See Note [static constexpr char* members for windows NVCC]
32
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(name, "aten::_foreach_log_")
33
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(overload_name, "")
34
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(schema_str, "_foreach_log_(Tensor(a!)[] self) -> ()")
35
+ static void call(at::TensorList self);
36
+ static void redispatch(c10::DispatchKeySet dispatchKeySet, at::TensorList self);
37
+ };
38
+
39
+ struct TORCH_API _foreach_log_out {
40
+ using schema = void (at::TensorList, at::TensorList);
41
+ using ptr_schema = schema*;
42
+ // See Note [static constexpr char* members for windows NVCC]
43
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(name, "aten::_foreach_log")
44
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(overload_name, "out")
45
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(schema_str, "_foreach_log.out(Tensor[] self, *, Tensor(a!)[] out) -> ()")
46
+ static void call(at::TensorList self, at::TensorList out);
47
+ static void redispatch(c10::DispatchKeySet dispatchKeySet, at::TensorList self, at::TensorList out);
48
+ };
49
+
50
+ }} // namespace at::_ops
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/_foreach_round_native.h ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from NativeFunction.h
4
+
5
+ #include <c10/core/Scalar.h>
6
+ #include <c10/core/Storage.h>
7
+ #include <c10/core/TensorOptions.h>
8
+ #include <c10/util/Deprecated.h>
9
+ #include <c10/util/Optional.h>
10
+ #include <c10/core/QScheme.h>
11
+ #include <ATen/core/Reduction.h>
12
+ #include <ATen/core/Tensor.h>
13
+ #include <tuple>
14
+ #include <vector>
15
+
16
+
17
+ namespace at {
18
+ namespace native {
19
+ TORCH_API void _foreach_round_out(at::TensorList self, at::TensorList out);
20
+ TORCH_API ::std::vector<at::Tensor> foreach_tensor_round_slow(at::TensorList self);
21
+ TORCH_API void foreach_tensor_round_slow_(at::TensorList self);
22
+ TORCH_API ::std::vector<at::Tensor> foreach_tensor_round_cuda(at::TensorList self);
23
+ TORCH_API void foreach_tensor_round_cuda_(at::TensorList self);
24
+ } // namespace native
25
+ } // namespace at
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/_fused_adamw_compositeexplicitautograd_dispatch.h ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+ // @generated by torchgen/gen.py from DispatchKeyFunction.h
3
+
4
+ // NB: The implementing C++ file is RegisterDispatchKey.cpp
5
+
6
+ // The only #includes we need are for custom classes that have defaults in the C++ API
7
+ #include <c10/core/MemoryFormat.h>
8
+ #include <c10/core/Scalar.h>
9
+ #include <ATen/core/Reduction.h>
10
+
11
+ // Forward declarations of any types needed in the operator signatures.
12
+ // We can't directly include these classes because it will cause circular include dependencies.
13
+ // This file is included by TensorBody.h, which defines the Tensor class.
14
+ #include <ATen/core/ATen_fwd.h>
15
+
16
+ namespace at {
17
+
18
+ namespace compositeexplicitautograd {
19
+
20
+ TORCH_API ::std::tuple<::std::vector<at::Tensor>,::std::vector<at::Tensor>,::std::vector<at::Tensor>,::std::vector<at::Tensor>,::std::vector<at::Tensor>> _fused_adamw(at::TensorList self, at::TensorList grads, at::TensorList exp_avgs, at::TensorList exp_avg_sqs, at::TensorList max_exp_avg_sqs, at::TensorList state_steps, double lr, double beta1, double beta2, double weight_decay, double eps, bool amsgrad, bool maximize, const c10::optional<at::Tensor> & grad_scale={}, const c10::optional<at::Tensor> & found_inf={});
21
+ TORCH_API void _fused_adamw_out(at::TensorList out, at::TensorList self, at::TensorList grads, at::TensorList exp_avgs, at::TensorList exp_avg_sqs, at::TensorList max_exp_avg_sqs, at::TensorList state_steps, double lr, double beta1, double beta2, double weight_decay, double eps, bool amsgrad, bool maximize, const c10::optional<at::Tensor> & grad_scale={}, const c10::optional<at::Tensor> & found_inf={});
22
+ TORCH_API void _fused_adamw_outf(at::TensorList self, at::TensorList grads, at::TensorList exp_avgs, at::TensorList exp_avg_sqs, at::TensorList max_exp_avg_sqs, at::TensorList state_steps, double lr, double beta1, double beta2, double weight_decay, double eps, bool amsgrad, bool maximize, const c10::optional<at::Tensor> & grad_scale, const c10::optional<at::Tensor> & found_inf, at::TensorList out);
23
+ TORCH_API ::std::tuple<::std::vector<at::Tensor>,::std::vector<at::Tensor>,::std::vector<at::Tensor>,::std::vector<at::Tensor>,::std::vector<at::Tensor>> _fused_adamw(at::TensorList self, at::TensorList grads, at::TensorList exp_avgs, at::TensorList exp_avg_sqs, at::TensorList max_exp_avg_sqs, at::TensorList state_steps, const at::Tensor & lr, double beta1, double beta2, double weight_decay, double eps, bool amsgrad, bool maximize, const c10::optional<at::Tensor> & grad_scale={}, const c10::optional<at::Tensor> & found_inf={});
24
+ TORCH_API void _fused_adamw_out(at::TensorList out, at::TensorList self, at::TensorList grads, at::TensorList exp_avgs, at::TensorList exp_avg_sqs, at::TensorList max_exp_avg_sqs, at::TensorList state_steps, const at::Tensor & lr, double beta1, double beta2, double weight_decay, double eps, bool amsgrad, bool maximize, const c10::optional<at::Tensor> & grad_scale={}, const c10::optional<at::Tensor> & found_inf={});
25
+ TORCH_API void _fused_adamw_outf(at::TensorList self, at::TensorList grads, at::TensorList exp_avgs, at::TensorList exp_avg_sqs, at::TensorList max_exp_avg_sqs, at::TensorList state_steps, const at::Tensor & lr, double beta1, double beta2, double weight_decay, double eps, bool amsgrad, bool maximize, const c10::optional<at::Tensor> & grad_scale, const c10::optional<at::Tensor> & found_inf, at::TensorList out);
26
+
27
+ } // namespace compositeexplicitautograd
28
+ } // namespace at
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/_nested_view_from_buffer_copy_native.h ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from NativeFunction.h
4
+
5
+ #include <c10/core/Scalar.h>
6
+ #include <c10/core/Storage.h>
7
+ #include <c10/core/TensorOptions.h>
8
+ #include <c10/util/Deprecated.h>
9
+ #include <c10/util/Optional.h>
10
+ #include <c10/core/QScheme.h>
11
+ #include <ATen/core/Reduction.h>
12
+ #include <ATen/core/Tensor.h>
13
+ #include <tuple>
14
+ #include <vector>
15
+
16
+
17
+ namespace at {
18
+ namespace native {
19
+ TORCH_API at::Tensor & _nested_view_from_buffer_copy_out(const at::Tensor & self, const at::Tensor & nested_size, const at::Tensor & nested_strides, const at::Tensor & offsets, at::Tensor & out);
20
+ TORCH_API at::Tensor _nested_view_from_buffer_copy(const at::Tensor & self, const at::Tensor & nested_size, const at::Tensor & nested_strides, const at::Tensor & offsets);
21
+ } // namespace native
22
+ } // namespace at
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/_shape_as_tensor_native.h ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from NativeFunction.h
4
+
5
+ #include <c10/core/Scalar.h>
6
+ #include <c10/core/Storage.h>
7
+ #include <c10/core/TensorOptions.h>
8
+ #include <c10/util/Deprecated.h>
9
+ #include <c10/util/Optional.h>
10
+ #include <c10/core/QScheme.h>
11
+ #include <ATen/core/Reduction.h>
12
+ #include <ATen/core/Tensor.h>
13
+ #include <tuple>
14
+ #include <vector>
15
+
16
+
17
+ namespace at {
18
+ namespace native {
19
+ TORCH_API at::Tensor _shape_as_tensor(const at::Tensor & self);
20
+ } // namespace native
21
+ } // namespace at
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/_sparse_broadcast_to_native.h ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from NativeFunction.h
4
+
5
+ #include <c10/core/Scalar.h>
6
+ #include <c10/core/Storage.h>
7
+ #include <c10/core/TensorOptions.h>
8
+ #include <c10/util/Deprecated.h>
9
+ #include <c10/util/Optional.h>
10
+ #include <c10/core/QScheme.h>
11
+ #include <ATen/core/Reduction.h>
12
+ #include <ATen/core/Tensor.h>
13
+ #include <tuple>
14
+ #include <vector>
15
+
16
+
17
+ namespace at {
18
+ namespace native {
19
+ TORCH_API at::Tensor sparse_broadcast_to(const at::Tensor & self, at::IntArrayRef size);
20
+ } // namespace native
21
+ } // namespace at
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/_to_sparse_csc_native.h ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from NativeFunction.h
4
+
5
+ #include <c10/core/Scalar.h>
6
+ #include <c10/core/Storage.h>
7
+ #include <c10/core/TensorOptions.h>
8
+ #include <c10/util/Deprecated.h>
9
+ #include <c10/util/Optional.h>
10
+ #include <c10/core/QScheme.h>
11
+ #include <ATen/core/Reduction.h>
12
+ #include <ATen/core/Tensor.h>
13
+ #include <tuple>
14
+ #include <vector>
15
+
16
+
17
+ namespace at {
18
+ namespace native {
19
+ TORCH_API at::Tensor & _to_sparse_csc_out(const at::Tensor & self, c10::optional<int64_t> dense_dim, at::Tensor & out);
20
+ TORCH_API at::Tensor dense_to_sparse_csc(const at::Tensor & self, c10::optional<int64_t> dense_dim=c10::nullopt);
21
+ TORCH_API at::Tensor coo_to_sparse_csc(const at::Tensor & self, c10::optional<int64_t> dense_dim=c10::nullopt);
22
+ TORCH_API at::Tensor sparse_compressed_to_sparse_csc(const at::Tensor & self, c10::optional<int64_t> dense_dim=c10::nullopt);
23
+ } // namespace native
24
+ } // namespace at
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/adaptive_avg_pool2d_ops.h ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from Operator.h
4
+
5
+ #include <tuple>
6
+ #include <vector>
7
+
8
+ // Forward declarations of any types needed in the operator signatures.
9
+ // We can't directly include these classes because it will cause circular include dependencies.
10
+ // This file is included by TensorBody.h, which defines the Tensor class.
11
+ #include <ATen/core/ATen_fwd.h>
12
+
13
+ namespace at {
14
+ namespace _ops {
15
+
16
+
17
+ struct TORCH_API adaptive_avg_pool2d_out {
18
+ using schema = at::Tensor & (const at::Tensor &, c10::SymIntArrayRef, at::Tensor &);
19
+ using ptr_schema = schema*;
20
+ // See Note [static constexpr char* members for windows NVCC]
21
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(name, "aten::adaptive_avg_pool2d")
22
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(overload_name, "out")
23
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(schema_str, "adaptive_avg_pool2d.out(Tensor self, SymInt[2] output_size, *, Tensor(a!) out) -> Tensor(a!)")
24
+ static at::Tensor & call(const at::Tensor & self, c10::SymIntArrayRef output_size, at::Tensor & out);
25
+ static at::Tensor & redispatch(c10::DispatchKeySet dispatchKeySet, const at::Tensor & self, c10::SymIntArrayRef output_size, at::Tensor & out);
26
+ };
27
+
28
+ struct TORCH_API adaptive_avg_pool2d {
29
+ using schema = at::Tensor (const at::Tensor &, c10::SymIntArrayRef);
30
+ using ptr_schema = schema*;
31
+ // See Note [static constexpr char* members for windows NVCC]
32
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(name, "aten::adaptive_avg_pool2d")
33
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(overload_name, "")
34
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(schema_str, "adaptive_avg_pool2d(Tensor self, SymInt[2] output_size) -> Tensor")
35
+ static at::Tensor call(const at::Tensor & self, c10::SymIntArrayRef output_size);
36
+ static at::Tensor redispatch(c10::DispatchKeySet dispatchKeySet, const at::Tensor & self, c10::SymIntArrayRef output_size);
37
+ };
38
+
39
+ }} // namespace at::_ops
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/adaptive_max_pool2d_native.h ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from NativeFunction.h
4
+
5
+ #include <c10/core/Scalar.h>
6
+ #include <c10/core/Storage.h>
7
+ #include <c10/core/TensorOptions.h>
8
+ #include <c10/util/Deprecated.h>
9
+ #include <c10/util/Optional.h>
10
+ #include <c10/core/QScheme.h>
11
+ #include <ATen/core/Reduction.h>
12
+ #include <ATen/core/Tensor.h>
13
+ #include <tuple>
14
+ #include <vector>
15
+ #include <ATen/ops/adaptive_max_pool2d_meta.h>
16
+
17
+ namespace at {
18
+ namespace native {
19
+ struct TORCH_API structured_adaptive_max_pool2d_out_cpu : public at::meta::structured_adaptive_max_pool2d {
20
+ void impl(const at::Tensor & self, at::IntArrayRef output_size, const at::Tensor & out, const at::Tensor & indices);
21
+ };
22
+ struct TORCH_API structured_adaptive_max_pool2d_out_cuda : public at::meta::structured_adaptive_max_pool2d {
23
+ void impl(const at::Tensor & self, at::IntArrayRef output_size, const at::Tensor & out, const at::Tensor & indices);
24
+ };
25
+ } // namespace native
26
+ } // namespace at
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/angle_cuda_dispatch.h ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+ // @generated by torchgen/gen.py from DispatchKeyFunction.h
3
+
4
+ // NB: The implementing C++ file is RegisterDispatchKey.cpp
5
+
6
+ // The only #includes we need are for custom classes that have defaults in the C++ API
7
+ #include <c10/core/MemoryFormat.h>
8
+ #include <c10/core/Scalar.h>
9
+ #include <ATen/core/Reduction.h>
10
+
11
+ // Forward declarations of any types needed in the operator signatures.
12
+ // We can't directly include these classes because it will cause circular include dependencies.
13
+ // This file is included by TensorBody.h, which defines the Tensor class.
14
+ #include <ATen/core/ATen_fwd.h>
15
+
16
+ namespace at {
17
+
18
+ namespace cuda {
19
+
20
+ TORCH_API at::Tensor angle(const at::Tensor & self);
21
+ TORCH_API at::Tensor & angle_out(at::Tensor & out, const at::Tensor & self);
22
+ TORCH_API at::Tensor & angle_outf(const at::Tensor & self, at::Tensor & out);
23
+
24
+ } // namespace cuda
25
+ } // namespace at
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/argsort_native.h ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from NativeFunction.h
4
+
5
+ #include <c10/core/Scalar.h>
6
+ #include <c10/core/Storage.h>
7
+ #include <c10/core/TensorOptions.h>
8
+ #include <c10/util/Deprecated.h>
9
+ #include <c10/util/Optional.h>
10
+ #include <c10/core/QScheme.h>
11
+ #include <ATen/core/Reduction.h>
12
+ #include <ATen/core/Tensor.h>
13
+ #include <tuple>
14
+ #include <vector>
15
+
16
+
17
+ namespace at {
18
+ namespace native {
19
+ TORCH_API at::Tensor argsort(const at::Tensor & self, int64_t dim=-1, bool descending=false);
20
+ TORCH_API at::Tensor & argsort_stable_out(const at::Tensor & self, bool stable, int64_t dim, bool descending, at::Tensor & out);
21
+ TORCH_API at::Tensor argsort_stable(const at::Tensor & self, bool stable, int64_t dim=-1, bool descending=false);
22
+ TORCH_API at::Tensor argsort(const at::Tensor & self, at::Dimname dim, bool descending=false);
23
+ } // namespace native
24
+ } // namespace at
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/cdist_native.h ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from NativeFunction.h
4
+
5
+ #include <c10/core/Scalar.h>
6
+ #include <c10/core/Storage.h>
7
+ #include <c10/core/TensorOptions.h>
8
+ #include <c10/util/Deprecated.h>
9
+ #include <c10/util/Optional.h>
10
+ #include <c10/core/QScheme.h>
11
+ #include <ATen/core/Reduction.h>
12
+ #include <ATen/core/Tensor.h>
13
+ #include <tuple>
14
+ #include <vector>
15
+
16
+
17
+ namespace at {
18
+ namespace native {
19
+ TORCH_API at::Tensor cdist(const at::Tensor & x1, const at::Tensor & x2, double p=2, c10::optional<int64_t> compute_mode=c10::nullopt);
20
+ } // namespace native
21
+ } // namespace at
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/col2im_ops.h ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from Operator.h
4
+
5
+ #include <tuple>
6
+ #include <vector>
7
+
8
+ // Forward declarations of any types needed in the operator signatures.
9
+ // We can't directly include these classes because it will cause circular include dependencies.
10
+ // This file is included by TensorBody.h, which defines the Tensor class.
11
+ #include <ATen/core/ATen_fwd.h>
12
+
13
+ namespace at {
14
+ namespace _ops {
15
+
16
+
17
+ struct TORCH_API col2im_out {
18
+ using schema = at::Tensor & (const at::Tensor &, c10::SymIntArrayRef, at::IntArrayRef, at::IntArrayRef, at::IntArrayRef, at::IntArrayRef, at::Tensor &);
19
+ using ptr_schema = schema*;
20
+ // See Note [static constexpr char* members for windows NVCC]
21
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(name, "aten::col2im")
22
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(overload_name, "out")
23
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(schema_str, "col2im.out(Tensor self, SymInt[2] output_size, int[2] kernel_size, int[2] dilation, int[2] padding, int[2] stride, *, Tensor(a!) out) -> Tensor(a!)")
24
+ static at::Tensor & call(const at::Tensor & self, c10::SymIntArrayRef output_size, at::IntArrayRef kernel_size, at::IntArrayRef dilation, at::IntArrayRef padding, at::IntArrayRef stride, at::Tensor & out);
25
+ static at::Tensor & redispatch(c10::DispatchKeySet dispatchKeySet, const at::Tensor & self, c10::SymIntArrayRef output_size, at::IntArrayRef kernel_size, at::IntArrayRef dilation, at::IntArrayRef padding, at::IntArrayRef stride, at::Tensor & out);
26
+ };
27
+
28
+ struct TORCH_API col2im {
29
+ using schema = at::Tensor (const at::Tensor &, c10::SymIntArrayRef, at::IntArrayRef, at::IntArrayRef, at::IntArrayRef, at::IntArrayRef);
30
+ using ptr_schema = schema*;
31
+ // See Note [static constexpr char* members for windows NVCC]
32
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(name, "aten::col2im")
33
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(overload_name, "")
34
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(schema_str, "col2im(Tensor self, SymInt[2] output_size, int[2] kernel_size, int[2] dilation, int[2] padding, int[2] stride) -> Tensor")
35
+ static at::Tensor call(const at::Tensor & self, c10::SymIntArrayRef output_size, at::IntArrayRef kernel_size, at::IntArrayRef dilation, at::IntArrayRef padding, at::IntArrayRef stride);
36
+ static at::Tensor redispatch(c10::DispatchKeySet dispatchKeySet, const at::Tensor & self, c10::SymIntArrayRef output_size, at::IntArrayRef kernel_size, at::IntArrayRef dilation, at::IntArrayRef padding, at::IntArrayRef stride);
37
+ };
38
+
39
+ }} // namespace at::_ops
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/copy_native.h ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from NativeFunction.h
4
+
5
+ #include <c10/core/Scalar.h>
6
+ #include <c10/core/Storage.h>
7
+ #include <c10/core/TensorOptions.h>
8
+ #include <c10/util/Deprecated.h>
9
+ #include <c10/util/Optional.h>
10
+ #include <c10/core/QScheme.h>
11
+ #include <ATen/core/Reduction.h>
12
+ #include <ATen/core/Tensor.h>
13
+ #include <tuple>
14
+ #include <vector>
15
+
16
+
17
+ namespace at {
18
+ namespace native {
19
+ TORCH_API at::Tensor & copy_out(const at::Tensor & self, const at::Tensor & src, bool non_blocking, at::Tensor & out);
20
+ TORCH_API at::Tensor & copy_(at::Tensor & self, const at::Tensor & src, bool non_blocking=false);
21
+ TORCH_API at::Tensor & copy_nested_(at::Tensor & self, const at::Tensor & src, bool non_blocking=false);
22
+ TORCH_API at::Tensor & copy_sparse_wrapper_(at::Tensor & self, const at::Tensor & src, bool non_blocking=false);
23
+ TORCH_API at::Tensor & copy_sparse_compressed_(at::Tensor & self, const at::Tensor & src, bool non_blocking=false);
24
+ TORCH_API at::Tensor & copy_mkldnn_(at::Tensor & self, const at::Tensor & src, bool non_blocking=false);
25
+ TORCH_API at::Tensor copy(const at::Tensor & self, const at::Tensor & src, bool non_blocking=false);
26
+ } // namespace native
27
+ } // namespace at
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/cudnn_grid_sampler_cuda_dispatch.h ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+ // @generated by torchgen/gen.py from DispatchKeyFunction.h
3
+
4
+ // NB: The implementing C++ file is RegisterDispatchKey.cpp
5
+
6
+ // The only #includes we need are for custom classes that have defaults in the C++ API
7
+ #include <c10/core/MemoryFormat.h>
8
+ #include <c10/core/Scalar.h>
9
+ #include <ATen/core/Reduction.h>
10
+
11
+ // Forward declarations of any types needed in the operator signatures.
12
+ // We can't directly include these classes because it will cause circular include dependencies.
13
+ // This file is included by TensorBody.h, which defines the Tensor class.
14
+ #include <ATen/core/ATen_fwd.h>
15
+
16
+ namespace at {
17
+
18
+ namespace cuda {
19
+
20
+ TORCH_API at::Tensor cudnn_grid_sampler(const at::Tensor & self, const at::Tensor & grid);
21
+
22
+ } // namespace cuda
23
+ } // namespace at
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/digamma_native.h ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from NativeFunction.h
4
+
5
+ #include <c10/core/Scalar.h>
6
+ #include <c10/core/Storage.h>
7
+ #include <c10/core/TensorOptions.h>
8
+ #include <c10/util/Deprecated.h>
9
+ #include <c10/util/Optional.h>
10
+ #include <c10/core/QScheme.h>
11
+ #include <ATen/core/Reduction.h>
12
+ #include <ATen/core/Tensor.h>
13
+ #include <tuple>
14
+ #include <vector>
15
+ #include <ATen/ops/digamma_meta.h>
16
+
17
+ namespace at {
18
+ namespace native {
19
+ struct TORCH_API structured_digamma_out : public at::meta::structured_digamma {
20
+ void impl(const at::Tensor & self, const at::Tensor & out);
21
+ };
22
+ } // namespace native
23
+ } // namespace at
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/dstack_ops.h ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from Operator.h
4
+
5
+ #include <tuple>
6
+ #include <vector>
7
+
8
+ // Forward declarations of any types needed in the operator signatures.
9
+ // We can't directly include these classes because it will cause circular include dependencies.
10
+ // This file is included by TensorBody.h, which defines the Tensor class.
11
+ #include <ATen/core/ATen_fwd.h>
12
+
13
+ namespace at {
14
+ namespace _ops {
15
+
16
+
17
+ struct TORCH_API dstack {
18
+ using schema = at::Tensor (at::TensorList);
19
+ using ptr_schema = schema*;
20
+ // See Note [static constexpr char* members for windows NVCC]
21
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(name, "aten::dstack")
22
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(overload_name, "")
23
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(schema_str, "dstack(Tensor[] tensors) -> Tensor")
24
+ static at::Tensor call(at::TensorList tensors);
25
+ static at::Tensor redispatch(c10::DispatchKeySet dispatchKeySet, at::TensorList tensors);
26
+ };
27
+
28
+ struct TORCH_API dstack_out {
29
+ using schema = at::Tensor & (at::TensorList, at::Tensor &);
30
+ using ptr_schema = schema*;
31
+ // See Note [static constexpr char* members for windows NVCC]
32
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(name, "aten::dstack")
33
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(overload_name, "out")
34
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(schema_str, "dstack.out(Tensor[] tensors, *, Tensor(a!) out) -> Tensor(a!)")
35
+ static at::Tensor & call(at::TensorList tensors, at::Tensor & out);
36
+ static at::Tensor & redispatch(c10::DispatchKeySet dispatchKeySet, at::TensorList tensors, at::Tensor & out);
37
+ };
38
+
39
+ }} // namespace at::_ops
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/embedding_backward_ops.h ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from Operator.h
4
+
5
+ #include <tuple>
6
+ #include <vector>
7
+
8
+ // Forward declarations of any types needed in the operator signatures.
9
+ // We can't directly include these classes because it will cause circular include dependencies.
10
+ // This file is included by TensorBody.h, which defines the Tensor class.
11
+ #include <ATen/core/ATen_fwd.h>
12
+
13
+ namespace at {
14
+ namespace _ops {
15
+
16
+
17
+ struct TORCH_API embedding_backward {
18
+ using schema = at::Tensor (const at::Tensor &, const at::Tensor &, c10::SymInt, c10::SymInt, bool, bool);
19
+ using ptr_schema = schema*;
20
+ // See Note [static constexpr char* members for windows NVCC]
21
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(name, "aten::embedding_backward")
22
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(overload_name, "")
23
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(schema_str, "embedding_backward(Tensor grad, Tensor indices, SymInt num_weights, SymInt padding_idx, bool scale_grad_by_freq, bool sparse) -> Tensor")
24
+ static at::Tensor call(const at::Tensor & grad, const at::Tensor & indices, c10::SymInt num_weights, c10::SymInt padding_idx, bool scale_grad_by_freq, bool sparse);
25
+ static at::Tensor redispatch(c10::DispatchKeySet dispatchKeySet, const at::Tensor & grad, const at::Tensor & indices, c10::SymInt num_weights, c10::SymInt padding_idx, bool scale_grad_by_freq, bool sparse);
26
+ };
27
+
28
+ }} // namespace at::_ops
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/exp_compositeexplicitautogradnonfunctional_dispatch.h ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+ // @generated by torchgen/gen.py from DispatchKeyFunction.h
3
+
4
+ // NB: The implementing C++ file is RegisterDispatchKey.cpp
5
+
6
+ // The only #includes we need are for custom classes that have defaults in the C++ API
7
+ #include <c10/core/MemoryFormat.h>
8
+ #include <c10/core/Scalar.h>
9
+ #include <ATen/core/Reduction.h>
10
+
11
+ // Forward declarations of any types needed in the operator signatures.
12
+ // We can't directly include these classes because it will cause circular include dependencies.
13
+ // This file is included by TensorBody.h, which defines the Tensor class.
14
+ #include <ATen/core/ATen_fwd.h>
15
+
16
+ namespace at {
17
+
18
+ namespace compositeexplicitautogradnonfunctional {
19
+
20
+ TORCH_API at::Tensor exp(const at::Tensor & self);
21
+ TORCH_API at::Tensor & exp_(at::Tensor & self);
22
+
23
+ } // namespace compositeexplicitautogradnonfunctional
24
+ } // namespace at
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/feature_alpha_dropout_compositeimplicitautograd_dispatch.h ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+ // @generated by torchgen/gen.py from DispatchKeyFunction.h
3
+
4
+ // NB: The implementing C++ file is RegisterDispatchKey.cpp
5
+
6
+ // The only #includes we need are for custom classes that have defaults in the C++ API
7
+ #include <c10/core/MemoryFormat.h>
8
+ #include <c10/core/Scalar.h>
9
+ #include <ATen/core/Reduction.h>
10
+
11
+ // Forward declarations of any types needed in the operator signatures.
12
+ // We can't directly include these classes because it will cause circular include dependencies.
13
+ // This file is included by TensorBody.h, which defines the Tensor class.
14
+ #include <ATen/core/ATen_fwd.h>
15
+
16
+ namespace at {
17
+
18
+ namespace compositeimplicitautograd {
19
+
20
+ TORCH_API at::Tensor feature_alpha_dropout(const at::Tensor & input, double p, bool train);
21
+ TORCH_API at::Tensor & feature_alpha_dropout_(at::Tensor & self, double p, bool train);
22
+
23
+ } // namespace compositeimplicitautograd
24
+ } // namespace at
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/flatten_dense_tensors.h ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from Function.h
4
+
5
+ #include <ATen/Context.h>
6
+ #include <ATen/DeviceGuard.h>
7
+ #include <ATen/TensorUtils.h>
8
+ #include <ATen/TracerMode.h>
9
+ #include <ATen/core/Generator.h>
10
+ #include <ATen/core/Reduction.h>
11
+ #include <ATen/core/Tensor.h>
12
+ #include <c10/core/Scalar.h>
13
+ #include <c10/core/Storage.h>
14
+ #include <c10/core/TensorOptions.h>
15
+ #include <c10/util/Deprecated.h>
16
+ #include <c10/util/Optional.h>
17
+
18
+
19
+
20
+ #include <ATen/ops/flatten_dense_tensors_ops.h>
21
+
22
+ namespace at {
23
+
24
+
25
+ // aten::flatten_dense_tensors(Tensor[] tensors) -> Tensor
26
+ inline at::Tensor flatten_dense_tensors(at::TensorList tensors) {
27
+ return at::_ops::flatten_dense_tensors::call(tensors);
28
+ }
29
+
30
+ }
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/flip_cpu_dispatch.h ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+ // @generated by torchgen/gen.py from DispatchKeyFunction.h
3
+
4
+ // NB: The implementing C++ file is RegisterDispatchKey.cpp
5
+
6
+ // The only #includes we need are for custom classes that have defaults in the C++ API
7
+ #include <c10/core/MemoryFormat.h>
8
+ #include <c10/core/Scalar.h>
9
+ #include <ATen/core/Reduction.h>
10
+
11
+ // Forward declarations of any types needed in the operator signatures.
12
+ // We can't directly include these classes because it will cause circular include dependencies.
13
+ // This file is included by TensorBody.h, which defines the Tensor class.
14
+ #include <ATen/core/ATen_fwd.h>
15
+
16
+ namespace at {
17
+
18
+ namespace cpu {
19
+
20
+ TORCH_API at::Tensor flip(const at::Tensor & self, at::IntArrayRef dims);
21
+
22
+ } // namespace cpu
23
+ } // namespace at
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/glu_ops.h ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from Operator.h
4
+
5
+ #include <tuple>
6
+ #include <vector>
7
+
8
+ // Forward declarations of any types needed in the operator signatures.
9
+ // We can't directly include these classes because it will cause circular include dependencies.
10
+ // This file is included by TensorBody.h, which defines the Tensor class.
11
+ #include <ATen/core/ATen_fwd.h>
12
+
13
+ namespace at {
14
+ namespace _ops {
15
+
16
+
17
+ struct TORCH_API glu_out {
18
+ using schema = at::Tensor & (const at::Tensor &, int64_t, at::Tensor &);
19
+ using ptr_schema = schema*;
20
+ // See Note [static constexpr char* members for windows NVCC]
21
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(name, "aten::glu")
22
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(overload_name, "out")
23
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(schema_str, "glu.out(Tensor self, int dim=-1, *, Tensor(a!) out) -> Tensor(a!)")
24
+ static at::Tensor & call(const at::Tensor & self, int64_t dim, at::Tensor & out);
25
+ static at::Tensor & redispatch(c10::DispatchKeySet dispatchKeySet, const at::Tensor & self, int64_t dim, at::Tensor & out);
26
+ };
27
+
28
+ struct TORCH_API glu {
29
+ using schema = at::Tensor (const at::Tensor &, int64_t);
30
+ using ptr_schema = schema*;
31
+ // See Note [static constexpr char* members for windows NVCC]
32
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(name, "aten::glu")
33
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(overload_name, "")
34
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(schema_str, "glu(Tensor self, int dim=-1) -> Tensor")
35
+ static at::Tensor call(const at::Tensor & self, int64_t dim);
36
+ static at::Tensor redispatch(c10::DispatchKeySet dispatchKeySet, const at::Tensor & self, int64_t dim);
37
+ };
38
+
39
+ }} // namespace at::_ops
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/grid_sampler_2d_compositeexplicitautograd_dispatch.h ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+ // @generated by torchgen/gen.py from DispatchKeyFunction.h
3
+
4
+ // NB: The implementing C++ file is RegisterDispatchKey.cpp
5
+
6
+ // The only #includes we need are for custom classes that have defaults in the C++ API
7
+ #include <c10/core/MemoryFormat.h>
8
+ #include <c10/core/Scalar.h>
9
+ #include <ATen/core/Reduction.h>
10
+
11
+ // Forward declarations of any types needed in the operator signatures.
12
+ // We can't directly include these classes because it will cause circular include dependencies.
13
+ // This file is included by TensorBody.h, which defines the Tensor class.
14
+ #include <ATen/core/ATen_fwd.h>
15
+
16
+ namespace at {
17
+
18
+ namespace compositeexplicitautograd {
19
+
20
+ TORCH_API at::Tensor & grid_sampler_2d_out(at::Tensor & out, const at::Tensor & input, const at::Tensor & grid, int64_t interpolation_mode, int64_t padding_mode, bool align_corners);
21
+ TORCH_API at::Tensor & grid_sampler_2d_outf(const at::Tensor & input, const at::Tensor & grid, int64_t interpolation_mode, int64_t padding_mode, bool align_corners, at::Tensor & out);
22
+
23
+ } // namespace compositeexplicitautograd
24
+ } // namespace at
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/hardswish_cuda_dispatch.h ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+ // @generated by torchgen/gen.py from DispatchKeyFunction.h
3
+
4
+ // NB: The implementing C++ file is RegisterDispatchKey.cpp
5
+
6
+ // The only #includes we need are for custom classes that have defaults in the C++ API
7
+ #include <c10/core/MemoryFormat.h>
8
+ #include <c10/core/Scalar.h>
9
+ #include <ATen/core/Reduction.h>
10
+
11
+ // Forward declarations of any types needed in the operator signatures.
12
+ // We can't directly include these classes because it will cause circular include dependencies.
13
+ // This file is included by TensorBody.h, which defines the Tensor class.
14
+ #include <ATen/core/ATen_fwd.h>
15
+
16
+ namespace at {
17
+
18
+ namespace cuda {
19
+
20
+ TORCH_API at::Tensor hardswish(const at::Tensor & self);
21
+ TORCH_API at::Tensor & hardswish_out(at::Tensor & out, const at::Tensor & self);
22
+ TORCH_API at::Tensor & hardswish_outf(const at::Tensor & self, at::Tensor & out);
23
+ TORCH_API at::Tensor & hardswish_(at::Tensor & self);
24
+
25
+ } // namespace cuda
26
+ } // namespace at
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/hardtanh_backward_native.h ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from NativeFunction.h
4
+
5
+ #include <c10/core/Scalar.h>
6
+ #include <c10/core/Storage.h>
7
+ #include <c10/core/TensorOptions.h>
8
+ #include <c10/util/Deprecated.h>
9
+ #include <c10/util/Optional.h>
10
+ #include <c10/core/QScheme.h>
11
+ #include <ATen/core/Reduction.h>
12
+ #include <ATen/core/Tensor.h>
13
+ #include <tuple>
14
+ #include <vector>
15
+
16
+
17
+ namespace at {
18
+ namespace native {
19
+ TORCH_API at::Tensor hardtanh_backward(const at::Tensor & grad_output, const at::Tensor & self, const at::Scalar & min_val, const at::Scalar & max_val);
20
+ TORCH_API at::Tensor & hardtanh_backward_out(const at::Tensor & grad_output, const at::Tensor & self, const at::Scalar & min_val, const at::Scalar & max_val, at::Tensor & grad_input);
21
+ } // namespace native
22
+ } // namespace at
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/histogramdd_compositeimplicitautograd_dispatch.h ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+ // @generated by torchgen/gen.py from DispatchKeyFunction.h
3
+
4
+ // NB: The implementing C++ file is RegisterDispatchKey.cpp
5
+
6
+ // The only #includes we need are for custom classes that have defaults in the C++ API
7
+ #include <c10/core/MemoryFormat.h>
8
+ #include <c10/core/Scalar.h>
9
+ #include <ATen/core/Reduction.h>
10
+
11
+ // Forward declarations of any types needed in the operator signatures.
12
+ // We can't directly include these classes because it will cause circular include dependencies.
13
+ // This file is included by TensorBody.h, which defines the Tensor class.
14
+ #include <ATen/core/ATen_fwd.h>
15
+
16
+ namespace at {
17
+
18
+ namespace compositeimplicitautograd {
19
+
20
+ TORCH_API ::std::tuple<at::Tensor,::std::vector<at::Tensor>> histogramdd(const at::Tensor & self, at::IntArrayRef bins, c10::optional<at::ArrayRef<double>> range=c10::nullopt, const c10::optional<at::Tensor> & weight={}, bool density=false);
21
+ TORCH_API ::std::tuple<at::Tensor,::std::vector<at::Tensor>> histogramdd(const at::Tensor & self, int64_t bins, c10::optional<at::ArrayRef<double>> range=c10::nullopt, const c10::optional<at::Tensor> & weight={}, bool density=false);
22
+ TORCH_API ::std::tuple<at::Tensor,::std::vector<at::Tensor>> histogramdd(const at::Tensor & self, at::TensorList bins, c10::optional<at::ArrayRef<double>> range=c10::nullopt, const c10::optional<at::Tensor> & weight={}, bool density=false);
23
+
24
+ } // namespace compositeimplicitautograd
25
+ } // namespace at
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/huber_loss_cuda_dispatch.h ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+ // @generated by torchgen/gen.py from DispatchKeyFunction.h
3
+
4
+ // NB: The implementing C++ file is RegisterDispatchKey.cpp
5
+
6
+ // The only #includes we need are for custom classes that have defaults in the C++ API
7
+ #include <c10/core/MemoryFormat.h>
8
+ #include <c10/core/Scalar.h>
9
+ #include <ATen/core/Reduction.h>
10
+
11
+ // Forward declarations of any types needed in the operator signatures.
12
+ // We can't directly include these classes because it will cause circular include dependencies.
13
+ // This file is included by TensorBody.h, which defines the Tensor class.
14
+ #include <ATen/core/ATen_fwd.h>
15
+
16
+ namespace at {
17
+
18
+ namespace cuda {
19
+
20
+ TORCH_API at::Tensor huber_loss(const at::Tensor & self, const at::Tensor & target, int64_t reduction=at::Reduction::Mean, double delta=1.0);
21
+ TORCH_API at::Tensor & huber_loss_out(at::Tensor & out, const at::Tensor & self, const at::Tensor & target, int64_t reduction=at::Reduction::Mean, double delta=1.0);
22
+ TORCH_API at::Tensor & huber_loss_outf(const at::Tensor & self, const at::Tensor & target, int64_t reduction, double delta, at::Tensor & out);
23
+
24
+ } // namespace cuda
25
+ } // namespace at
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/igammac_meta_dispatch.h ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+ // @generated by torchgen/gen.py from DispatchKeyFunction.h
3
+
4
+ // NB: The implementing C++ file is RegisterDispatchKey.cpp
5
+
6
+ // The only #includes we need are for custom classes that have defaults in the C++ API
7
+ #include <c10/core/MemoryFormat.h>
8
+ #include <c10/core/Scalar.h>
9
+ #include <ATen/core/Reduction.h>
10
+
11
+ // Forward declarations of any types needed in the operator signatures.
12
+ // We can't directly include these classes because it will cause circular include dependencies.
13
+ // This file is included by TensorBody.h, which defines the Tensor class.
14
+ #include <ATen/core/ATen_fwd.h>
15
+
16
+ namespace at {
17
+
18
+ namespace meta {
19
+
20
+ TORCH_API at::Tensor igammac(const at::Tensor & self, const at::Tensor & other);
21
+ TORCH_API at::Tensor & igammac_out(at::Tensor & out, const at::Tensor & self, const at::Tensor & other);
22
+ TORCH_API at::Tensor & igammac_outf(const at::Tensor & self, const at::Tensor & other, at::Tensor & out);
23
+ TORCH_API at::Tensor & igammac_(at::Tensor & self, const at::Tensor & other);
24
+
25
+ } // namespace meta
26
+ } // namespace at
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/instance_norm_native.h ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from NativeFunction.h
4
+
5
+ #include <c10/core/Scalar.h>
6
+ #include <c10/core/Storage.h>
7
+ #include <c10/core/TensorOptions.h>
8
+ #include <c10/util/Deprecated.h>
9
+ #include <c10/util/Optional.h>
10
+ #include <c10/core/QScheme.h>
11
+ #include <ATen/core/Reduction.h>
12
+ #include <ATen/core/Tensor.h>
13
+ #include <tuple>
14
+ #include <vector>
15
+
16
+
17
+ namespace at {
18
+ namespace native {
19
+ TORCH_API at::Tensor instance_norm(const at::Tensor & input, const c10::optional<at::Tensor> & weight, const c10::optional<at::Tensor> & bias, const c10::optional<at::Tensor> & running_mean, const c10::optional<at::Tensor> & running_var, bool use_input_stats, double momentum, double eps, bool cudnn_enabled);
20
+ } // namespace native
21
+ } // namespace at
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/int_repr_native.h ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from NativeFunction.h
4
+
5
+ #include <c10/core/Scalar.h>
6
+ #include <c10/core/Storage.h>
7
+ #include <c10/core/TensorOptions.h>
8
+ #include <c10/util/Deprecated.h>
9
+ #include <c10/util/Optional.h>
10
+ #include <c10/core/QScheme.h>
11
+ #include <ATen/core/Reduction.h>
12
+ #include <ATen/core/Tensor.h>
13
+ #include <tuple>
14
+ #include <vector>
15
+
16
+
17
+ namespace at {
18
+ namespace native {
19
+ TORCH_API at::Tensor & int_repr_out(const at::Tensor & self, at::Tensor & out);
20
+ TORCH_API at::Tensor int_repr_quantized_cpu(const at::Tensor & self);
21
+ TORCH_API at::Tensor int_repr_quantized_cuda(const at::Tensor & self);
22
+ } // namespace native
23
+ } // namespace at
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/is_distributed_ops.h ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from Operator.h
4
+
5
+ #include <tuple>
6
+ #include <vector>
7
+
8
+ // Forward declarations of any types needed in the operator signatures.
9
+ // We can't directly include these classes because it will cause circular include dependencies.
10
+ // This file is included by TensorBody.h, which defines the Tensor class.
11
+ #include <ATen/core/ATen_fwd.h>
12
+
13
+ namespace at {
14
+ namespace _ops {
15
+
16
+
17
+ struct TORCH_API is_distributed {
18
+ using schema = bool (const at::Tensor &);
19
+ using ptr_schema = schema*;
20
+ // See Note [static constexpr char* members for windows NVCC]
21
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(name, "aten::is_distributed")
22
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(overload_name, "")
23
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(schema_str, "is_distributed(Tensor self) -> bool")
24
+ static bool call(const at::Tensor & self);
25
+ static bool redispatch(c10::DispatchKeySet dispatchKeySet, const at::Tensor & self);
26
+ };
27
+
28
+ }} // namespace at::_ops
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/isnan_cpu_dispatch.h ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+ // @generated by torchgen/gen.py from DispatchKeyFunction.h
3
+
4
+ // NB: The implementing C++ file is RegisterDispatchKey.cpp
5
+
6
+ // The only #includes we need are for custom classes that have defaults in the C++ API
7
+ #include <c10/core/MemoryFormat.h>
8
+ #include <c10/core/Scalar.h>
9
+ #include <ATen/core/Reduction.h>
10
+
11
+ // Forward declarations of any types needed in the operator signatures.
12
+ // We can't directly include these classes because it will cause circular include dependencies.
13
+ // This file is included by TensorBody.h, which defines the Tensor class.
14
+ #include <ATen/core/ATen_fwd.h>
15
+
16
+ namespace at {
17
+
18
+ namespace cpu {
19
+
20
+ TORCH_API at::Tensor isnan(const at::Tensor & self);
21
+
22
+ } // namespace cpu
23
+ } // namespace at
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/isneginf_native.h ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from NativeFunction.h
4
+
5
+ #include <c10/core/Scalar.h>
6
+ #include <c10/core/Storage.h>
7
+ #include <c10/core/TensorOptions.h>
8
+ #include <c10/util/Deprecated.h>
9
+ #include <c10/util/Optional.h>
10
+ #include <c10/core/QScheme.h>
11
+ #include <ATen/core/Reduction.h>
12
+ #include <ATen/core/Tensor.h>
13
+ #include <tuple>
14
+ #include <vector>
15
+ #include <ATen/ops/isneginf_meta.h>
16
+
17
+ namespace at {
18
+ namespace native {
19
+ struct TORCH_API structured_isneginf_out : public at::meta::structured_isneginf {
20
+ void impl(const at::Tensor & self, const at::Tensor & out);
21
+ };
22
+ TORCH_API at::Tensor isneginf_sparse(const at::Tensor & self);
23
+ TORCH_API at::Tensor & isneginf_sparse_out(const at::Tensor & self, at::Tensor & out);
24
+ TORCH_API at::Tensor isneginf_sparse_csr(const at::Tensor & self);
25
+ TORCH_API at::Tensor & isneginf_sparse_csr_out(const at::Tensor & self, at::Tensor & out);
26
+ } // namespace native
27
+ } // namespace at
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/isposinf_meta_dispatch.h ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+ // @generated by torchgen/gen.py from DispatchKeyFunction.h
3
+
4
+ // NB: The implementing C++ file is RegisterDispatchKey.cpp
5
+
6
+ // The only #includes we need are for custom classes that have defaults in the C++ API
7
+ #include <c10/core/MemoryFormat.h>
8
+ #include <c10/core/Scalar.h>
9
+ #include <ATen/core/Reduction.h>
10
+
11
+ // Forward declarations of any types needed in the operator signatures.
12
+ // We can't directly include these classes because it will cause circular include dependencies.
13
+ // This file is included by TensorBody.h, which defines the Tensor class.
14
+ #include <ATen/core/ATen_fwd.h>
15
+
16
+ namespace at {
17
+
18
+ namespace meta {
19
+
20
+ TORCH_API at::Tensor isposinf(const at::Tensor & self);
21
+ TORCH_API at::Tensor & isposinf_out(at::Tensor & out, const at::Tensor & self);
22
+ TORCH_API at::Tensor & isposinf_outf(const at::Tensor & self, at::Tensor & out);
23
+
24
+ } // namespace meta
25
+ } // namespace at
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/le_compositeexplicitautogradnonfunctional_dispatch.h ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+ // @generated by torchgen/gen.py from DispatchKeyFunction.h
3
+
4
+ // NB: The implementing C++ file is RegisterDispatchKey.cpp
5
+
6
+ // The only #includes we need are for custom classes that have defaults in the C++ API
7
+ #include <c10/core/MemoryFormat.h>
8
+ #include <c10/core/Scalar.h>
9
+ #include <ATen/core/Reduction.h>
10
+
11
+ // Forward declarations of any types needed in the operator signatures.
12
+ // We can't directly include these classes because it will cause circular include dependencies.
13
+ // This file is included by TensorBody.h, which defines the Tensor class.
14
+ #include <ATen/core/ATen_fwd.h>
15
+
16
+ namespace at {
17
+
18
+ namespace compositeexplicitautogradnonfunctional {
19
+
20
+ TORCH_API at::Tensor le(const at::Tensor & self, const at::Scalar & other);
21
+ TORCH_API at::Tensor & le_(at::Tensor & self, const at::Scalar & other);
22
+ TORCH_API at::Tensor le(const at::Tensor & self, const at::Tensor & other);
23
+ TORCH_API at::Tensor & le_(at::Tensor & self, const at::Tensor & other);
24
+
25
+ } // namespace compositeexplicitautogradnonfunctional
26
+ } // namespace at
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/linalg_cholesky_ex_native.h ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from NativeFunction.h
4
+
5
+ #include <c10/core/Scalar.h>
6
+ #include <c10/core/Storage.h>
7
+ #include <c10/core/TensorOptions.h>
8
+ #include <c10/util/Deprecated.h>
9
+ #include <c10/util/Optional.h>
10
+ #include <c10/core/QScheme.h>
11
+ #include <ATen/core/Reduction.h>
12
+ #include <ATen/core/Tensor.h>
13
+ #include <tuple>
14
+ #include <vector>
15
+ #include <ATen/ops/linalg_cholesky_ex_meta.h>
16
+
17
+ namespace at {
18
+ namespace native {
19
+ struct TORCH_API structured_linalg_cholesky_ex_out : public at::meta::structured_linalg_cholesky_ex {
20
+ void impl(const at::Tensor & self, bool upper, bool check_errors, const at::Tensor & L, const at::Tensor & info);
21
+ };
22
+ } // namespace native
23
+ } // namespace at
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/linalg_eig_cuda_dispatch.h ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+ // @generated by torchgen/gen.py from DispatchKeyFunction.h
3
+
4
+ // NB: The implementing C++ file is RegisterDispatchKey.cpp
5
+
6
+ // The only #includes we need are for custom classes that have defaults in the C++ API
7
+ #include <c10/core/MemoryFormat.h>
8
+ #include <c10/core/Scalar.h>
9
+ #include <ATen/core/Reduction.h>
10
+
11
+ // Forward declarations of any types needed in the operator signatures.
12
+ // We can't directly include these classes because it will cause circular include dependencies.
13
+ // This file is included by TensorBody.h, which defines the Tensor class.
14
+ #include <ATen/core/ATen_fwd.h>
15
+
16
+ namespace at {
17
+
18
+ namespace cuda {
19
+
20
+ TORCH_API ::std::tuple<at::Tensor,at::Tensor> linalg_eig(const at::Tensor & self);
21
+ TORCH_API ::std::tuple<at::Tensor &,at::Tensor &> linalg_eig_out(at::Tensor & eigenvalues, at::Tensor & eigenvectors, const at::Tensor & self);
22
+ TORCH_API ::std::tuple<at::Tensor &,at::Tensor &> linalg_eig_outf(const at::Tensor & self, at::Tensor & eigenvalues, at::Tensor & eigenvectors);
23
+
24
+ } // namespace cuda
25
+ } // namespace at
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/linalg_eigh_native.h ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from NativeFunction.h
4
+
5
+ #include <c10/core/Scalar.h>
6
+ #include <c10/core/Storage.h>
7
+ #include <c10/core/TensorOptions.h>
8
+ #include <c10/util/Deprecated.h>
9
+ #include <c10/util/Optional.h>
10
+ #include <c10/core/QScheme.h>
11
+ #include <ATen/core/Reduction.h>
12
+ #include <ATen/core/Tensor.h>
13
+ #include <tuple>
14
+ #include <vector>
15
+
16
+
17
+ namespace at {
18
+ namespace native {
19
+ TORCH_API ::std::tuple<at::Tensor,at::Tensor> linalg_eigh(const at::Tensor & self, c10::string_view UPLO="L");
20
+ TORCH_API ::std::tuple<at::Tensor &,at::Tensor &> linalg_eigh_out(const at::Tensor & self, c10::string_view UPLO, at::Tensor & eigvals, at::Tensor & eigvecs);
21
+ } // namespace native
22
+ } // namespace at
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/linalg_lstsq.h ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from Function.h
4
+
5
+ #include <ATen/Context.h>
6
+ #include <ATen/DeviceGuard.h>
7
+ #include <ATen/TensorUtils.h>
8
+ #include <ATen/TracerMode.h>
9
+ #include <ATen/core/Generator.h>
10
+ #include <ATen/core/Reduction.h>
11
+ #include <ATen/core/Tensor.h>
12
+ #include <c10/core/Scalar.h>
13
+ #include <c10/core/Storage.h>
14
+ #include <c10/core/TensorOptions.h>
15
+ #include <c10/util/Deprecated.h>
16
+ #include <c10/util/Optional.h>
17
+
18
+
19
+
20
+ #include <ATen/ops/linalg_lstsq_ops.h>
21
+
22
+ namespace at {
23
+
24
+
25
+ // aten::linalg_lstsq(Tensor self, Tensor b, float? rcond=None, *, str? driver=None) -> (Tensor solution, Tensor residuals, Tensor rank, Tensor singular_values)
26
+ inline ::std::tuple<at::Tensor,at::Tensor,at::Tensor,at::Tensor> linalg_lstsq(const at::Tensor & self, const at::Tensor & b, c10::optional<double> rcond=c10::nullopt, c10::optional<c10::string_view> driver=c10::nullopt) {
27
+ return at::_ops::linalg_lstsq::call(self, b, rcond, driver);
28
+ }
29
+
30
+ // aten::linalg_lstsq.out(Tensor self, Tensor b, float? rcond=None, *, str? driver=None, Tensor(a!) solution, Tensor(b!) residuals, Tensor(c!) rank, Tensor(d!) singular_values) -> (Tensor(a!) solution, Tensor(b!) residuals, Tensor(c!) rank, Tensor(d!) singular_values)
31
+ inline ::std::tuple<at::Tensor &,at::Tensor &,at::Tensor &,at::Tensor &> linalg_lstsq_out(at::Tensor & solution, at::Tensor & residuals, at::Tensor & rank, at::Tensor & singular_values, const at::Tensor & self, const at::Tensor & b, c10::optional<double> rcond=c10::nullopt, c10::optional<c10::string_view> driver=c10::nullopt) {
32
+ return at::_ops::linalg_lstsq_out::call(self, b, rcond, driver, solution, residuals, rank, singular_values);
33
+ }
34
+ // aten::linalg_lstsq.out(Tensor self, Tensor b, float? rcond=None, *, str? driver=None, Tensor(a!) solution, Tensor(b!) residuals, Tensor(c!) rank, Tensor(d!) singular_values) -> (Tensor(a!) solution, Tensor(b!) residuals, Tensor(c!) rank, Tensor(d!) singular_values)
35
+ inline ::std::tuple<at::Tensor &,at::Tensor &,at::Tensor &,at::Tensor &> linalg_lstsq_outf(const at::Tensor & self, const at::Tensor & b, c10::optional<double> rcond, c10::optional<c10::string_view> driver, at::Tensor & solution, at::Tensor & residuals, at::Tensor & rank, at::Tensor & singular_values) {
36
+ return at::_ops::linalg_lstsq_out::call(self, b, rcond, driver, solution, residuals, rank, singular_values);
37
+ }
38
+
39
+ }
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/logaddexp_native.h ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from NativeFunction.h
4
+
5
+ #include <c10/core/Scalar.h>
6
+ #include <c10/core/Storage.h>
7
+ #include <c10/core/TensorOptions.h>
8
+ #include <c10/util/Deprecated.h>
9
+ #include <c10/util/Optional.h>
10
+ #include <c10/core/QScheme.h>
11
+ #include <ATen/core/Reduction.h>
12
+ #include <ATen/core/Tensor.h>
13
+ #include <tuple>
14
+ #include <vector>
15
+ #include <ATen/ops/logaddexp_meta.h>
16
+
17
+ namespace at {
18
+ namespace native {
19
+ struct TORCH_API structured_logaddexp_out : public at::meta::structured_logaddexp {
20
+ void impl(const at::Tensor & self, const at::Tensor & other, const at::Tensor & out);
21
+ };
22
+ } // namespace native
23
+ } // namespace at
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/logspace_native.h ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from NativeFunction.h
4
+
5
+ #include <c10/core/Scalar.h>
6
+ #include <c10/core/Storage.h>
7
+ #include <c10/core/TensorOptions.h>
8
+ #include <c10/util/Deprecated.h>
9
+ #include <c10/util/Optional.h>
10
+ #include <c10/core/QScheme.h>
11
+ #include <ATen/core/Reduction.h>
12
+ #include <ATen/core/Tensor.h>
13
+ #include <tuple>
14
+ #include <vector>
15
+
16
+
17
+ namespace at {
18
+ namespace native {
19
+ TORCH_API at::Tensor logspace(const at::Scalar & start, const at::Scalar & end, int64_t steps, double base=10.0, c10::optional<at::ScalarType> dtype={}, c10::optional<at::Layout> layout={}, c10::optional<at::Device> device={}, c10::optional<bool> pin_memory={});
20
+ TORCH_API at::Tensor & logspace_out(const at::Scalar & start, const at::Scalar & end, int64_t steps, double base, at::Tensor & out);
21
+ TORCH_API at::Tensor & logspace_cuda_out(const at::Scalar & start, const at::Scalar & end, int64_t steps, double base, at::Tensor & out);
22
+ TORCH_API at::Tensor logspace(const at::Tensor & start, const at::Tensor & end, int64_t steps, double base=10.0, c10::optional<at::ScalarType> dtype={}, c10::optional<at::Layout> layout={}, c10::optional<at::Device> device={}, c10::optional<bool> pin_memory={});
23
+ TORCH_API at::Tensor & logspace_out(const at::Tensor & start, const at::Tensor & end, int64_t steps, double base, at::Tensor & out);
24
+ TORCH_API at::Tensor logspace(const at::Tensor & start, const at::Scalar & end, int64_t steps, double base=10.0, c10::optional<at::ScalarType> dtype={}, c10::optional<at::Layout> layout={}, c10::optional<at::Device> device={}, c10::optional<bool> pin_memory={});
25
+ TORCH_API at::Tensor & logspace_out(const at::Tensor & start, const at::Scalar & end, int64_t steps, double base, at::Tensor & out);
26
+ TORCH_API at::Tensor logspace(const at::Scalar & start, const at::Tensor & end, int64_t steps, double base=10.0, c10::optional<at::ScalarType> dtype={}, c10::optional<at::Layout> layout={}, c10::optional<at::Device> device={}, c10::optional<bool> pin_memory={});
27
+ TORCH_API at::Tensor & logspace_out(const at::Scalar & start, const at::Tensor & end, int64_t steps, double base, at::Tensor & out);
28
+ } // namespace native
29
+ } // namespace at
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/margin_ranking_loss_ops.h ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from Operator.h
4
+
5
+ #include <tuple>
6
+ #include <vector>
7
+
8
+ // Forward declarations of any types needed in the operator signatures.
9
+ // We can't directly include these classes because it will cause circular include dependencies.
10
+ // This file is included by TensorBody.h, which defines the Tensor class.
11
+ #include <ATen/core/ATen_fwd.h>
12
+
13
+ namespace at {
14
+ namespace _ops {
15
+
16
+
17
+ struct TORCH_API margin_ranking_loss {
18
+ using schema = at::Tensor (const at::Tensor &, const at::Tensor &, const at::Tensor &, double, int64_t);
19
+ using ptr_schema = schema*;
20
+ // See Note [static constexpr char* members for windows NVCC]
21
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(name, "aten::margin_ranking_loss")
22
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(overload_name, "")
23
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(schema_str, "margin_ranking_loss(Tensor input1, Tensor input2, Tensor target, float margin=0.0, int reduction=Mean) -> Tensor")
24
+ static at::Tensor call(const at::Tensor & input1, const at::Tensor & input2, const at::Tensor & target, double margin, int64_t reduction);
25
+ static at::Tensor redispatch(c10::DispatchKeySet dispatchKeySet, const at::Tensor & input1, const at::Tensor & input2, const at::Tensor & target, double margin, int64_t reduction);
26
+ };
27
+
28
+ }} // namespace at::_ops
moondream/lib/python3.10/site-packages/torch/include/ATen/ops/mode_ops.h ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #pragma once
2
+
3
+ // @generated by torchgen/gen.py from Operator.h
4
+
5
+ #include <tuple>
6
+ #include <vector>
7
+
8
+ // Forward declarations of any types needed in the operator signatures.
9
+ // We can't directly include these classes because it will cause circular include dependencies.
10
+ // This file is included by TensorBody.h, which defines the Tensor class.
11
+ #include <ATen/core/ATen_fwd.h>
12
+
13
+ namespace at {
14
+ namespace _ops {
15
+
16
+
17
+ struct TORCH_API mode {
18
+ using schema = ::std::tuple<at::Tensor,at::Tensor> (const at::Tensor &, int64_t, bool);
19
+ using ptr_schema = schema*;
20
+ // See Note [static constexpr char* members for windows NVCC]
21
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(name, "aten::mode")
22
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(overload_name, "")
23
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(schema_str, "mode(Tensor self, int dim=-1, bool keepdim=False) -> (Tensor values, Tensor indices)")
24
+ static ::std::tuple<at::Tensor,at::Tensor> call(const at::Tensor & self, int64_t dim, bool keepdim);
25
+ static ::std::tuple<at::Tensor,at::Tensor> redispatch(c10::DispatchKeySet dispatchKeySet, const at::Tensor & self, int64_t dim, bool keepdim);
26
+ };
27
+
28
+ struct TORCH_API mode_values {
29
+ using schema = ::std::tuple<at::Tensor &,at::Tensor &> (const at::Tensor &, int64_t, bool, at::Tensor &, at::Tensor &);
30
+ using ptr_schema = schema*;
31
+ // See Note [static constexpr char* members for windows NVCC]
32
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(name, "aten::mode")
33
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(overload_name, "values")
34
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(schema_str, "mode.values(Tensor self, int dim=-1, bool keepdim=False, *, Tensor(a!) values, Tensor(b!) indices) -> (Tensor(a!) values, Tensor(b!) indices)")
35
+ static ::std::tuple<at::Tensor &,at::Tensor &> call(const at::Tensor & self, int64_t dim, bool keepdim, at::Tensor & values, at::Tensor & indices);
36
+ static ::std::tuple<at::Tensor &,at::Tensor &> redispatch(c10::DispatchKeySet dispatchKeySet, const at::Tensor & self, int64_t dim, bool keepdim, at::Tensor & values, at::Tensor & indices);
37
+ };
38
+
39
+ struct TORCH_API mode_dimname {
40
+ using schema = ::std::tuple<at::Tensor,at::Tensor> (const at::Tensor &, at::Dimname, bool);
41
+ using ptr_schema = schema*;
42
+ // See Note [static constexpr char* members for windows NVCC]
43
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(name, "aten::mode")
44
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(overload_name, "dimname")
45
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(schema_str, "mode.dimname(Tensor self, Dimname dim, bool keepdim=False) -> (Tensor values, Tensor indices)")
46
+ static ::std::tuple<at::Tensor,at::Tensor> call(const at::Tensor & self, at::Dimname dim, bool keepdim);
47
+ static ::std::tuple<at::Tensor,at::Tensor> redispatch(c10::DispatchKeySet dispatchKeySet, const at::Tensor & self, at::Dimname dim, bool keepdim);
48
+ };
49
+
50
+ struct TORCH_API mode_dimname_out {
51
+ using schema = ::std::tuple<at::Tensor &,at::Tensor &> (const at::Tensor &, at::Dimname, bool, at::Tensor &, at::Tensor &);
52
+ using ptr_schema = schema*;
53
+ // See Note [static constexpr char* members for windows NVCC]
54
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(name, "aten::mode")
55
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(overload_name, "dimname_out")
56
+ STATIC_CONSTEXPR_STR_INL_EXCEPT_WIN_CUDA(schema_str, "mode.dimname_out(Tensor self, Dimname dim, bool keepdim=False, *, Tensor(a!) values, Tensor(b!) indices) -> (Tensor(a!) values, Tensor(b!) indices)")
57
+ static ::std::tuple<at::Tensor &,at::Tensor &> call(const at::Tensor & self, at::Dimname dim, bool keepdim, at::Tensor & values, at::Tensor & indices);
58
+ static ::std::tuple<at::Tensor &,at::Tensor &> redispatch(c10::DispatchKeySet dispatchKeySet, const at::Tensor & self, at::Dimname dim, bool keepdim, at::Tensor & values, at::Tensor & indices);
59
+ };
60
+
61
+ }} // namespace at::_ops