nnilayy commited on
Commit
8d8f679
·
verified ·
1 Parent(s): add1da8

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +2 -0
  2. lib/python3.10/site-packages/av/codec/context.cpython-310-x86_64-linux-gnu.so +3 -0
  3. lib/python3.10/site-packages/av/subtitles/codeccontext.cpython-310-x86_64-linux-gnu.so +3 -0
  4. lib/python3.10/site-packages/keras_core/_tf_keras/activations/__init__.py +28 -0
  5. lib/python3.10/site-packages/keras_core/_tf_keras/constraints/__init__.py +19 -0
  6. lib/python3.10/site-packages/keras_core/_tf_keras/distribution/__init__.py +15 -0
  7. lib/python3.10/site-packages/keras_core/_tf_keras/initializers/__init__.py +43 -0
  8. lib/python3.10/site-packages/keras_core/_tf_keras/losses/__init__.py +66 -0
  9. lib/python3.10/site-packages/keras_core/_tf_keras/metrics/__init__.py +95 -0
  10. lib/python3.10/site-packages/keras_core/_tf_keras/mixed_precision/__init__.py +14 -0
  11. lib/python3.10/site-packages/keras_core/_tf_keras/random/__init__.py +15 -0
  12. lib/python3.10/site-packages/keras_core/_tf_keras/utils/__init__.py +44 -0
  13. lib/python3.10/site-packages/keras_core/_tf_keras/utils/legacy/__init__.py +9 -0
  14. lib/python3.10/site-packages/pandas/_libs/__init__.py +27 -0
  15. lib/python3.10/site-packages/pandas/_libs/arrays.pyi +40 -0
  16. lib/python3.10/site-packages/pandas/_libs/byteswap.cpython-310-x86_64-linux-gnu.so +0 -0
  17. lib/python3.10/site-packages/pandas/_libs/byteswap.pyi +5 -0
  18. lib/python3.10/site-packages/pandas/_libs/groupby.pyi +216 -0
  19. lib/python3.10/site-packages/pandas/_libs/hashing.pyi +9 -0
  20. lib/python3.10/site-packages/pandas/_libs/hashtable.pyi +252 -0
  21. lib/python3.10/site-packages/pandas/_libs/index.pyi +100 -0
  22. lib/python3.10/site-packages/pandas/_libs/indexing.cpython-310-x86_64-linux-gnu.so +0 -0
  23. lib/python3.10/site-packages/pandas/_libs/indexing.pyi +17 -0
  24. lib/python3.10/site-packages/pandas/_libs/internals.pyi +94 -0
  25. lib/python3.10/site-packages/pandas/_libs/interval.pyi +174 -0
  26. lib/python3.10/site-packages/pandas/_libs/join.pyi +79 -0
  27. lib/python3.10/site-packages/pandas/_libs/json.cpython-310-x86_64-linux-gnu.so +0 -0
  28. lib/python3.10/site-packages/pandas/_libs/json.pyi +23 -0
  29. lib/python3.10/site-packages/pandas/_libs/lib.pyi +213 -0
  30. lib/python3.10/site-packages/pandas/_libs/ops.pyi +51 -0
  31. lib/python3.10/site-packages/pandas/_libs/ops_dispatch.cpython-310-x86_64-linux-gnu.so +0 -0
  32. lib/python3.10/site-packages/pandas/_libs/ops_dispatch.pyi +5 -0
  33. lib/python3.10/site-packages/pandas/_libs/pandas_datetime.cpython-310-x86_64-linux-gnu.so +0 -0
  34. lib/python3.10/site-packages/pandas/_libs/pandas_parser.cpython-310-x86_64-linux-gnu.so +0 -0
  35. lib/python3.10/site-packages/pandas/_libs/parsers.pyi +77 -0
  36. lib/python3.10/site-packages/pandas/_libs/properties.cpython-310-x86_64-linux-gnu.so +0 -0
  37. lib/python3.10/site-packages/pandas/_libs/properties.pyi +27 -0
  38. lib/python3.10/site-packages/pandas/_libs/reshape.pyi +16 -0
  39. lib/python3.10/site-packages/pandas/_libs/sas.pyi +7 -0
  40. lib/python3.10/site-packages/pandas/_libs/sparse.pyi +51 -0
  41. lib/python3.10/site-packages/pandas/_libs/testing.pyi +12 -0
  42. lib/python3.10/site-packages/pandas/_libs/tslib.pyi +37 -0
  43. lib/python3.10/site-packages/pandas/_libs/tslibs/__init__.py +87 -0
  44. lib/python3.10/site-packages/pandas/_libs/tslibs/base.cpython-310-x86_64-linux-gnu.so +0 -0
  45. lib/python3.10/site-packages/pandas/_libs/tslibs/ccalendar.pyi +12 -0
  46. lib/python3.10/site-packages/pandas/_libs/tslibs/conversion.pyi +14 -0
  47. lib/python3.10/site-packages/pandas/_libs/tslibs/dtypes.pyi +83 -0
  48. lib/python3.10/site-packages/pandas/_libs/tslibs/fields.pyi +62 -0
  49. lib/python3.10/site-packages/pandas/_libs/tslibs/nattype.pyi +141 -0
  50. lib/python3.10/site-packages/pandas/_libs/tslibs/np_datetime.pyi +27 -0
.gitattributes CHANGED
@@ -116,3 +116,5 @@ lib/python3.10/site-packages/av/subtitles/stream.cpython-310-x86_64-linux-gnu.so
116
  lib/python3.10/site-packages/av/subtitles/subtitle.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
117
  lib/python3.10/site-packages/av/attachments/stream.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
118
  lib/python3.10/site-packages/av/codec/codec.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
 
 
 
116
  lib/python3.10/site-packages/av/subtitles/subtitle.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
117
  lib/python3.10/site-packages/av/attachments/stream.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
118
  lib/python3.10/site-packages/av/codec/codec.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
119
+ lib/python3.10/site-packages/av/codec/context.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
120
+ lib/python3.10/site-packages/av/subtitles/codeccontext.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
lib/python3.10/site-packages/av/codec/context.cpython-310-x86_64-linux-gnu.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee422ad46e430a883a7cfc973c94c2cbabad986b5603783181dae456ddf328fc
3
+ size 1270145
lib/python3.10/site-packages/av/subtitles/codeccontext.cpython-310-x86_64-linux-gnu.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:33e57873951bbccdbe24f57115e02409918748483bb32c8b678f37d0af7802c9
3
+ size 351201
lib/python3.10/site-packages/keras_core/_tf_keras/activations/__init__.py ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """DO NOT EDIT.
2
+
3
+ This file was autogenerated. Do not edit it by hand,
4
+ since your modifications would be overwritten.
5
+ """
6
+
7
+
8
+ from keras_core.src.activations import deserialize
9
+ from keras_core.src.activations import get
10
+ from keras_core.src.activations import serialize
11
+ from keras_core.src.activations.activations import elu
12
+ from keras_core.src.activations.activations import exponential
13
+ from keras_core.src.activations.activations import gelu
14
+ from keras_core.src.activations.activations import hard_sigmoid
15
+ from keras_core.src.activations.activations import leaky_relu
16
+ from keras_core.src.activations.activations import linear
17
+ from keras_core.src.activations.activations import log_softmax
18
+ from keras_core.src.activations.activations import mish
19
+ from keras_core.src.activations.activations import relu
20
+ from keras_core.src.activations.activations import relu6
21
+ from keras_core.src.activations.activations import selu
22
+ from keras_core.src.activations.activations import sigmoid
23
+ from keras_core.src.activations.activations import silu
24
+ from keras_core.src.activations.activations import silu as swish
25
+ from keras_core.src.activations.activations import softmax
26
+ from keras_core.src.activations.activations import softplus
27
+ from keras_core.src.activations.activations import softsign
28
+ from keras_core.src.activations.activations import tanh
lib/python3.10/site-packages/keras_core/_tf_keras/constraints/__init__.py ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """DO NOT EDIT.
2
+
3
+ This file was autogenerated. Do not edit it by hand,
4
+ since your modifications would be overwritten.
5
+ """
6
+
7
+
8
+ from keras_core.src.constraints import deserialize
9
+ from keras_core.src.constraints import get
10
+ from keras_core.src.constraints import serialize
11
+ from keras_core.src.constraints.constraints import Constraint
12
+ from keras_core.src.constraints.constraints import MaxNorm
13
+ from keras_core.src.constraints.constraints import MaxNorm as max_norm
14
+ from keras_core.src.constraints.constraints import MinMaxNorm
15
+ from keras_core.src.constraints.constraints import MinMaxNorm as min_max_norm
16
+ from keras_core.src.constraints.constraints import NonNeg
17
+ from keras_core.src.constraints.constraints import NonNeg as non_neg
18
+ from keras_core.src.constraints.constraints import UnitNorm
19
+ from keras_core.src.constraints.constraints import UnitNorm as unit_norm
lib/python3.10/site-packages/keras_core/_tf_keras/distribution/__init__.py ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """DO NOT EDIT.
2
+
3
+ This file was autogenerated. Do not edit it by hand,
4
+ since your modifications would be overwritten.
5
+ """
6
+
7
+
8
+ from keras_core.src.distribution.distribution_lib import DataParallel
9
+ from keras_core.src.distribution.distribution_lib import DeviceMesh
10
+ from keras_core.src.distribution.distribution_lib import LayoutMap
11
+ from keras_core.src.distribution.distribution_lib import ModelParallel
12
+ from keras_core.src.distribution.distribution_lib import TensorLayout
13
+ from keras_core.src.distribution.distribution_lib import distribution
14
+ from keras_core.src.distribution.distribution_lib import list_devices
15
+ from keras_core.src.distribution.distribution_lib import set_distribution
lib/python3.10/site-packages/keras_core/_tf_keras/initializers/__init__.py ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """DO NOT EDIT.
2
+
3
+ This file was autogenerated. Do not edit it by hand,
4
+ since your modifications would be overwritten.
5
+ """
6
+
7
+
8
+ from keras_core.src.initializers import deserialize
9
+ from keras_core.src.initializers import get
10
+ from keras_core.src.initializers import serialize
11
+ from keras_core.src.initializers.constant_initializers import Constant
12
+ from keras_core.src.initializers.constant_initializers import Constant as constant
13
+ from keras_core.src.initializers.constant_initializers import Identity
14
+ from keras_core.src.initializers.constant_initializers import Identity as IdentityInitializer
15
+ from keras_core.src.initializers.constant_initializers import Identity as identity
16
+ from keras_core.src.initializers.constant_initializers import Ones
17
+ from keras_core.src.initializers.constant_initializers import Ones as ones
18
+ from keras_core.src.initializers.constant_initializers import Zeros
19
+ from keras_core.src.initializers.constant_initializers import Zeros as zeros
20
+ from keras_core.src.initializers.initializer import Initializer
21
+ from keras_core.src.initializers.random_initializers import GlorotNormal
22
+ from keras_core.src.initializers.random_initializers import GlorotNormal as glorot_normal
23
+ from keras_core.src.initializers.random_initializers import GlorotUniform
24
+ from keras_core.src.initializers.random_initializers import GlorotUniform as glorot_uniform
25
+ from keras_core.src.initializers.random_initializers import HeNormal
26
+ from keras_core.src.initializers.random_initializers import HeNormal as he_normal
27
+ from keras_core.src.initializers.random_initializers import HeUniform
28
+ from keras_core.src.initializers.random_initializers import HeUniform as he_uniform
29
+ from keras_core.src.initializers.random_initializers import LecunNormal
30
+ from keras_core.src.initializers.random_initializers import LecunNormal as lecun_normal
31
+ from keras_core.src.initializers.random_initializers import LecunUniform
32
+ from keras_core.src.initializers.random_initializers import LecunUniform as lecun_uniform
33
+ from keras_core.src.initializers.random_initializers import OrthogonalInitializer
34
+ from keras_core.src.initializers.random_initializers import OrthogonalInitializer as Orthogonal
35
+ from keras_core.src.initializers.random_initializers import OrthogonalInitializer as orthogonal
36
+ from keras_core.src.initializers.random_initializers import RandomNormal
37
+ from keras_core.src.initializers.random_initializers import RandomNormal as random_normal
38
+ from keras_core.src.initializers.random_initializers import RandomUniform
39
+ from keras_core.src.initializers.random_initializers import RandomUniform as random_uniform
40
+ from keras_core.src.initializers.random_initializers import TruncatedNormal
41
+ from keras_core.src.initializers.random_initializers import TruncatedNormal as truncated_normal
42
+ from keras_core.src.initializers.random_initializers import VarianceScaling
43
+ from keras_core.src.initializers.random_initializers import VarianceScaling as variance_scaling
lib/python3.10/site-packages/keras_core/_tf_keras/losses/__init__.py ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """DO NOT EDIT.
2
+
3
+ This file was autogenerated. Do not edit it by hand,
4
+ since your modifications would be overwritten.
5
+ """
6
+
7
+
8
+ from keras_core.src.losses import deserialize
9
+ from keras_core.src.losses import get
10
+ from keras_core.src.losses import serialize
11
+ from keras_core.src.losses.loss import Loss
12
+ from keras_core.src.losses.losses import BinaryCrossentropy
13
+ from keras_core.src.losses.losses import BinaryFocalCrossentropy
14
+ from keras_core.src.losses.losses import CategoricalCrossentropy
15
+ from keras_core.src.losses.losses import CategoricalFocalCrossentropy
16
+ from keras_core.src.losses.losses import CategoricalHinge
17
+ from keras_core.src.losses.losses import CosineSimilarity
18
+ from keras_core.src.losses.losses import Hinge
19
+ from keras_core.src.losses.losses import Huber
20
+ from keras_core.src.losses.losses import KLDivergence
21
+ from keras_core.src.losses.losses import LogCosh
22
+ from keras_core.src.losses.losses import MeanAbsoluteError
23
+ from keras_core.src.losses.losses import MeanAbsolutePercentageError
24
+ from keras_core.src.losses.losses import MeanSquaredError
25
+ from keras_core.src.losses.losses import MeanSquaredLogarithmicError
26
+ from keras_core.src.losses.losses import Poisson
27
+ from keras_core.src.losses.losses import SparseCategoricalCrossentropy
28
+ from keras_core.src.losses.losses import SquaredHinge
29
+ from keras_core.src.losses.losses import binary_crossentropy
30
+ from keras_core.src.losses.losses import binary_focal_crossentropy
31
+ from keras_core.src.losses.losses import categorical_crossentropy
32
+ from keras_core.src.losses.losses import categorical_focal_crossentropy
33
+ from keras_core.src.losses.losses import categorical_hinge
34
+ from keras_core.src.losses.losses import cosine_similarity
35
+ from keras_core.src.losses.losses import hinge
36
+ from keras_core.src.losses.losses import huber
37
+ from keras_core.src.losses.losses import kl_divergence
38
+ from keras_core.src.losses.losses import log_cosh
39
+ from keras_core.src.losses.losses import mean_absolute_error
40
+ from keras_core.src.losses.losses import mean_absolute_percentage_error
41
+ from keras_core.src.losses.losses import mean_squared_error
42
+ from keras_core.src.losses.losses import mean_squared_logarithmic_error
43
+ from keras_core.src.losses.losses import poisson
44
+ from keras_core.src.losses.losses import sparse_categorical_crossentropy
45
+ from keras_core.src.losses.losses import squared_hinge
46
+
47
+ """DO NOT EDIT.
48
+
49
+ This file was autogenerated. Do not edit it by hand,
50
+ since your modifications would be overwritten.
51
+ """
52
+
53
+
54
+ from keras_core.src.legacy.losses import Reduction
55
+ from keras_core.src.losses.losses import kl_divergence as KLD
56
+ from keras_core.src.losses.losses import kl_divergence as kld
57
+ from keras_core.src.losses.losses import kl_divergence as kullback_leibler_divergence
58
+ from keras_core.src.losses.losses import log_cosh as logcosh
59
+ from keras_core.src.losses.losses import mean_absolute_error as MAE
60
+ from keras_core.src.losses.losses import mean_absolute_error as mae
61
+ from keras_core.src.losses.losses import mean_absolute_percentage_error as MAPE
62
+ from keras_core.src.losses.losses import mean_absolute_percentage_error as mape
63
+ from keras_core.src.losses.losses import mean_squared_error as MSE
64
+ from keras_core.src.losses.losses import mean_squared_error as mse
65
+ from keras_core.src.losses.losses import mean_squared_logarithmic_error as MSLE
66
+ from keras_core.src.losses.losses import mean_squared_logarithmic_error as msle
lib/python3.10/site-packages/keras_core/_tf_keras/metrics/__init__.py ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """DO NOT EDIT.
2
+
3
+ This file was autogenerated. Do not edit it by hand,
4
+ since your modifications would be overwritten.
5
+ """
6
+
7
+
8
+ from keras_core.src.losses.losses import binary_crossentropy
9
+ from keras_core.src.losses.losses import binary_focal_crossentropy
10
+ from keras_core.src.losses.losses import categorical_crossentropy
11
+ from keras_core.src.losses.losses import categorical_focal_crossentropy
12
+ from keras_core.src.losses.losses import categorical_hinge
13
+ from keras_core.src.losses.losses import hinge
14
+ from keras_core.src.losses.losses import huber
15
+ from keras_core.src.losses.losses import kl_divergence
16
+ from keras_core.src.losses.losses import log_cosh
17
+ from keras_core.src.losses.losses import mean_absolute_error
18
+ from keras_core.src.losses.losses import mean_absolute_percentage_error
19
+ from keras_core.src.losses.losses import mean_squared_error
20
+ from keras_core.src.losses.losses import mean_squared_logarithmic_error
21
+ from keras_core.src.losses.losses import poisson
22
+ from keras_core.src.losses.losses import sparse_categorical_crossentropy
23
+ from keras_core.src.losses.losses import squared_hinge
24
+ from keras_core.src.metrics import deserialize
25
+ from keras_core.src.metrics import get
26
+ from keras_core.src.metrics import serialize
27
+ from keras_core.src.metrics.accuracy_metrics import Accuracy
28
+ from keras_core.src.metrics.accuracy_metrics import BinaryAccuracy
29
+ from keras_core.src.metrics.accuracy_metrics import CategoricalAccuracy
30
+ from keras_core.src.metrics.accuracy_metrics import SparseCategoricalAccuracy
31
+ from keras_core.src.metrics.accuracy_metrics import SparseTopKCategoricalAccuracy
32
+ from keras_core.src.metrics.accuracy_metrics import TopKCategoricalAccuracy
33
+ from keras_core.src.metrics.accuracy_metrics import binary_accuracy
34
+ from keras_core.src.metrics.accuracy_metrics import categorical_accuracy
35
+ from keras_core.src.metrics.accuracy_metrics import sparse_categorical_accuracy
36
+ from keras_core.src.metrics.accuracy_metrics import sparse_top_k_categorical_accuracy
37
+ from keras_core.src.metrics.accuracy_metrics import top_k_categorical_accuracy
38
+ from keras_core.src.metrics.confusion_metrics import AUC
39
+ from keras_core.src.metrics.confusion_metrics import FalseNegatives
40
+ from keras_core.src.metrics.confusion_metrics import FalsePositives
41
+ from keras_core.src.metrics.confusion_metrics import Precision
42
+ from keras_core.src.metrics.confusion_metrics import PrecisionAtRecall
43
+ from keras_core.src.metrics.confusion_metrics import Recall
44
+ from keras_core.src.metrics.confusion_metrics import RecallAtPrecision
45
+ from keras_core.src.metrics.confusion_metrics import SensitivityAtSpecificity
46
+ from keras_core.src.metrics.confusion_metrics import SpecificityAtSensitivity
47
+ from keras_core.src.metrics.confusion_metrics import TrueNegatives
48
+ from keras_core.src.metrics.confusion_metrics import TruePositives
49
+ from keras_core.src.metrics.f_score_metrics import F1Score
50
+ from keras_core.src.metrics.f_score_metrics import FBetaScore
51
+ from keras_core.src.metrics.hinge_metrics import CategoricalHinge
52
+ from keras_core.src.metrics.hinge_metrics import Hinge
53
+ from keras_core.src.metrics.hinge_metrics import SquaredHinge
54
+ from keras_core.src.metrics.iou_metrics import BinaryIoU
55
+ from keras_core.src.metrics.iou_metrics import IoU
56
+ from keras_core.src.metrics.iou_metrics import MeanIoU
57
+ from keras_core.src.metrics.iou_metrics import OneHotIoU
58
+ from keras_core.src.metrics.iou_metrics import OneHotMeanIoU
59
+ from keras_core.src.metrics.metric import Metric
60
+ from keras_core.src.metrics.probabilistic_metrics import BinaryCrossentropy
61
+ from keras_core.src.metrics.probabilistic_metrics import CategoricalCrossentropy
62
+ from keras_core.src.metrics.probabilistic_metrics import KLDivergence
63
+ from keras_core.src.metrics.probabilistic_metrics import Poisson
64
+ from keras_core.src.metrics.probabilistic_metrics import SparseCategoricalCrossentropy
65
+ from keras_core.src.metrics.reduction_metrics import Mean
66
+ from keras_core.src.metrics.reduction_metrics import MeanMetricWrapper
67
+ from keras_core.src.metrics.reduction_metrics import Sum
68
+ from keras_core.src.metrics.regression_metrics import CosineSimilarity
69
+ from keras_core.src.metrics.regression_metrics import LogCoshError
70
+ from keras_core.src.metrics.regression_metrics import MeanAbsoluteError
71
+ from keras_core.src.metrics.regression_metrics import MeanAbsolutePercentageError
72
+ from keras_core.src.metrics.regression_metrics import MeanSquaredError
73
+ from keras_core.src.metrics.regression_metrics import MeanSquaredLogarithmicError
74
+ from keras_core.src.metrics.regression_metrics import R2Score
75
+ from keras_core.src.metrics.regression_metrics import RootMeanSquaredError
76
+
77
+ """DO NOT EDIT.
78
+
79
+ This file was autogenerated. Do not edit it by hand,
80
+ since your modifications would be overwritten.
81
+ """
82
+
83
+
84
+ from keras_core.src.losses.losses import kl_divergence as KLD
85
+ from keras_core.src.losses.losses import kl_divergence as kld
86
+ from keras_core.src.losses.losses import kl_divergence as kullback_leibler_divergence
87
+ from keras_core.src.losses.losses import log_cosh as logcosh
88
+ from keras_core.src.losses.losses import mean_absolute_error as MAE
89
+ from keras_core.src.losses.losses import mean_absolute_error as mae
90
+ from keras_core.src.losses.losses import mean_absolute_percentage_error as MAPE
91
+ from keras_core.src.losses.losses import mean_absolute_percentage_error as mape
92
+ from keras_core.src.losses.losses import mean_squared_error as MSE
93
+ from keras_core.src.losses.losses import mean_squared_error as mse
94
+ from keras_core.src.losses.losses import mean_squared_logarithmic_error as MSLE
95
+ from keras_core.src.losses.losses import mean_squared_logarithmic_error as msle
lib/python3.10/site-packages/keras_core/_tf_keras/mixed_precision/__init__.py ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """DO NOT EDIT.
2
+
3
+ This file was autogenerated. Do not edit it by hand,
4
+ since your modifications would be overwritten.
5
+ """
6
+
7
+
8
+ from keras_core.src.mixed_precision.dtype_policy import DTypePolicy
9
+ from keras_core.src.mixed_precision.dtype_policy import DTypePolicy as Policy
10
+ from keras_core.src.mixed_precision.dtype_policy import dtype_policy
11
+ from keras_core.src.mixed_precision.dtype_policy import dtype_policy as global_policy
12
+ from keras_core.src.mixed_precision.dtype_policy import set_dtype_policy
13
+ from keras_core.src.mixed_precision.dtype_policy import set_dtype_policy as set_global_policy
14
+ from keras_core.src.optimizers.loss_scale_optimizer import LossScaleOptimizer
lib/python3.10/site-packages/keras_core/_tf_keras/random/__init__.py ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """DO NOT EDIT.
2
+
3
+ This file was autogenerated. Do not edit it by hand,
4
+ since your modifications would be overwritten.
5
+ """
6
+
7
+
8
+ from keras_core.src.random.random import categorical
9
+ from keras_core.src.random.random import dropout
10
+ from keras_core.src.random.random import normal
11
+ from keras_core.src.random.random import randint
12
+ from keras_core.src.random.random import shuffle
13
+ from keras_core.src.random.random import truncated_normal
14
+ from keras_core.src.random.random import uniform
15
+ from keras_core.src.random.seed_generator import SeedGenerator
lib/python3.10/site-packages/keras_core/_tf_keras/utils/__init__.py ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """DO NOT EDIT.
2
+
3
+ This file was autogenerated. Do not edit it by hand,
4
+ since your modifications would be overwritten.
5
+ """
6
+
7
+
8
+ from keras_core.src.backend.common.global_state import clear_session
9
+ from keras_core.src.backend.common.keras_tensor import is_keras_tensor
10
+ from keras_core.src.layers.preprocessing.feature_space import FeatureSpace
11
+ from keras_core.src.ops.operation_utils import get_source_inputs
12
+ from keras_core.src.saving.object_registration import CustomObjectScope
13
+ from keras_core.src.saving.object_registration import CustomObjectScope as custom_object_scope
14
+ from keras_core.src.saving.object_registration import get_custom_objects
15
+ from keras_core.src.saving.object_registration import get_registered_name
16
+ from keras_core.src.saving.object_registration import get_registered_object
17
+ from keras_core.src.saving.object_registration import register_keras_serializable
18
+ from keras_core.src.saving.serialization_lib import deserialize_keras_object
19
+ from keras_core.src.saving.serialization_lib import serialize_keras_object
20
+ from keras_core.src.trainers.data_adapters.data_adapter_utils import pack_x_y_sample_weight
21
+ from keras_core.src.trainers.data_adapters.data_adapter_utils import unpack_x_y_sample_weight
22
+ from keras_core.src.trainers.data_adapters.py_dataset_adapter import PyDataset
23
+ from keras_core.src.trainers.data_adapters.py_dataset_adapter import PyDataset as Sequence
24
+ from keras_core.src.utils.audio_dataset_utils import audio_dataset_from_directory
25
+ from keras_core.src.utils.dataset_utils import split_dataset
26
+ from keras_core.src.utils.file_utils import get_file
27
+ from keras_core.src.utils.image_dataset_utils import image_dataset_from_directory
28
+ from keras_core.src.utils.image_utils import array_to_img
29
+ from keras_core.src.utils.image_utils import img_to_array
30
+ from keras_core.src.utils.image_utils import load_img
31
+ from keras_core.src.utils.image_utils import save_img
32
+ from keras_core.src.utils.io_utils import disable_interactive_logging
33
+ from keras_core.src.utils.io_utils import enable_interactive_logging
34
+ from keras_core.src.utils.io_utils import is_interactive_logging_enabled
35
+ from keras_core.src.utils.model_visualization import model_to_dot
36
+ from keras_core.src.utils.model_visualization import plot_model
37
+ from keras_core.src.utils.numerical_utils import normalize
38
+ from keras_core.src.utils.numerical_utils import to_categorical
39
+ from keras_core.src.utils.progbar import Progbar
40
+ from keras_core.src.utils.rng_utils import set_random_seed
41
+ from keras_core.src.utils.sequence_utils import pad_sequences
42
+ from keras_core.src.utils.text_dataset_utils import text_dataset_from_directory
43
+ from keras_core.src.utils.timeseries_dataset_utils import timeseries_dataset_from_array
44
+ from keras_core.utils import legacy
lib/python3.10/site-packages/keras_core/_tf_keras/utils/legacy/__init__.py ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ """DO NOT EDIT.
2
+
3
+ This file was autogenerated. Do not edit it by hand,
4
+ since your modifications would be overwritten.
5
+ """
6
+
7
+
8
+ from keras_core.src.legacy.saving.serialization import deserialize_keras_object
9
+ from keras_core.src.legacy.saving.serialization import serialize_keras_object
lib/python3.10/site-packages/pandas/_libs/__init__.py ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ __all__ = [
2
+ "NaT",
3
+ "NaTType",
4
+ "OutOfBoundsDatetime",
5
+ "Period",
6
+ "Timedelta",
7
+ "Timestamp",
8
+ "iNaT",
9
+ "Interval",
10
+ ]
11
+
12
+
13
+ # Below imports needs to happen first to ensure pandas top level
14
+ # module gets monkeypatched with the pandas_datetime_CAPI
15
+ # see pandas_datetime_exec in pd_datetime.c
16
+ import pandas._libs.pandas_parser # isort: skip # type: ignore[reportUnusedImport]
17
+ import pandas._libs.pandas_datetime # noqa: F401 # isort: skip # type: ignore[reportUnusedImport]
18
+ from pandas._libs.interval import Interval
19
+ from pandas._libs.tslibs import (
20
+ NaT,
21
+ NaTType,
22
+ OutOfBoundsDatetime,
23
+ Period,
24
+ Timedelta,
25
+ Timestamp,
26
+ iNaT,
27
+ )
lib/python3.10/site-packages/pandas/_libs/arrays.pyi ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from typing import Sequence
2
+
3
+ import numpy as np
4
+
5
+ from pandas._typing import (
6
+ AxisInt,
7
+ DtypeObj,
8
+ Self,
9
+ Shape,
10
+ )
11
+
12
+ class NDArrayBacked:
13
+ _dtype: DtypeObj
14
+ _ndarray: np.ndarray
15
+ def __init__(self, values: np.ndarray, dtype: DtypeObj) -> None: ...
16
+ @classmethod
17
+ def _simple_new(cls, values: np.ndarray, dtype: DtypeObj): ...
18
+ def _from_backing_data(self, values: np.ndarray): ...
19
+ def __setstate__(self, state): ...
20
+ def __len__(self) -> int: ...
21
+ @property
22
+ def shape(self) -> Shape: ...
23
+ @property
24
+ def ndim(self) -> int: ...
25
+ @property
26
+ def size(self) -> int: ...
27
+ @property
28
+ def nbytes(self) -> int: ...
29
+ def copy(self, order=...): ...
30
+ def delete(self, loc, axis=...): ...
31
+ def swapaxes(self, axis1, axis2): ...
32
+ def repeat(self, repeats: int | Sequence[int], axis: int | None = ...): ...
33
+ def reshape(self, *args, **kwargs): ...
34
+ def ravel(self, order=...): ...
35
+ @property
36
+ def T(self): ...
37
+ @classmethod
38
+ def _concat_same_type(
39
+ cls, to_concat: Sequence[Self], axis: AxisInt = ...
40
+ ) -> Self: ...
lib/python3.10/site-packages/pandas/_libs/byteswap.cpython-310-x86_64-linux-gnu.so ADDED
Binary file (61.7 kB). View file
 
lib/python3.10/site-packages/pandas/_libs/byteswap.pyi ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ def read_float_with_byteswap(data: bytes, offset: int, byteswap: bool) -> float: ...
2
+ def read_double_with_byteswap(data: bytes, offset: int, byteswap: bool) -> float: ...
3
+ def read_uint16_with_byteswap(data: bytes, offset: int, byteswap: bool) -> int: ...
4
+ def read_uint32_with_byteswap(data: bytes, offset: int, byteswap: bool) -> int: ...
5
+ def read_uint64_with_byteswap(data: bytes, offset: int, byteswap: bool) -> int: ...
lib/python3.10/site-packages/pandas/_libs/groupby.pyi ADDED
@@ -0,0 +1,216 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from typing import Literal
2
+
3
+ import numpy as np
4
+
5
+ from pandas._typing import npt
6
+
7
+ def group_median_float64(
8
+ out: np.ndarray, # ndarray[float64_t, ndim=2]
9
+ counts: npt.NDArray[np.int64],
10
+ values: np.ndarray, # ndarray[float64_t, ndim=2]
11
+ labels: npt.NDArray[np.int64],
12
+ min_count: int = ..., # Py_ssize_t
13
+ mask: np.ndarray | None = ...,
14
+ result_mask: np.ndarray | None = ...,
15
+ ) -> None: ...
16
+ def group_cumprod(
17
+ out: np.ndarray, # float64_t[:, ::1]
18
+ values: np.ndarray, # const float64_t[:, :]
19
+ labels: np.ndarray, # const int64_t[:]
20
+ ngroups: int,
21
+ is_datetimelike: bool,
22
+ skipna: bool = ...,
23
+ mask: np.ndarray | None = ...,
24
+ result_mask: np.ndarray | None = ...,
25
+ ) -> None: ...
26
+ def group_cumsum(
27
+ out: np.ndarray, # int64float_t[:, ::1]
28
+ values: np.ndarray, # ndarray[int64float_t, ndim=2]
29
+ labels: np.ndarray, # const int64_t[:]
30
+ ngroups: int,
31
+ is_datetimelike: bool,
32
+ skipna: bool = ...,
33
+ mask: np.ndarray | None = ...,
34
+ result_mask: np.ndarray | None = ...,
35
+ ) -> None: ...
36
+ def group_shift_indexer(
37
+ out: np.ndarray, # int64_t[::1]
38
+ labels: np.ndarray, # const int64_t[:]
39
+ ngroups: int,
40
+ periods: int,
41
+ ) -> None: ...
42
+ def group_fillna_indexer(
43
+ out: np.ndarray, # ndarray[intp_t]
44
+ labels: np.ndarray, # ndarray[int64_t]
45
+ sorted_labels: npt.NDArray[np.intp],
46
+ mask: npt.NDArray[np.uint8],
47
+ limit: int, # int64_t
48
+ dropna: bool,
49
+ ) -> None: ...
50
+ def group_any_all(
51
+ out: np.ndarray, # uint8_t[::1]
52
+ values: np.ndarray, # const uint8_t[::1]
53
+ labels: np.ndarray, # const int64_t[:]
54
+ mask: np.ndarray, # const uint8_t[::1]
55
+ val_test: Literal["any", "all"],
56
+ skipna: bool,
57
+ result_mask: np.ndarray | None,
58
+ ) -> None: ...
59
+ def group_sum(
60
+ out: np.ndarray, # complexfloatingintuint_t[:, ::1]
61
+ counts: np.ndarray, # int64_t[::1]
62
+ values: np.ndarray, # ndarray[complexfloatingintuint_t, ndim=2]
63
+ labels: np.ndarray, # const intp_t[:]
64
+ mask: np.ndarray | None,
65
+ result_mask: np.ndarray | None = ...,
66
+ min_count: int = ...,
67
+ is_datetimelike: bool = ...,
68
+ ) -> None: ...
69
+ def group_prod(
70
+ out: np.ndarray, # int64float_t[:, ::1]
71
+ counts: np.ndarray, # int64_t[::1]
72
+ values: np.ndarray, # ndarray[int64float_t, ndim=2]
73
+ labels: np.ndarray, # const intp_t[:]
74
+ mask: np.ndarray | None,
75
+ result_mask: np.ndarray | None = ...,
76
+ min_count: int = ...,
77
+ ) -> None: ...
78
+ def group_var(
79
+ out: np.ndarray, # floating[:, ::1]
80
+ counts: np.ndarray, # int64_t[::1]
81
+ values: np.ndarray, # ndarray[floating, ndim=2]
82
+ labels: np.ndarray, # const intp_t[:]
83
+ min_count: int = ..., # Py_ssize_t
84
+ ddof: int = ..., # int64_t
85
+ mask: np.ndarray | None = ...,
86
+ result_mask: np.ndarray | None = ...,
87
+ is_datetimelike: bool = ...,
88
+ name: str = ...,
89
+ ) -> None: ...
90
+ def group_skew(
91
+ out: np.ndarray, # float64_t[:, ::1]
92
+ counts: np.ndarray, # int64_t[::1]
93
+ values: np.ndarray, # ndarray[float64_T, ndim=2]
94
+ labels: np.ndarray, # const intp_t[::1]
95
+ mask: np.ndarray | None = ...,
96
+ result_mask: np.ndarray | None = ...,
97
+ skipna: bool = ...,
98
+ ) -> None: ...
99
+ def group_mean(
100
+ out: np.ndarray, # floating[:, ::1]
101
+ counts: np.ndarray, # int64_t[::1]
102
+ values: np.ndarray, # ndarray[floating, ndim=2]
103
+ labels: np.ndarray, # const intp_t[:]
104
+ min_count: int = ..., # Py_ssize_t
105
+ is_datetimelike: bool = ..., # bint
106
+ mask: np.ndarray | None = ...,
107
+ result_mask: np.ndarray | None = ...,
108
+ ) -> None: ...
109
+ def group_ohlc(
110
+ out: np.ndarray, # floatingintuint_t[:, ::1]
111
+ counts: np.ndarray, # int64_t[::1]
112
+ values: np.ndarray, # ndarray[floatingintuint_t, ndim=2]
113
+ labels: np.ndarray, # const intp_t[:]
114
+ min_count: int = ...,
115
+ mask: np.ndarray | None = ...,
116
+ result_mask: np.ndarray | None = ...,
117
+ ) -> None: ...
118
+ def group_quantile(
119
+ out: npt.NDArray[np.float64],
120
+ values: np.ndarray, # ndarray[numeric, ndim=1]
121
+ labels: npt.NDArray[np.intp],
122
+ mask: npt.NDArray[np.uint8],
123
+ qs: npt.NDArray[np.float64], # const
124
+ starts: npt.NDArray[np.int64],
125
+ ends: npt.NDArray[np.int64],
126
+ interpolation: Literal["linear", "lower", "higher", "nearest", "midpoint"],
127
+ result_mask: np.ndarray | None,
128
+ is_datetimelike: bool,
129
+ ) -> None: ...
130
+ def group_last(
131
+ out: np.ndarray, # rank_t[:, ::1]
132
+ counts: np.ndarray, # int64_t[::1]
133
+ values: np.ndarray, # ndarray[rank_t, ndim=2]
134
+ labels: np.ndarray, # const int64_t[:]
135
+ mask: npt.NDArray[np.bool_] | None,
136
+ result_mask: npt.NDArray[np.bool_] | None = ...,
137
+ min_count: int = ..., # Py_ssize_t
138
+ is_datetimelike: bool = ...,
139
+ skipna: bool = ...,
140
+ ) -> None: ...
141
+ def group_nth(
142
+ out: np.ndarray, # rank_t[:, ::1]
143
+ counts: np.ndarray, # int64_t[::1]
144
+ values: np.ndarray, # ndarray[rank_t, ndim=2]
145
+ labels: np.ndarray, # const int64_t[:]
146
+ mask: npt.NDArray[np.bool_] | None,
147
+ result_mask: npt.NDArray[np.bool_] | None = ...,
148
+ min_count: int = ..., # int64_t
149
+ rank: int = ..., # int64_t
150
+ is_datetimelike: bool = ...,
151
+ skipna: bool = ...,
152
+ ) -> None: ...
153
+ def group_rank(
154
+ out: np.ndarray, # float64_t[:, ::1]
155
+ values: np.ndarray, # ndarray[rank_t, ndim=2]
156
+ labels: np.ndarray, # const int64_t[:]
157
+ ngroups: int,
158
+ is_datetimelike: bool,
159
+ ties_method: Literal["average", "min", "max", "first", "dense"] = ...,
160
+ ascending: bool = ...,
161
+ pct: bool = ...,
162
+ na_option: Literal["keep", "top", "bottom"] = ...,
163
+ mask: npt.NDArray[np.bool_] | None = ...,
164
+ ) -> None: ...
165
+ def group_max(
166
+ out: np.ndarray, # groupby_t[:, ::1]
167
+ counts: np.ndarray, # int64_t[::1]
168
+ values: np.ndarray, # ndarray[groupby_t, ndim=2]
169
+ labels: np.ndarray, # const int64_t[:]
170
+ min_count: int = ...,
171
+ is_datetimelike: bool = ...,
172
+ mask: np.ndarray | None = ...,
173
+ result_mask: np.ndarray | None = ...,
174
+ ) -> None: ...
175
+ def group_min(
176
+ out: np.ndarray, # groupby_t[:, ::1]
177
+ counts: np.ndarray, # int64_t[::1]
178
+ values: np.ndarray, # ndarray[groupby_t, ndim=2]
179
+ labels: np.ndarray, # const int64_t[:]
180
+ min_count: int = ...,
181
+ is_datetimelike: bool = ...,
182
+ mask: np.ndarray | None = ...,
183
+ result_mask: np.ndarray | None = ...,
184
+ ) -> None: ...
185
+ def group_idxmin_idxmax(
186
+ out: npt.NDArray[np.intp],
187
+ counts: npt.NDArray[np.int64],
188
+ values: np.ndarray, # ndarray[groupby_t, ndim=2]
189
+ labels: npt.NDArray[np.intp],
190
+ min_count: int = ...,
191
+ is_datetimelike: bool = ...,
192
+ mask: np.ndarray | None = ...,
193
+ name: str = ...,
194
+ skipna: bool = ...,
195
+ result_mask: np.ndarray | None = ...,
196
+ ) -> None: ...
197
+ def group_cummin(
198
+ out: np.ndarray, # groupby_t[:, ::1]
199
+ values: np.ndarray, # ndarray[groupby_t, ndim=2]
200
+ labels: np.ndarray, # const int64_t[:]
201
+ ngroups: int,
202
+ is_datetimelike: bool,
203
+ mask: np.ndarray | None = ...,
204
+ result_mask: np.ndarray | None = ...,
205
+ skipna: bool = ...,
206
+ ) -> None: ...
207
+ def group_cummax(
208
+ out: np.ndarray, # groupby_t[:, ::1]
209
+ values: np.ndarray, # ndarray[groupby_t, ndim=2]
210
+ labels: np.ndarray, # const int64_t[:]
211
+ ngroups: int,
212
+ is_datetimelike: bool,
213
+ mask: np.ndarray | None = ...,
214
+ result_mask: np.ndarray | None = ...,
215
+ skipna: bool = ...,
216
+ ) -> None: ...
lib/python3.10/site-packages/pandas/_libs/hashing.pyi ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ import numpy as np
2
+
3
+ from pandas._typing import npt
4
+
5
+ def hash_object_array(
6
+ arr: npt.NDArray[np.object_],
7
+ key: str,
8
+ encoding: str = ...,
9
+ ) -> npt.NDArray[np.uint64]: ...
lib/python3.10/site-packages/pandas/_libs/hashtable.pyi ADDED
@@ -0,0 +1,252 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from typing import (
2
+ Any,
3
+ Hashable,
4
+ Literal,
5
+ )
6
+
7
+ import numpy as np
8
+
9
+ from pandas._typing import npt
10
+
11
+ def unique_label_indices(
12
+ labels: np.ndarray, # const int64_t[:]
13
+ ) -> np.ndarray: ...
14
+
15
+ class Factorizer:
16
+ count: int
17
+ uniques: Any
18
+ def __init__(self, size_hint: int) -> None: ...
19
+ def get_count(self) -> int: ...
20
+ def factorize(
21
+ self,
22
+ values: np.ndarray,
23
+ na_sentinel=...,
24
+ na_value=...,
25
+ mask=...,
26
+ ) -> npt.NDArray[np.intp]: ...
27
+
28
+ class ObjectFactorizer(Factorizer):
29
+ table: PyObjectHashTable
30
+ uniques: ObjectVector
31
+
32
+ class Int64Factorizer(Factorizer):
33
+ table: Int64HashTable
34
+ uniques: Int64Vector
35
+
36
+ class UInt64Factorizer(Factorizer):
37
+ table: UInt64HashTable
38
+ uniques: UInt64Vector
39
+
40
+ class Int32Factorizer(Factorizer):
41
+ table: Int32HashTable
42
+ uniques: Int32Vector
43
+
44
+ class UInt32Factorizer(Factorizer):
45
+ table: UInt32HashTable
46
+ uniques: UInt32Vector
47
+
48
+ class Int16Factorizer(Factorizer):
49
+ table: Int16HashTable
50
+ uniques: Int16Vector
51
+
52
+ class UInt16Factorizer(Factorizer):
53
+ table: UInt16HashTable
54
+ uniques: UInt16Vector
55
+
56
+ class Int8Factorizer(Factorizer):
57
+ table: Int8HashTable
58
+ uniques: Int8Vector
59
+
60
+ class UInt8Factorizer(Factorizer):
61
+ table: UInt8HashTable
62
+ uniques: UInt8Vector
63
+
64
+ class Float64Factorizer(Factorizer):
65
+ table: Float64HashTable
66
+ uniques: Float64Vector
67
+
68
+ class Float32Factorizer(Factorizer):
69
+ table: Float32HashTable
70
+ uniques: Float32Vector
71
+
72
+ class Complex64Factorizer(Factorizer):
73
+ table: Complex64HashTable
74
+ uniques: Complex64Vector
75
+
76
+ class Complex128Factorizer(Factorizer):
77
+ table: Complex128HashTable
78
+ uniques: Complex128Vector
79
+
80
+ class Int64Vector:
81
+ def __init__(self, *args) -> None: ...
82
+ def __len__(self) -> int: ...
83
+ def to_array(self) -> npt.NDArray[np.int64]: ...
84
+
85
+ class Int32Vector:
86
+ def __init__(self, *args) -> None: ...
87
+ def __len__(self) -> int: ...
88
+ def to_array(self) -> npt.NDArray[np.int32]: ...
89
+
90
+ class Int16Vector:
91
+ def __init__(self, *args) -> None: ...
92
+ def __len__(self) -> int: ...
93
+ def to_array(self) -> npt.NDArray[np.int16]: ...
94
+
95
+ class Int8Vector:
96
+ def __init__(self, *args) -> None: ...
97
+ def __len__(self) -> int: ...
98
+ def to_array(self) -> npt.NDArray[np.int8]: ...
99
+
100
+ class UInt64Vector:
101
+ def __init__(self, *args) -> None: ...
102
+ def __len__(self) -> int: ...
103
+ def to_array(self) -> npt.NDArray[np.uint64]: ...
104
+
105
+ class UInt32Vector:
106
+ def __init__(self, *args) -> None: ...
107
+ def __len__(self) -> int: ...
108
+ def to_array(self) -> npt.NDArray[np.uint32]: ...
109
+
110
+ class UInt16Vector:
111
+ def __init__(self, *args) -> None: ...
112
+ def __len__(self) -> int: ...
113
+ def to_array(self) -> npt.NDArray[np.uint16]: ...
114
+
115
+ class UInt8Vector:
116
+ def __init__(self, *args) -> None: ...
117
+ def __len__(self) -> int: ...
118
+ def to_array(self) -> npt.NDArray[np.uint8]: ...
119
+
120
+ class Float64Vector:
121
+ def __init__(self, *args) -> None: ...
122
+ def __len__(self) -> int: ...
123
+ def to_array(self) -> npt.NDArray[np.float64]: ...
124
+
125
+ class Float32Vector:
126
+ def __init__(self, *args) -> None: ...
127
+ def __len__(self) -> int: ...
128
+ def to_array(self) -> npt.NDArray[np.float32]: ...
129
+
130
+ class Complex128Vector:
131
+ def __init__(self, *args) -> None: ...
132
+ def __len__(self) -> int: ...
133
+ def to_array(self) -> npt.NDArray[np.complex128]: ...
134
+
135
+ class Complex64Vector:
136
+ def __init__(self, *args) -> None: ...
137
+ def __len__(self) -> int: ...
138
+ def to_array(self) -> npt.NDArray[np.complex64]: ...
139
+
140
+ class StringVector:
141
+ def __init__(self, *args) -> None: ...
142
+ def __len__(self) -> int: ...
143
+ def to_array(self) -> npt.NDArray[np.object_]: ...
144
+
145
+ class ObjectVector:
146
+ def __init__(self, *args) -> None: ...
147
+ def __len__(self) -> int: ...
148
+ def to_array(self) -> npt.NDArray[np.object_]: ...
149
+
150
+ class HashTable:
151
+ # NB: The base HashTable class does _not_ actually have these methods;
152
+ # we are putting them here for the sake of mypy to avoid
153
+ # reproducing them in each subclass below.
154
+ def __init__(self, size_hint: int = ..., uses_mask: bool = ...) -> None: ...
155
+ def __len__(self) -> int: ...
156
+ def __contains__(self, key: Hashable) -> bool: ...
157
+ def sizeof(self, deep: bool = ...) -> int: ...
158
+ def get_state(self) -> dict[str, int]: ...
159
+ # TODO: `val/key` type is subclass-specific
160
+ def get_item(self, val): ... # TODO: return type?
161
+ def set_item(self, key, val) -> None: ...
162
+ def get_na(self): ... # TODO: return type?
163
+ def set_na(self, val) -> None: ...
164
+ def map_locations(
165
+ self,
166
+ values: np.ndarray, # np.ndarray[subclass-specific]
167
+ mask: npt.NDArray[np.bool_] | None = ...,
168
+ ) -> None: ...
169
+ def lookup(
170
+ self,
171
+ values: np.ndarray, # np.ndarray[subclass-specific]
172
+ mask: npt.NDArray[np.bool_] | None = ...,
173
+ ) -> npt.NDArray[np.intp]: ...
174
+ def get_labels(
175
+ self,
176
+ values: np.ndarray, # np.ndarray[subclass-specific]
177
+ uniques, # SubclassTypeVector
178
+ count_prior: int = ...,
179
+ na_sentinel: int = ...,
180
+ na_value: object = ...,
181
+ mask=...,
182
+ ) -> npt.NDArray[np.intp]: ...
183
+ def unique(
184
+ self,
185
+ values: np.ndarray, # np.ndarray[subclass-specific]
186
+ return_inverse: bool = ...,
187
+ mask=...,
188
+ ) -> (
189
+ tuple[
190
+ np.ndarray, # np.ndarray[subclass-specific]
191
+ npt.NDArray[np.intp],
192
+ ]
193
+ | np.ndarray
194
+ ): ... # np.ndarray[subclass-specific]
195
+ def factorize(
196
+ self,
197
+ values: np.ndarray, # np.ndarray[subclass-specific]
198
+ na_sentinel: int = ...,
199
+ na_value: object = ...,
200
+ mask=...,
201
+ ignore_na: bool = True,
202
+ ) -> tuple[np.ndarray, npt.NDArray[np.intp]]: ... # np.ndarray[subclass-specific]
203
+
204
+ class Complex128HashTable(HashTable): ...
205
+ class Complex64HashTable(HashTable): ...
206
+ class Float64HashTable(HashTable): ...
207
+ class Float32HashTable(HashTable): ...
208
+
209
+ class Int64HashTable(HashTable):
210
+ # Only Int64HashTable has get_labels_groupby, map_keys_to_values
211
+ def get_labels_groupby(
212
+ self,
213
+ values: npt.NDArray[np.int64], # const int64_t[:]
214
+ ) -> tuple[npt.NDArray[np.intp], npt.NDArray[np.int64]]: ...
215
+ def map_keys_to_values(
216
+ self,
217
+ keys: npt.NDArray[np.int64],
218
+ values: npt.NDArray[np.int64], # const int64_t[:]
219
+ ) -> None: ...
220
+
221
+ class Int32HashTable(HashTable): ...
222
+ class Int16HashTable(HashTable): ...
223
+ class Int8HashTable(HashTable): ...
224
+ class UInt64HashTable(HashTable): ...
225
+ class UInt32HashTable(HashTable): ...
226
+ class UInt16HashTable(HashTable): ...
227
+ class UInt8HashTable(HashTable): ...
228
+ class StringHashTable(HashTable): ...
229
+ class PyObjectHashTable(HashTable): ...
230
+ class IntpHashTable(HashTable): ...
231
+
232
+ def duplicated(
233
+ values: np.ndarray,
234
+ keep: Literal["last", "first", False] = ...,
235
+ mask: npt.NDArray[np.bool_] | None = ...,
236
+ ) -> npt.NDArray[np.bool_]: ...
237
+ def mode(
238
+ values: np.ndarray, dropna: bool, mask: npt.NDArray[np.bool_] | None = ...
239
+ ) -> np.ndarray: ...
240
+ def value_count(
241
+ values: np.ndarray,
242
+ dropna: bool,
243
+ mask: npt.NDArray[np.bool_] | None = ...,
244
+ ) -> tuple[np.ndarray, npt.NDArray[np.int64], int]: ... # np.ndarray[same-as-values]
245
+
246
+ # arr and values should have same dtype
247
+ def ismember(
248
+ arr: np.ndarray,
249
+ values: np.ndarray,
250
+ ) -> npt.NDArray[np.bool_]: ...
251
+ def object_hash(obj) -> int: ...
252
+ def objects_are_equal(a, b) -> bool: ...
lib/python3.10/site-packages/pandas/_libs/index.pyi ADDED
@@ -0,0 +1,100 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import numpy as np
2
+
3
+ from pandas._typing import npt
4
+
5
+ from pandas import MultiIndex
6
+ from pandas.core.arrays import ExtensionArray
7
+
8
+ multiindex_nulls_shift: int
9
+
10
+ class IndexEngine:
11
+ over_size_threshold: bool
12
+ def __init__(self, values: np.ndarray) -> None: ...
13
+ def __contains__(self, val: object) -> bool: ...
14
+
15
+ # -> int | slice | np.ndarray[bool]
16
+ def get_loc(self, val: object) -> int | slice | np.ndarray: ...
17
+ def sizeof(self, deep: bool = ...) -> int: ...
18
+ def __sizeof__(self) -> int: ...
19
+ @property
20
+ def is_unique(self) -> bool: ...
21
+ @property
22
+ def is_monotonic_increasing(self) -> bool: ...
23
+ @property
24
+ def is_monotonic_decreasing(self) -> bool: ...
25
+ @property
26
+ def is_mapping_populated(self) -> bool: ...
27
+ def clear_mapping(self): ...
28
+ def get_indexer(self, values: np.ndarray) -> npt.NDArray[np.intp]: ...
29
+ def get_indexer_non_unique(
30
+ self,
31
+ targets: np.ndarray,
32
+ ) -> tuple[npt.NDArray[np.intp], npt.NDArray[np.intp]]: ...
33
+
34
+ class MaskedIndexEngine(IndexEngine):
35
+ def __init__(self, values: object) -> None: ...
36
+ def get_indexer_non_unique(
37
+ self, targets: object
38
+ ) -> tuple[npt.NDArray[np.intp], npt.NDArray[np.intp]]: ...
39
+
40
+ class Float64Engine(IndexEngine): ...
41
+ class Float32Engine(IndexEngine): ...
42
+ class Complex128Engine(IndexEngine): ...
43
+ class Complex64Engine(IndexEngine): ...
44
+ class Int64Engine(IndexEngine): ...
45
+ class Int32Engine(IndexEngine): ...
46
+ class Int16Engine(IndexEngine): ...
47
+ class Int8Engine(IndexEngine): ...
48
+ class UInt64Engine(IndexEngine): ...
49
+ class UInt32Engine(IndexEngine): ...
50
+ class UInt16Engine(IndexEngine): ...
51
+ class UInt8Engine(IndexEngine): ...
52
+ class ObjectEngine(IndexEngine): ...
53
+ class DatetimeEngine(Int64Engine): ...
54
+ class TimedeltaEngine(DatetimeEngine): ...
55
+ class PeriodEngine(Int64Engine): ...
56
+ class BoolEngine(UInt8Engine): ...
57
+ class MaskedFloat64Engine(MaskedIndexEngine): ...
58
+ class MaskedFloat32Engine(MaskedIndexEngine): ...
59
+ class MaskedComplex128Engine(MaskedIndexEngine): ...
60
+ class MaskedComplex64Engine(MaskedIndexEngine): ...
61
+ class MaskedInt64Engine(MaskedIndexEngine): ...
62
+ class MaskedInt32Engine(MaskedIndexEngine): ...
63
+ class MaskedInt16Engine(MaskedIndexEngine): ...
64
+ class MaskedInt8Engine(MaskedIndexEngine): ...
65
+ class MaskedUInt64Engine(MaskedIndexEngine): ...
66
+ class MaskedUInt32Engine(MaskedIndexEngine): ...
67
+ class MaskedUInt16Engine(MaskedIndexEngine): ...
68
+ class MaskedUInt8Engine(MaskedIndexEngine): ...
69
+ class MaskedBoolEngine(MaskedUInt8Engine): ...
70
+
71
+ class BaseMultiIndexCodesEngine:
72
+ levels: list[np.ndarray]
73
+ offsets: np.ndarray # ndarray[uint64_t, ndim=1]
74
+
75
+ def __init__(
76
+ self,
77
+ levels: list[np.ndarray], # all entries hashable
78
+ labels: list[np.ndarray], # all entries integer-dtyped
79
+ offsets: np.ndarray, # np.ndarray[np.uint64, ndim=1]
80
+ ) -> None: ...
81
+ def get_indexer(self, target: npt.NDArray[np.object_]) -> npt.NDArray[np.intp]: ...
82
+ def _extract_level_codes(self, target: MultiIndex) -> np.ndarray: ...
83
+
84
+ class ExtensionEngine:
85
+ def __init__(self, values: ExtensionArray) -> None: ...
86
+ def __contains__(self, val: object) -> bool: ...
87
+ def get_loc(self, val: object) -> int | slice | np.ndarray: ...
88
+ def get_indexer(self, values: np.ndarray) -> npt.NDArray[np.intp]: ...
89
+ def get_indexer_non_unique(
90
+ self,
91
+ targets: np.ndarray,
92
+ ) -> tuple[npt.NDArray[np.intp], npt.NDArray[np.intp]]: ...
93
+ @property
94
+ def is_unique(self) -> bool: ...
95
+ @property
96
+ def is_monotonic_increasing(self) -> bool: ...
97
+ @property
98
+ def is_monotonic_decreasing(self) -> bool: ...
99
+ def sizeof(self, deep: bool = ...) -> int: ...
100
+ def clear_mapping(self): ...
lib/python3.10/site-packages/pandas/_libs/indexing.cpython-310-x86_64-linux-gnu.so ADDED
Binary file (66.6 kB). View file
 
lib/python3.10/site-packages/pandas/_libs/indexing.pyi ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from typing import (
2
+ Generic,
3
+ TypeVar,
4
+ )
5
+
6
+ from pandas.core.indexing import IndexingMixin
7
+
8
+ _IndexingMixinT = TypeVar("_IndexingMixinT", bound=IndexingMixin)
9
+
10
+ class NDFrameIndexerBase(Generic[_IndexingMixinT]):
11
+ name: str
12
+ # in practice obj is either a DataFrame or a Series
13
+ obj: _IndexingMixinT
14
+
15
+ def __init__(self, name: str, obj: _IndexingMixinT) -> None: ...
16
+ @property
17
+ def ndim(self) -> int: ...
lib/python3.10/site-packages/pandas/_libs/internals.pyi ADDED
@@ -0,0 +1,94 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from typing import (
2
+ Iterator,
3
+ Sequence,
4
+ final,
5
+ overload,
6
+ )
7
+ import weakref
8
+
9
+ import numpy as np
10
+
11
+ from pandas._typing import (
12
+ ArrayLike,
13
+ Self,
14
+ npt,
15
+ )
16
+
17
+ from pandas import Index
18
+ from pandas.core.internals.blocks import Block as B
19
+
20
+ def slice_len(slc: slice, objlen: int = ...) -> int: ...
21
+ def get_concat_blkno_indexers(
22
+ blknos_list: list[npt.NDArray[np.intp]],
23
+ ) -> list[tuple[npt.NDArray[np.intp], BlockPlacement]]: ...
24
+ def get_blkno_indexers(
25
+ blknos: np.ndarray, # int64_t[:]
26
+ group: bool = ...,
27
+ ) -> list[tuple[int, slice | np.ndarray]]: ...
28
+ def get_blkno_placements(
29
+ blknos: np.ndarray,
30
+ group: bool = ...,
31
+ ) -> Iterator[tuple[int, BlockPlacement]]: ...
32
+ def update_blklocs_and_blknos(
33
+ blklocs: npt.NDArray[np.intp],
34
+ blknos: npt.NDArray[np.intp],
35
+ loc: int,
36
+ nblocks: int,
37
+ ) -> tuple[npt.NDArray[np.intp], npt.NDArray[np.intp]]: ...
38
+ @final
39
+ class BlockPlacement:
40
+ def __init__(self, val: int | slice | np.ndarray) -> None: ...
41
+ @property
42
+ def indexer(self) -> np.ndarray | slice: ...
43
+ @property
44
+ def as_array(self) -> np.ndarray: ...
45
+ @property
46
+ def as_slice(self) -> slice: ...
47
+ @property
48
+ def is_slice_like(self) -> bool: ...
49
+ @overload
50
+ def __getitem__(
51
+ self, loc: slice | Sequence[int] | npt.NDArray[np.intp]
52
+ ) -> BlockPlacement: ...
53
+ @overload
54
+ def __getitem__(self, loc: int) -> int: ...
55
+ def __iter__(self) -> Iterator[int]: ...
56
+ def __len__(self) -> int: ...
57
+ def delete(self, loc) -> BlockPlacement: ...
58
+ def add(self, other) -> BlockPlacement: ...
59
+ def append(self, others: list[BlockPlacement]) -> BlockPlacement: ...
60
+ def tile_for_unstack(self, factor: int) -> npt.NDArray[np.intp]: ...
61
+
62
+ class Block:
63
+ _mgr_locs: BlockPlacement
64
+ ndim: int
65
+ values: ArrayLike
66
+ refs: BlockValuesRefs
67
+ def __init__(
68
+ self,
69
+ values: ArrayLike,
70
+ placement: BlockPlacement,
71
+ ndim: int,
72
+ refs: BlockValuesRefs | None = ...,
73
+ ) -> None: ...
74
+ def slice_block_rows(self, slicer: slice) -> Self: ...
75
+
76
+ class BlockManager:
77
+ blocks: tuple[B, ...]
78
+ axes: list[Index]
79
+ _known_consolidated: bool
80
+ _is_consolidated: bool
81
+ _blknos: np.ndarray
82
+ _blklocs: np.ndarray
83
+ def __init__(
84
+ self, blocks: tuple[B, ...], axes: list[Index], verify_integrity=...
85
+ ) -> None: ...
86
+ def get_slice(self, slobj: slice, axis: int = ...) -> Self: ...
87
+ def _rebuild_blknos_and_blklocs(self) -> None: ...
88
+
89
+ class BlockValuesRefs:
90
+ referenced_blocks: list[weakref.ref]
91
+ def __init__(self, blk: Block | None = ...) -> None: ...
92
+ def add_reference(self, blk: Block) -> None: ...
93
+ def add_index_reference(self, index: Index) -> None: ...
94
+ def has_reference(self) -> bool: ...
lib/python3.10/site-packages/pandas/_libs/interval.pyi ADDED
@@ -0,0 +1,174 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from typing import (
2
+ Any,
3
+ Generic,
4
+ TypeVar,
5
+ overload,
6
+ )
7
+
8
+ import numpy as np
9
+ import numpy.typing as npt
10
+
11
+ from pandas._typing import (
12
+ IntervalClosedType,
13
+ Timedelta,
14
+ Timestamp,
15
+ )
16
+
17
+ VALID_CLOSED: frozenset[str]
18
+
19
+ _OrderableScalarT = TypeVar("_OrderableScalarT", int, float)
20
+ _OrderableTimesT = TypeVar("_OrderableTimesT", Timestamp, Timedelta)
21
+ _OrderableT = TypeVar("_OrderableT", int, float, Timestamp, Timedelta)
22
+
23
+ class _LengthDescriptor:
24
+ @overload
25
+ def __get__(
26
+ self, instance: Interval[_OrderableScalarT], owner: Any
27
+ ) -> _OrderableScalarT: ...
28
+ @overload
29
+ def __get__(
30
+ self, instance: Interval[_OrderableTimesT], owner: Any
31
+ ) -> Timedelta: ...
32
+
33
+ class _MidDescriptor:
34
+ @overload
35
+ def __get__(self, instance: Interval[_OrderableScalarT], owner: Any) -> float: ...
36
+ @overload
37
+ def __get__(
38
+ self, instance: Interval[_OrderableTimesT], owner: Any
39
+ ) -> _OrderableTimesT: ...
40
+
41
+ class IntervalMixin:
42
+ @property
43
+ def closed_left(self) -> bool: ...
44
+ @property
45
+ def closed_right(self) -> bool: ...
46
+ @property
47
+ def open_left(self) -> bool: ...
48
+ @property
49
+ def open_right(self) -> bool: ...
50
+ @property
51
+ def is_empty(self) -> bool: ...
52
+ def _check_closed_matches(self, other: IntervalMixin, name: str = ...) -> None: ...
53
+
54
+ class Interval(IntervalMixin, Generic[_OrderableT]):
55
+ @property
56
+ def left(self: Interval[_OrderableT]) -> _OrderableT: ...
57
+ @property
58
+ def right(self: Interval[_OrderableT]) -> _OrderableT: ...
59
+ @property
60
+ def closed(self) -> IntervalClosedType: ...
61
+ mid: _MidDescriptor
62
+ length: _LengthDescriptor
63
+ def __init__(
64
+ self,
65
+ left: _OrderableT,
66
+ right: _OrderableT,
67
+ closed: IntervalClosedType = ...,
68
+ ) -> None: ...
69
+ def __hash__(self) -> int: ...
70
+ @overload
71
+ def __contains__(
72
+ self: Interval[Timedelta], key: Timedelta | Interval[Timedelta]
73
+ ) -> bool: ...
74
+ @overload
75
+ def __contains__(
76
+ self: Interval[Timestamp], key: Timestamp | Interval[Timestamp]
77
+ ) -> bool: ...
78
+ @overload
79
+ def __contains__(
80
+ self: Interval[_OrderableScalarT],
81
+ key: _OrderableScalarT | Interval[_OrderableScalarT],
82
+ ) -> bool: ...
83
+ @overload
84
+ def __add__(
85
+ self: Interval[_OrderableTimesT], y: Timedelta
86
+ ) -> Interval[_OrderableTimesT]: ...
87
+ @overload
88
+ def __add__(
89
+ self: Interval[int], y: _OrderableScalarT
90
+ ) -> Interval[_OrderableScalarT]: ...
91
+ @overload
92
+ def __add__(self: Interval[float], y: float) -> Interval[float]: ...
93
+ @overload
94
+ def __radd__(
95
+ self: Interval[_OrderableTimesT], y: Timedelta
96
+ ) -> Interval[_OrderableTimesT]: ...
97
+ @overload
98
+ def __radd__(
99
+ self: Interval[int], y: _OrderableScalarT
100
+ ) -> Interval[_OrderableScalarT]: ...
101
+ @overload
102
+ def __radd__(self: Interval[float], y: float) -> Interval[float]: ...
103
+ @overload
104
+ def __sub__(
105
+ self: Interval[_OrderableTimesT], y: Timedelta
106
+ ) -> Interval[_OrderableTimesT]: ...
107
+ @overload
108
+ def __sub__(
109
+ self: Interval[int], y: _OrderableScalarT
110
+ ) -> Interval[_OrderableScalarT]: ...
111
+ @overload
112
+ def __sub__(self: Interval[float], y: float) -> Interval[float]: ...
113
+ @overload
114
+ def __rsub__(
115
+ self: Interval[_OrderableTimesT], y: Timedelta
116
+ ) -> Interval[_OrderableTimesT]: ...
117
+ @overload
118
+ def __rsub__(
119
+ self: Interval[int], y: _OrderableScalarT
120
+ ) -> Interval[_OrderableScalarT]: ...
121
+ @overload
122
+ def __rsub__(self: Interval[float], y: float) -> Interval[float]: ...
123
+ @overload
124
+ def __mul__(
125
+ self: Interval[int], y: _OrderableScalarT
126
+ ) -> Interval[_OrderableScalarT]: ...
127
+ @overload
128
+ def __mul__(self: Interval[float], y: float) -> Interval[float]: ...
129
+ @overload
130
+ def __rmul__(
131
+ self: Interval[int], y: _OrderableScalarT
132
+ ) -> Interval[_OrderableScalarT]: ...
133
+ @overload
134
+ def __rmul__(self: Interval[float], y: float) -> Interval[float]: ...
135
+ @overload
136
+ def __truediv__(
137
+ self: Interval[int], y: _OrderableScalarT
138
+ ) -> Interval[_OrderableScalarT]: ...
139
+ @overload
140
+ def __truediv__(self: Interval[float], y: float) -> Interval[float]: ...
141
+ @overload
142
+ def __floordiv__(
143
+ self: Interval[int], y: _OrderableScalarT
144
+ ) -> Interval[_OrderableScalarT]: ...
145
+ @overload
146
+ def __floordiv__(self: Interval[float], y: float) -> Interval[float]: ...
147
+ def overlaps(self: Interval[_OrderableT], other: Interval[_OrderableT]) -> bool: ...
148
+
149
+ def intervals_to_interval_bounds(
150
+ intervals: np.ndarray, validate_closed: bool = ...
151
+ ) -> tuple[np.ndarray, np.ndarray, IntervalClosedType]: ...
152
+
153
+ class IntervalTree(IntervalMixin):
154
+ def __init__(
155
+ self,
156
+ left: np.ndarray,
157
+ right: np.ndarray,
158
+ closed: IntervalClosedType = ...,
159
+ leaf_size: int = ...,
160
+ ) -> None: ...
161
+ @property
162
+ def mid(self) -> np.ndarray: ...
163
+ @property
164
+ def length(self) -> np.ndarray: ...
165
+ def get_indexer(self, target) -> npt.NDArray[np.intp]: ...
166
+ def get_indexer_non_unique(
167
+ self, target
168
+ ) -> tuple[npt.NDArray[np.intp], npt.NDArray[np.intp]]: ...
169
+ _na_count: int
170
+ @property
171
+ def is_overlapping(self) -> bool: ...
172
+ @property
173
+ def is_monotonic_increasing(self) -> bool: ...
174
+ def clear_mapping(self) -> None: ...
lib/python3.10/site-packages/pandas/_libs/join.pyi ADDED
@@ -0,0 +1,79 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import numpy as np
2
+
3
+ from pandas._typing import npt
4
+
5
+ def inner_join(
6
+ left: np.ndarray, # const intp_t[:]
7
+ right: np.ndarray, # const intp_t[:]
8
+ max_groups: int,
9
+ sort: bool = ...,
10
+ ) -> tuple[npt.NDArray[np.intp], npt.NDArray[np.intp]]: ...
11
+ def left_outer_join(
12
+ left: np.ndarray, # const intp_t[:]
13
+ right: np.ndarray, # const intp_t[:]
14
+ max_groups: int,
15
+ sort: bool = ...,
16
+ ) -> tuple[npt.NDArray[np.intp], npt.NDArray[np.intp]]: ...
17
+ def full_outer_join(
18
+ left: np.ndarray, # const intp_t[:]
19
+ right: np.ndarray, # const intp_t[:]
20
+ max_groups: int,
21
+ ) -> tuple[npt.NDArray[np.intp], npt.NDArray[np.intp]]: ...
22
+ def ffill_indexer(
23
+ indexer: np.ndarray, # const intp_t[:]
24
+ ) -> npt.NDArray[np.intp]: ...
25
+ def left_join_indexer_unique(
26
+ left: np.ndarray, # ndarray[join_t]
27
+ right: np.ndarray, # ndarray[join_t]
28
+ ) -> npt.NDArray[np.intp]: ...
29
+ def left_join_indexer(
30
+ left: np.ndarray, # ndarray[join_t]
31
+ right: np.ndarray, # ndarray[join_t]
32
+ ) -> tuple[
33
+ np.ndarray, # np.ndarray[join_t]
34
+ npt.NDArray[np.intp],
35
+ npt.NDArray[np.intp],
36
+ ]: ...
37
+ def inner_join_indexer(
38
+ left: np.ndarray, # ndarray[join_t]
39
+ right: np.ndarray, # ndarray[join_t]
40
+ ) -> tuple[
41
+ np.ndarray, # np.ndarray[join_t]
42
+ npt.NDArray[np.intp],
43
+ npt.NDArray[np.intp],
44
+ ]: ...
45
+ def outer_join_indexer(
46
+ left: np.ndarray, # ndarray[join_t]
47
+ right: np.ndarray, # ndarray[join_t]
48
+ ) -> tuple[
49
+ np.ndarray, # np.ndarray[join_t]
50
+ npt.NDArray[np.intp],
51
+ npt.NDArray[np.intp],
52
+ ]: ...
53
+ def asof_join_backward_on_X_by_Y(
54
+ left_values: np.ndarray, # ndarray[numeric_t]
55
+ right_values: np.ndarray, # ndarray[numeric_t]
56
+ left_by_values: np.ndarray, # const int64_t[:]
57
+ right_by_values: np.ndarray, # const int64_t[:]
58
+ allow_exact_matches: bool = ...,
59
+ tolerance: np.number | float | None = ...,
60
+ use_hashtable: bool = ...,
61
+ ) -> tuple[npt.NDArray[np.intp], npt.NDArray[np.intp]]: ...
62
+ def asof_join_forward_on_X_by_Y(
63
+ left_values: np.ndarray, # ndarray[numeric_t]
64
+ right_values: np.ndarray, # ndarray[numeric_t]
65
+ left_by_values: np.ndarray, # const int64_t[:]
66
+ right_by_values: np.ndarray, # const int64_t[:]
67
+ allow_exact_matches: bool = ...,
68
+ tolerance: np.number | float | None = ...,
69
+ use_hashtable: bool = ...,
70
+ ) -> tuple[npt.NDArray[np.intp], npt.NDArray[np.intp]]: ...
71
+ def asof_join_nearest_on_X_by_Y(
72
+ left_values: np.ndarray, # ndarray[numeric_t]
73
+ right_values: np.ndarray, # ndarray[numeric_t]
74
+ left_by_values: np.ndarray, # const int64_t[:]
75
+ right_by_values: np.ndarray, # const int64_t[:]
76
+ allow_exact_matches: bool = ...,
77
+ tolerance: np.number | float | None = ...,
78
+ use_hashtable: bool = ...,
79
+ ) -> tuple[npt.NDArray[np.intp], npt.NDArray[np.intp]]: ...
lib/python3.10/site-packages/pandas/_libs/json.cpython-310-x86_64-linux-gnu.so ADDED
Binary file (64.3 kB). View file
 
lib/python3.10/site-packages/pandas/_libs/json.pyi ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from typing import (
2
+ Any,
3
+ Callable,
4
+ )
5
+
6
+ def ujson_dumps(
7
+ obj: Any,
8
+ ensure_ascii: bool = ...,
9
+ double_precision: int = ...,
10
+ indent: int = ...,
11
+ orient: str = ...,
12
+ date_unit: str = ...,
13
+ iso_dates: bool = ...,
14
+ default_handler: None
15
+ | Callable[[Any], str | float | bool | list | dict | None] = ...,
16
+ ) -> str: ...
17
+ def ujson_loads(
18
+ s: str,
19
+ precise_float: bool = ...,
20
+ numpy: bool = ...,
21
+ dtype: None = ...,
22
+ labelled: bool = ...,
23
+ ) -> Any: ...
lib/python3.10/site-packages/pandas/_libs/lib.pyi ADDED
@@ -0,0 +1,213 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # TODO(npdtypes): Many types specified here can be made more specific/accurate;
2
+ # the more specific versions are specified in comments
3
+ from decimal import Decimal
4
+ from typing import (
5
+ Any,
6
+ Callable,
7
+ Final,
8
+ Generator,
9
+ Hashable,
10
+ Literal,
11
+ TypeAlias,
12
+ overload,
13
+ )
14
+
15
+ import numpy as np
16
+
17
+ from pandas._libs.interval import Interval
18
+ from pandas._libs.tslibs import Period
19
+ from pandas._typing import (
20
+ ArrayLike,
21
+ DtypeObj,
22
+ TypeGuard,
23
+ npt,
24
+ )
25
+
26
+ # placeholder until we can specify np.ndarray[object, ndim=2]
27
+ ndarray_obj_2d = np.ndarray
28
+
29
+ from enum import Enum
30
+
31
+ class _NoDefault(Enum):
32
+ no_default = ...
33
+
34
+ no_default: Final = _NoDefault.no_default
35
+ NoDefault: TypeAlias = Literal[_NoDefault.no_default]
36
+
37
+ i8max: int
38
+ u8max: int
39
+
40
+ def is_np_dtype(dtype: object, kinds: str | None = ...) -> TypeGuard[np.dtype]: ...
41
+ def item_from_zerodim(val: object) -> object: ...
42
+ def infer_dtype(value: object, skipna: bool = ...) -> str: ...
43
+ def is_iterator(obj: object) -> bool: ...
44
+ def is_scalar(val: object) -> bool: ...
45
+ def is_list_like(obj: object, allow_sets: bool = ...) -> bool: ...
46
+ def is_pyarrow_array(obj: object) -> bool: ...
47
+ def is_period(val: object) -> TypeGuard[Period]: ...
48
+ def is_interval(obj: object) -> TypeGuard[Interval]: ...
49
+ def is_decimal(obj: object) -> TypeGuard[Decimal]: ...
50
+ def is_complex(obj: object) -> TypeGuard[complex]: ...
51
+ def is_bool(obj: object) -> TypeGuard[bool | np.bool_]: ...
52
+ def is_integer(obj: object) -> TypeGuard[int | np.integer]: ...
53
+ def is_int_or_none(obj) -> bool: ...
54
+ def is_float(obj: object) -> TypeGuard[float]: ...
55
+ def is_interval_array(values: np.ndarray) -> bool: ...
56
+ def is_datetime64_array(values: np.ndarray, skipna: bool = True) -> bool: ...
57
+ def is_timedelta_or_timedelta64_array(
58
+ values: np.ndarray, skipna: bool = True
59
+ ) -> bool: ...
60
+ def is_datetime_with_singletz_array(values: np.ndarray) -> bool: ...
61
+ def is_time_array(values: np.ndarray, skipna: bool = ...): ...
62
+ def is_date_array(values: np.ndarray, skipna: bool = ...): ...
63
+ def is_datetime_array(values: np.ndarray, skipna: bool = ...): ...
64
+ def is_string_array(values: np.ndarray, skipna: bool = ...): ...
65
+ def is_float_array(values: np.ndarray): ...
66
+ def is_integer_array(values: np.ndarray, skipna: bool = ...): ...
67
+ def is_bool_array(values: np.ndarray, skipna: bool = ...): ...
68
+ def fast_multiget(
69
+ mapping: dict,
70
+ keys: np.ndarray, # object[:]
71
+ default=...,
72
+ ) -> np.ndarray: ...
73
+ def fast_unique_multiple_list_gen(gen: Generator, sort: bool = ...) -> list: ...
74
+ def fast_unique_multiple_list(lists: list, sort: bool | None = ...) -> list: ...
75
+ def map_infer(
76
+ arr: np.ndarray,
77
+ f: Callable[[Any], Any],
78
+ convert: bool = ...,
79
+ ignore_na: bool = ...,
80
+ ) -> np.ndarray: ...
81
+ @overload
82
+ def maybe_convert_objects(
83
+ objects: npt.NDArray[np.object_],
84
+ *,
85
+ try_float: bool = ...,
86
+ safe: bool = ...,
87
+ convert_numeric: bool = ...,
88
+ convert_non_numeric: Literal[False] = ...,
89
+ convert_to_nullable_dtype: Literal[False] = ...,
90
+ dtype_if_all_nat: DtypeObj | None = ...,
91
+ ) -> npt.NDArray[np.object_ | np.number]: ...
92
+ @overload
93
+ def maybe_convert_objects(
94
+ objects: npt.NDArray[np.object_],
95
+ *,
96
+ try_float: bool = ...,
97
+ safe: bool = ...,
98
+ convert_numeric: bool = ...,
99
+ convert_non_numeric: bool = ...,
100
+ convert_to_nullable_dtype: Literal[True] = ...,
101
+ dtype_if_all_nat: DtypeObj | None = ...,
102
+ ) -> ArrayLike: ...
103
+ @overload
104
+ def maybe_convert_objects(
105
+ objects: npt.NDArray[np.object_],
106
+ *,
107
+ try_float: bool = ...,
108
+ safe: bool = ...,
109
+ convert_numeric: bool = ...,
110
+ convert_non_numeric: bool = ...,
111
+ convert_to_nullable_dtype: bool = ...,
112
+ dtype_if_all_nat: DtypeObj | None = ...,
113
+ ) -> ArrayLike: ...
114
+ @overload
115
+ def maybe_convert_numeric(
116
+ values: npt.NDArray[np.object_],
117
+ na_values: set,
118
+ convert_empty: bool = ...,
119
+ coerce_numeric: bool = ...,
120
+ convert_to_masked_nullable: Literal[False] = ...,
121
+ ) -> tuple[np.ndarray, None]: ...
122
+ @overload
123
+ def maybe_convert_numeric(
124
+ values: npt.NDArray[np.object_],
125
+ na_values: set,
126
+ convert_empty: bool = ...,
127
+ coerce_numeric: bool = ...,
128
+ *,
129
+ convert_to_masked_nullable: Literal[True],
130
+ ) -> tuple[np.ndarray, np.ndarray]: ...
131
+
132
+ # TODO: restrict `arr`?
133
+ def ensure_string_array(
134
+ arr,
135
+ na_value: object = ...,
136
+ convert_na_value: bool = ...,
137
+ copy: bool = ...,
138
+ skipna: bool = ...,
139
+ ) -> npt.NDArray[np.object_]: ...
140
+ def convert_nans_to_NA(
141
+ arr: npt.NDArray[np.object_],
142
+ ) -> npt.NDArray[np.object_]: ...
143
+ def fast_zip(ndarrays: list) -> npt.NDArray[np.object_]: ...
144
+
145
+ # TODO: can we be more specific about rows?
146
+ def to_object_array_tuples(rows: object) -> ndarray_obj_2d: ...
147
+ def tuples_to_object_array(
148
+ tuples: npt.NDArray[np.object_],
149
+ ) -> ndarray_obj_2d: ...
150
+
151
+ # TODO: can we be more specific about rows?
152
+ def to_object_array(rows: object, min_width: int = ...) -> ndarray_obj_2d: ...
153
+ def dicts_to_array(dicts: list, columns: list) -> ndarray_obj_2d: ...
154
+ def maybe_booleans_to_slice(
155
+ mask: npt.NDArray[np.uint8],
156
+ ) -> slice | npt.NDArray[np.uint8]: ...
157
+ def maybe_indices_to_slice(
158
+ indices: npt.NDArray[np.intp],
159
+ max_len: int,
160
+ ) -> slice | npt.NDArray[np.intp]: ...
161
+ def is_all_arraylike(obj: list) -> bool: ...
162
+
163
+ # -----------------------------------------------------------------
164
+ # Functions which in reality take memoryviews
165
+
166
+ def memory_usage_of_objects(arr: np.ndarray) -> int: ... # object[:] # np.int64
167
+ def map_infer_mask(
168
+ arr: np.ndarray,
169
+ f: Callable[[Any], Any],
170
+ mask: np.ndarray, # const uint8_t[:]
171
+ convert: bool = ...,
172
+ na_value: Any = ...,
173
+ dtype: np.dtype = ...,
174
+ ) -> np.ndarray: ...
175
+ def indices_fast(
176
+ index: npt.NDArray[np.intp],
177
+ labels: np.ndarray, # const int64_t[:]
178
+ keys: list,
179
+ sorted_labels: list[npt.NDArray[np.int64]],
180
+ ) -> dict[Hashable, npt.NDArray[np.intp]]: ...
181
+ def generate_slices(
182
+ labels: np.ndarray, ngroups: int # const intp_t[:]
183
+ ) -> tuple[npt.NDArray[np.int64], npt.NDArray[np.int64]]: ...
184
+ def count_level_2d(
185
+ mask: np.ndarray, # ndarray[uint8_t, ndim=2, cast=True],
186
+ labels: np.ndarray, # const intp_t[:]
187
+ max_bin: int,
188
+ ) -> np.ndarray: ... # np.ndarray[np.int64, ndim=2]
189
+ def get_level_sorter(
190
+ codes: np.ndarray, # const int64_t[:]
191
+ starts: np.ndarray, # const intp_t[:]
192
+ ) -> np.ndarray: ... # np.ndarray[np.intp, ndim=1]
193
+ def generate_bins_dt64(
194
+ values: npt.NDArray[np.int64],
195
+ binner: np.ndarray, # const int64_t[:]
196
+ closed: object = ...,
197
+ hasnans: bool = ...,
198
+ ) -> np.ndarray: ... # np.ndarray[np.int64, ndim=1]
199
+ def array_equivalent_object(
200
+ left: npt.NDArray[np.object_],
201
+ right: npt.NDArray[np.object_],
202
+ ) -> bool: ...
203
+ def has_infs(arr: np.ndarray) -> bool: ... # const floating[:]
204
+ def has_only_ints_or_nan(arr: np.ndarray) -> bool: ... # const floating[:]
205
+ def get_reverse_indexer(
206
+ indexer: np.ndarray, # const intp_t[:]
207
+ length: int,
208
+ ) -> npt.NDArray[np.intp]: ...
209
+ def is_bool_list(obj: list) -> bool: ...
210
+ def dtypes_all_equal(types: list[DtypeObj]) -> bool: ...
211
+ def is_range_indexer(
212
+ left: np.ndarray, n: int # np.ndarray[np.int64, ndim=1]
213
+ ) -> bool: ...
lib/python3.10/site-packages/pandas/_libs/ops.pyi ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from typing import (
2
+ Any,
3
+ Callable,
4
+ Iterable,
5
+ Literal,
6
+ TypeAlias,
7
+ overload,
8
+ )
9
+
10
+ import numpy as np
11
+
12
+ from pandas._typing import npt
13
+
14
+ _BinOp: TypeAlias = Callable[[Any, Any], Any]
15
+ _BoolOp: TypeAlias = Callable[[Any, Any], bool]
16
+
17
+ def scalar_compare(
18
+ values: np.ndarray, # object[:]
19
+ val: object,
20
+ op: _BoolOp, # {operator.eq, operator.ne, ...}
21
+ ) -> npt.NDArray[np.bool_]: ...
22
+ def vec_compare(
23
+ left: npt.NDArray[np.object_],
24
+ right: npt.NDArray[np.object_],
25
+ op: _BoolOp, # {operator.eq, operator.ne, ...}
26
+ ) -> npt.NDArray[np.bool_]: ...
27
+ def scalar_binop(
28
+ values: np.ndarray, # object[:]
29
+ val: object,
30
+ op: _BinOp, # binary operator
31
+ ) -> np.ndarray: ...
32
+ def vec_binop(
33
+ left: np.ndarray, # object[:]
34
+ right: np.ndarray, # object[:]
35
+ op: _BinOp, # binary operator
36
+ ) -> np.ndarray: ...
37
+ @overload
38
+ def maybe_convert_bool(
39
+ arr: npt.NDArray[np.object_],
40
+ true_values: Iterable | None = None,
41
+ false_values: Iterable | None = None,
42
+ convert_to_masked_nullable: Literal[False] = ...,
43
+ ) -> tuple[np.ndarray, None]: ...
44
+ @overload
45
+ def maybe_convert_bool(
46
+ arr: npt.NDArray[np.object_],
47
+ true_values: Iterable = ...,
48
+ false_values: Iterable = ...,
49
+ *,
50
+ convert_to_masked_nullable: Literal[True],
51
+ ) -> tuple[np.ndarray, np.ndarray]: ...
lib/python3.10/site-packages/pandas/_libs/ops_dispatch.cpython-310-x86_64-linux-gnu.so ADDED
Binary file (61.7 kB). View file
 
lib/python3.10/site-packages/pandas/_libs/ops_dispatch.pyi ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ import numpy as np
2
+
3
+ def maybe_dispatch_ufunc_to_dunder_op(
4
+ self, ufunc: np.ufunc, method: str, *inputs, **kwargs
5
+ ): ...
lib/python3.10/site-packages/pandas/_libs/pandas_datetime.cpython-310-x86_64-linux-gnu.so ADDED
Binary file (39.3 kB). View file
 
lib/python3.10/site-packages/pandas/_libs/pandas_parser.cpython-310-x86_64-linux-gnu.so ADDED
Binary file (43.4 kB). View file
 
lib/python3.10/site-packages/pandas/_libs/parsers.pyi ADDED
@@ -0,0 +1,77 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from typing import (
2
+ Hashable,
3
+ Literal,
4
+ )
5
+
6
+ import numpy as np
7
+
8
+ from pandas._typing import (
9
+ ArrayLike,
10
+ Dtype,
11
+ npt,
12
+ )
13
+
14
+ STR_NA_VALUES: set[str]
15
+ DEFAULT_BUFFER_HEURISTIC: int
16
+
17
+ def sanitize_objects(
18
+ values: npt.NDArray[np.object_],
19
+ na_values: set,
20
+ ) -> int: ...
21
+
22
+ class TextReader:
23
+ unnamed_cols: set[str]
24
+ table_width: int # int64_t
25
+ leading_cols: int # int64_t
26
+ header: list[list[int]] # non-negative integers
27
+ def __init__(
28
+ self,
29
+ source,
30
+ delimiter: bytes | str = ..., # single-character only
31
+ header=...,
32
+ header_start: int = ..., # int64_t
33
+ header_end: int = ..., # uint64_t
34
+ index_col=...,
35
+ names=...,
36
+ tokenize_chunksize: int = ..., # int64_t
37
+ delim_whitespace: bool = ...,
38
+ converters=...,
39
+ skipinitialspace: bool = ...,
40
+ escapechar: bytes | str | None = ..., # single-character only
41
+ doublequote: bool = ...,
42
+ quotechar: str | bytes | None = ..., # at most 1 character
43
+ quoting: int = ...,
44
+ lineterminator: bytes | str | None = ..., # at most 1 character
45
+ comment=...,
46
+ decimal: bytes | str = ..., # single-character only
47
+ thousands: bytes | str | None = ..., # single-character only
48
+ dtype: Dtype | dict[Hashable, Dtype] = ...,
49
+ usecols=...,
50
+ error_bad_lines: bool = ...,
51
+ warn_bad_lines: bool = ...,
52
+ na_filter: bool = ...,
53
+ na_values=...,
54
+ na_fvalues=...,
55
+ keep_default_na: bool = ...,
56
+ true_values=...,
57
+ false_values=...,
58
+ allow_leading_cols: bool = ...,
59
+ skiprows=...,
60
+ skipfooter: int = ..., # int64_t
61
+ verbose: bool = ...,
62
+ float_precision: Literal["round_trip", "legacy", "high"] | None = ...,
63
+ skip_blank_lines: bool = ...,
64
+ encoding_errors: bytes | str = ...,
65
+ ) -> None: ...
66
+ def set_noconvert(self, i: int) -> None: ...
67
+ def remove_noconvert(self, i: int) -> None: ...
68
+ def close(self) -> None: ...
69
+ def read(self, rows: int | None = ...) -> dict[int, ArrayLike]: ...
70
+ def read_low_memory(self, rows: int | None) -> list[dict[int, ArrayLike]]: ...
71
+
72
+ # _maybe_upcast, na_values are only exposed for testing
73
+ na_values: dict
74
+
75
+ def _maybe_upcast(
76
+ arr, use_dtype_backend: bool = ..., dtype_backend: str = ...
77
+ ) -> np.ndarray: ...
lib/python3.10/site-packages/pandas/_libs/properties.cpython-310-x86_64-linux-gnu.so ADDED
Binary file (91.9 kB). View file
 
lib/python3.10/site-packages/pandas/_libs/properties.pyi ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from typing import (
2
+ Sequence,
3
+ overload,
4
+ )
5
+
6
+ from pandas._typing import (
7
+ AnyArrayLike,
8
+ DataFrame,
9
+ Index,
10
+ Series,
11
+ )
12
+
13
+ # note: this is a lie to make type checkers happy (they special
14
+ # case property). cache_readonly uses attribute names similar to
15
+ # property (fget) but it does not provide fset and fdel.
16
+ cache_readonly = property
17
+
18
+ class AxisProperty:
19
+ axis: int
20
+ def __init__(self, axis: int = ..., doc: str = ...) -> None: ...
21
+ @overload
22
+ def __get__(self, obj: DataFrame | Series, type) -> Index: ...
23
+ @overload
24
+ def __get__(self, obj: None, type) -> AxisProperty: ...
25
+ def __set__(
26
+ self, obj: DataFrame | Series, value: AnyArrayLike | Sequence
27
+ ) -> None: ...
lib/python3.10/site-packages/pandas/_libs/reshape.pyi ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import numpy as np
2
+
3
+ from pandas._typing import npt
4
+
5
+ def unstack(
6
+ values: np.ndarray, # reshape_t[:, :]
7
+ mask: np.ndarray, # const uint8_t[:]
8
+ stride: int,
9
+ length: int,
10
+ width: int,
11
+ new_values: np.ndarray, # reshape_t[:, :]
12
+ new_mask: np.ndarray, # uint8_t[:, :]
13
+ ) -> None: ...
14
+ def explode(
15
+ values: npt.NDArray[np.object_],
16
+ ) -> tuple[npt.NDArray[np.object_], npt.NDArray[np.int64]]: ...
lib/python3.10/site-packages/pandas/_libs/sas.pyi ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ from pandas.io.sas.sas7bdat import SAS7BDATReader
2
+
3
+ class Parser:
4
+ def __init__(self, parser: SAS7BDATReader) -> None: ...
5
+ def read(self, nrows: int) -> None: ...
6
+
7
+ def get_subheader_index(signature: bytes) -> int: ...
lib/python3.10/site-packages/pandas/_libs/sparse.pyi ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from typing import Sequence
2
+
3
+ import numpy as np
4
+
5
+ from pandas._typing import (
6
+ Self,
7
+ npt,
8
+ )
9
+
10
+ class SparseIndex:
11
+ length: int
12
+ npoints: int
13
+ def __init__(self) -> None: ...
14
+ @property
15
+ def ngaps(self) -> int: ...
16
+ @property
17
+ def nbytes(self) -> int: ...
18
+ @property
19
+ def indices(self) -> npt.NDArray[np.int32]: ...
20
+ def equals(self, other) -> bool: ...
21
+ def lookup(self, index: int) -> np.int32: ...
22
+ def lookup_array(self, indexer: npt.NDArray[np.int32]) -> npt.NDArray[np.int32]: ...
23
+ def to_int_index(self) -> IntIndex: ...
24
+ def to_block_index(self) -> BlockIndex: ...
25
+ def intersect(self, y_: SparseIndex) -> Self: ...
26
+ def make_union(self, y_: SparseIndex) -> Self: ...
27
+
28
+ class IntIndex(SparseIndex):
29
+ indices: npt.NDArray[np.int32]
30
+ def __init__(
31
+ self, length: int, indices: Sequence[int], check_integrity: bool = ...
32
+ ) -> None: ...
33
+
34
+ class BlockIndex(SparseIndex):
35
+ nblocks: int
36
+ blocs: np.ndarray
37
+ blengths: np.ndarray
38
+ def __init__(
39
+ self, length: int, blocs: np.ndarray, blengths: np.ndarray
40
+ ) -> None: ...
41
+
42
+ # Override to have correct parameters
43
+ def intersect(self, other: SparseIndex) -> Self: ...
44
+ def make_union(self, y: SparseIndex) -> Self: ...
45
+
46
+ def make_mask_object_ndarray(
47
+ arr: npt.NDArray[np.object_], fill_value
48
+ ) -> npt.NDArray[np.bool_]: ...
49
+ def get_blocks(
50
+ indices: npt.NDArray[np.int32],
51
+ ) -> tuple[npt.NDArray[np.int32], npt.NDArray[np.int32]]: ...
lib/python3.10/site-packages/pandas/_libs/testing.pyi ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ def assert_dict_equal(a, b, compare_keys: bool = ...): ...
2
+ def assert_almost_equal(
3
+ a,
4
+ b,
5
+ rtol: float = ...,
6
+ atol: float = ...,
7
+ check_dtype: bool = ...,
8
+ obj=...,
9
+ lobj=...,
10
+ robj=...,
11
+ index_values=...,
12
+ ): ...
lib/python3.10/site-packages/pandas/_libs/tslib.pyi ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from datetime import tzinfo
2
+
3
+ import numpy as np
4
+
5
+ from pandas._typing import npt
6
+
7
+ def format_array_from_datetime(
8
+ values: npt.NDArray[np.int64],
9
+ tz: tzinfo | None = ...,
10
+ format: str | None = ...,
11
+ na_rep: str | float = ...,
12
+ reso: int = ..., # NPY_DATETIMEUNIT
13
+ ) -> npt.NDArray[np.object_]: ...
14
+ def array_with_unit_to_datetime(
15
+ values: npt.NDArray[np.object_],
16
+ unit: str,
17
+ errors: str = ...,
18
+ ) -> tuple[np.ndarray, tzinfo | None]: ...
19
+ def first_non_null(values: np.ndarray) -> int: ...
20
+ def array_to_datetime(
21
+ values: npt.NDArray[np.object_],
22
+ errors: str = ...,
23
+ dayfirst: bool = ...,
24
+ yearfirst: bool = ...,
25
+ utc: bool = ...,
26
+ creso: int = ...,
27
+ ) -> tuple[np.ndarray, tzinfo | None]: ...
28
+
29
+ # returned ndarray may be object dtype or datetime64[ns]
30
+
31
+ def array_to_datetime_with_tz(
32
+ values: npt.NDArray[np.object_],
33
+ tz: tzinfo,
34
+ dayfirst: bool,
35
+ yearfirst: bool,
36
+ creso: int,
37
+ ) -> npt.NDArray[np.int64]: ...
lib/python3.10/site-packages/pandas/_libs/tslibs/__init__.py ADDED
@@ -0,0 +1,87 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ __all__ = [
2
+ "dtypes",
3
+ "localize_pydatetime",
4
+ "NaT",
5
+ "NaTType",
6
+ "iNaT",
7
+ "nat_strings",
8
+ "OutOfBoundsDatetime",
9
+ "OutOfBoundsTimedelta",
10
+ "IncompatibleFrequency",
11
+ "Period",
12
+ "Resolution",
13
+ "Timedelta",
14
+ "normalize_i8_timestamps",
15
+ "is_date_array_normalized",
16
+ "dt64arr_to_periodarr",
17
+ "delta_to_nanoseconds",
18
+ "ints_to_pydatetime",
19
+ "ints_to_pytimedelta",
20
+ "get_resolution",
21
+ "Timestamp",
22
+ "tz_convert_from_utc_single",
23
+ "tz_convert_from_utc",
24
+ "to_offset",
25
+ "Tick",
26
+ "BaseOffset",
27
+ "tz_compare",
28
+ "is_unitless",
29
+ "astype_overflowsafe",
30
+ "get_unit_from_dtype",
31
+ "periods_per_day",
32
+ "periods_per_second",
33
+ "guess_datetime_format",
34
+ "add_overflowsafe",
35
+ "get_supported_dtype",
36
+ "is_supported_dtype",
37
+ ]
38
+
39
+ from pandas._libs.tslibs import dtypes # pylint: disable=import-self
40
+ from pandas._libs.tslibs.conversion import localize_pydatetime
41
+ from pandas._libs.tslibs.dtypes import (
42
+ Resolution,
43
+ periods_per_day,
44
+ periods_per_second,
45
+ )
46
+ from pandas._libs.tslibs.nattype import (
47
+ NaT,
48
+ NaTType,
49
+ iNaT,
50
+ nat_strings,
51
+ )
52
+ from pandas._libs.tslibs.np_datetime import (
53
+ OutOfBoundsDatetime,
54
+ OutOfBoundsTimedelta,
55
+ add_overflowsafe,
56
+ astype_overflowsafe,
57
+ get_supported_dtype,
58
+ is_supported_dtype,
59
+ is_unitless,
60
+ py_get_unit_from_dtype as get_unit_from_dtype,
61
+ )
62
+ from pandas._libs.tslibs.offsets import (
63
+ BaseOffset,
64
+ Tick,
65
+ to_offset,
66
+ )
67
+ from pandas._libs.tslibs.parsing import guess_datetime_format
68
+ from pandas._libs.tslibs.period import (
69
+ IncompatibleFrequency,
70
+ Period,
71
+ )
72
+ from pandas._libs.tslibs.timedeltas import (
73
+ Timedelta,
74
+ delta_to_nanoseconds,
75
+ ints_to_pytimedelta,
76
+ )
77
+ from pandas._libs.tslibs.timestamps import Timestamp
78
+ from pandas._libs.tslibs.timezones import tz_compare
79
+ from pandas._libs.tslibs.tzconversion import tz_convert_from_utc_single
80
+ from pandas._libs.tslibs.vectorized import (
81
+ dt64arr_to_periodarr,
82
+ get_resolution,
83
+ ints_to_pydatetime,
84
+ is_date_array_normalized,
85
+ normalize_i8_timestamps,
86
+ tz_convert_from_utc,
87
+ )
lib/python3.10/site-packages/pandas/_libs/tslibs/base.cpython-310-x86_64-linux-gnu.so ADDED
Binary file (62.3 kB). View file
 
lib/python3.10/site-packages/pandas/_libs/tslibs/ccalendar.pyi ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ DAYS: list[str]
2
+ MONTH_ALIASES: dict[int, str]
3
+ MONTH_NUMBERS: dict[str, int]
4
+ MONTHS: list[str]
5
+ int_to_weekday: dict[int, str]
6
+
7
+ def get_firstbday(year: int, month: int) -> int: ...
8
+ def get_lastbday(year: int, month: int) -> int: ...
9
+ def get_day_of_year(year: int, month: int, day: int) -> int: ...
10
+ def get_iso_calendar(year: int, month: int, day: int) -> tuple[int, int, int]: ...
11
+ def get_week_of_year(year: int, month: int, day: int) -> int: ...
12
+ def get_days_in_month(year: int, month: int) -> int: ...
lib/python3.10/site-packages/pandas/_libs/tslibs/conversion.pyi ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from datetime import (
2
+ datetime,
3
+ tzinfo,
4
+ )
5
+
6
+ import numpy as np
7
+
8
+ DT64NS_DTYPE: np.dtype
9
+ TD64NS_DTYPE: np.dtype
10
+
11
+ def localize_pydatetime(dt: datetime, tz: tzinfo | None) -> datetime: ...
12
+ def cast_from_unit_vectorized(
13
+ values: np.ndarray, unit: str, out_unit: str = ...
14
+ ) -> np.ndarray: ...
lib/python3.10/site-packages/pandas/_libs/tslibs/dtypes.pyi ADDED
@@ -0,0 +1,83 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from enum import Enum
2
+
3
+ OFFSET_TO_PERIOD_FREQSTR: dict[str, str]
4
+
5
+ def periods_per_day(reso: int = ...) -> int: ...
6
+ def periods_per_second(reso: int) -> int: ...
7
+ def abbrev_to_npy_unit(abbrev: str | None) -> int: ...
8
+ def freq_to_period_freqstr(freq_n: int, freq_name: str) -> str: ...
9
+
10
+ class PeriodDtypeBase:
11
+ _dtype_code: int # PeriodDtypeCode
12
+ _n: int
13
+
14
+ # actually __cinit__
15
+ def __new__(cls, code: int, n: int): ...
16
+ @property
17
+ def _freq_group_code(self) -> int: ...
18
+ @property
19
+ def _resolution_obj(self) -> Resolution: ...
20
+ def _get_to_timestamp_base(self) -> int: ...
21
+ @property
22
+ def _freqstr(self) -> str: ...
23
+ def __hash__(self) -> int: ...
24
+ def _is_tick_like(self) -> bool: ...
25
+ @property
26
+ def _creso(self) -> int: ...
27
+ @property
28
+ def _td64_unit(self) -> str: ...
29
+
30
+ class FreqGroup(Enum):
31
+ FR_ANN: int
32
+ FR_QTR: int
33
+ FR_MTH: int
34
+ FR_WK: int
35
+ FR_BUS: int
36
+ FR_DAY: int
37
+ FR_HR: int
38
+ FR_MIN: int
39
+ FR_SEC: int
40
+ FR_MS: int
41
+ FR_US: int
42
+ FR_NS: int
43
+ FR_UND: int
44
+ @staticmethod
45
+ def from_period_dtype_code(code: int) -> FreqGroup: ...
46
+
47
+ class Resolution(Enum):
48
+ RESO_NS: int
49
+ RESO_US: int
50
+ RESO_MS: int
51
+ RESO_SEC: int
52
+ RESO_MIN: int
53
+ RESO_HR: int
54
+ RESO_DAY: int
55
+ RESO_MTH: int
56
+ RESO_QTR: int
57
+ RESO_YR: int
58
+ def __lt__(self, other: Resolution) -> bool: ...
59
+ def __ge__(self, other: Resolution) -> bool: ...
60
+ @property
61
+ def attrname(self) -> str: ...
62
+ @classmethod
63
+ def from_attrname(cls, attrname: str) -> Resolution: ...
64
+ @classmethod
65
+ def get_reso_from_freqstr(cls, freq: str) -> Resolution: ...
66
+ @property
67
+ def attr_abbrev(self) -> str: ...
68
+
69
+ class NpyDatetimeUnit(Enum):
70
+ NPY_FR_Y: int
71
+ NPY_FR_M: int
72
+ NPY_FR_W: int
73
+ NPY_FR_D: int
74
+ NPY_FR_h: int
75
+ NPY_FR_m: int
76
+ NPY_FR_s: int
77
+ NPY_FR_ms: int
78
+ NPY_FR_us: int
79
+ NPY_FR_ns: int
80
+ NPY_FR_ps: int
81
+ NPY_FR_fs: int
82
+ NPY_FR_as: int
83
+ NPY_FR_GENERIC: int
lib/python3.10/site-packages/pandas/_libs/tslibs/fields.pyi ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import numpy as np
2
+
3
+ from pandas._typing import npt
4
+
5
+ def build_field_sarray(
6
+ dtindex: npt.NDArray[np.int64], # const int64_t[:]
7
+ reso: int, # NPY_DATETIMEUNIT
8
+ ) -> np.ndarray: ...
9
+ def month_position_check(fields, weekdays) -> str | None: ...
10
+ def get_date_name_field(
11
+ dtindex: npt.NDArray[np.int64], # const int64_t[:]
12
+ field: str,
13
+ locale: str | None = ...,
14
+ reso: int = ..., # NPY_DATETIMEUNIT
15
+ ) -> npt.NDArray[np.object_]: ...
16
+ def get_start_end_field(
17
+ dtindex: npt.NDArray[np.int64],
18
+ field: str,
19
+ freqstr: str | None = ...,
20
+ month_kw: int = ...,
21
+ reso: int = ..., # NPY_DATETIMEUNIT
22
+ ) -> npt.NDArray[np.bool_]: ...
23
+ def get_date_field(
24
+ dtindex: npt.NDArray[np.int64], # const int64_t[:]
25
+ field: str,
26
+ reso: int = ..., # NPY_DATETIMEUNIT
27
+ ) -> npt.NDArray[np.int32]: ...
28
+ def get_timedelta_field(
29
+ tdindex: npt.NDArray[np.int64], # const int64_t[:]
30
+ field: str,
31
+ reso: int = ..., # NPY_DATETIMEUNIT
32
+ ) -> npt.NDArray[np.int32]: ...
33
+ def get_timedelta_days(
34
+ tdindex: npt.NDArray[np.int64], # const int64_t[:]
35
+ reso: int = ..., # NPY_DATETIMEUNIT
36
+ ) -> npt.NDArray[np.int64]: ...
37
+ def isleapyear_arr(
38
+ years: np.ndarray,
39
+ ) -> npt.NDArray[np.bool_]: ...
40
+ def build_isocalendar_sarray(
41
+ dtindex: npt.NDArray[np.int64], # const int64_t[:]
42
+ reso: int, # NPY_DATETIMEUNIT
43
+ ) -> np.ndarray: ...
44
+ def _get_locale_names(name_type: str, locale: str | None = ...): ...
45
+
46
+ class RoundTo:
47
+ @property
48
+ def MINUS_INFTY(self) -> int: ...
49
+ @property
50
+ def PLUS_INFTY(self) -> int: ...
51
+ @property
52
+ def NEAREST_HALF_EVEN(self) -> int: ...
53
+ @property
54
+ def NEAREST_HALF_PLUS_INFTY(self) -> int: ...
55
+ @property
56
+ def NEAREST_HALF_MINUS_INFTY(self) -> int: ...
57
+
58
+ def round_nsint64(
59
+ values: npt.NDArray[np.int64],
60
+ mode: RoundTo,
61
+ nanos: int,
62
+ ) -> npt.NDArray[np.int64]: ...
lib/python3.10/site-packages/pandas/_libs/tslibs/nattype.pyi ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from datetime import (
2
+ datetime,
3
+ timedelta,
4
+ tzinfo as _tzinfo,
5
+ )
6
+ import typing
7
+
8
+ import numpy as np
9
+
10
+ from pandas._libs.tslibs.period import Period
11
+ from pandas._typing import Self
12
+
13
+ NaT: NaTType
14
+ iNaT: int
15
+ nat_strings: set[str]
16
+
17
+ _NaTComparisonTypes: typing.TypeAlias = (
18
+ datetime | timedelta | Period | np.datetime64 | np.timedelta64
19
+ )
20
+
21
+ class _NatComparison:
22
+ def __call__(self, other: _NaTComparisonTypes) -> bool: ...
23
+
24
+ class NaTType:
25
+ _value: np.int64
26
+ @property
27
+ def value(self) -> int: ...
28
+ @property
29
+ def asm8(self) -> np.datetime64: ...
30
+ def to_datetime64(self) -> np.datetime64: ...
31
+ def to_numpy(
32
+ self, dtype: np.dtype | str | None = ..., copy: bool = ...
33
+ ) -> np.datetime64 | np.timedelta64: ...
34
+ @property
35
+ def is_leap_year(self) -> bool: ...
36
+ @property
37
+ def is_month_start(self) -> bool: ...
38
+ @property
39
+ def is_quarter_start(self) -> bool: ...
40
+ @property
41
+ def is_year_start(self) -> bool: ...
42
+ @property
43
+ def is_month_end(self) -> bool: ...
44
+ @property
45
+ def is_quarter_end(self) -> bool: ...
46
+ @property
47
+ def is_year_end(self) -> bool: ...
48
+ @property
49
+ def day_of_year(self) -> float: ...
50
+ @property
51
+ def dayofyear(self) -> float: ...
52
+ @property
53
+ def days_in_month(self) -> float: ...
54
+ @property
55
+ def daysinmonth(self) -> float: ...
56
+ @property
57
+ def day_of_week(self) -> float: ...
58
+ @property
59
+ def dayofweek(self) -> float: ...
60
+ @property
61
+ def week(self) -> float: ...
62
+ @property
63
+ def weekofyear(self) -> float: ...
64
+ def day_name(self) -> float: ...
65
+ def month_name(self) -> float: ...
66
+ def weekday(self) -> float: ...
67
+ def isoweekday(self) -> float: ...
68
+ def total_seconds(self) -> float: ...
69
+ def today(self, *args, **kwargs) -> NaTType: ...
70
+ def now(self, *args, **kwargs) -> NaTType: ...
71
+ def to_pydatetime(self) -> NaTType: ...
72
+ def date(self) -> NaTType: ...
73
+ def round(self) -> NaTType: ...
74
+ def floor(self) -> NaTType: ...
75
+ def ceil(self) -> NaTType: ...
76
+ @property
77
+ def tzinfo(self) -> None: ...
78
+ @property
79
+ def tz(self) -> None: ...
80
+ def tz_convert(self, tz: _tzinfo | str | None) -> NaTType: ...
81
+ def tz_localize(
82
+ self,
83
+ tz: _tzinfo | str | None,
84
+ ambiguous: str = ...,
85
+ nonexistent: str = ...,
86
+ ) -> NaTType: ...
87
+ def replace(
88
+ self,
89
+ year: int | None = ...,
90
+ month: int | None = ...,
91
+ day: int | None = ...,
92
+ hour: int | None = ...,
93
+ minute: int | None = ...,
94
+ second: int | None = ...,
95
+ microsecond: int | None = ...,
96
+ nanosecond: int | None = ...,
97
+ tzinfo: _tzinfo | None = ...,
98
+ fold: int | None = ...,
99
+ ) -> NaTType: ...
100
+ @property
101
+ def year(self) -> float: ...
102
+ @property
103
+ def quarter(self) -> float: ...
104
+ @property
105
+ def month(self) -> float: ...
106
+ @property
107
+ def day(self) -> float: ...
108
+ @property
109
+ def hour(self) -> float: ...
110
+ @property
111
+ def minute(self) -> float: ...
112
+ @property
113
+ def second(self) -> float: ...
114
+ @property
115
+ def millisecond(self) -> float: ...
116
+ @property
117
+ def microsecond(self) -> float: ...
118
+ @property
119
+ def nanosecond(self) -> float: ...
120
+ # inject Timedelta properties
121
+ @property
122
+ def days(self) -> float: ...
123
+ @property
124
+ def microseconds(self) -> float: ...
125
+ @property
126
+ def nanoseconds(self) -> float: ...
127
+ # inject Period properties
128
+ @property
129
+ def qyear(self) -> float: ...
130
+ def __eq__(self, other: object) -> bool: ...
131
+ def __ne__(self, other: object) -> bool: ...
132
+ __lt__: _NatComparison
133
+ __le__: _NatComparison
134
+ __gt__: _NatComparison
135
+ __ge__: _NatComparison
136
+ def __sub__(self, other: Self | timedelta | datetime) -> Self: ...
137
+ def __rsub__(self, other: Self | timedelta | datetime) -> Self: ...
138
+ def __add__(self, other: Self | timedelta | datetime) -> Self: ...
139
+ def __radd__(self, other: Self | timedelta | datetime) -> Self: ...
140
+ def __hash__(self) -> int: ...
141
+ def as_unit(self, unit: str, round_ok: bool = ...) -> NaTType: ...
lib/python3.10/site-packages/pandas/_libs/tslibs/np_datetime.pyi ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import numpy as np
2
+
3
+ from pandas._typing import npt
4
+
5
+ class OutOfBoundsDatetime(ValueError): ...
6
+ class OutOfBoundsTimedelta(ValueError): ...
7
+
8
+ # only exposed for testing
9
+ def py_get_unit_from_dtype(dtype: np.dtype): ...
10
+ def py_td64_to_tdstruct(td64: int, unit: int) -> dict: ...
11
+ def astype_overflowsafe(
12
+ values: np.ndarray,
13
+ dtype: np.dtype,
14
+ copy: bool = ...,
15
+ round_ok: bool = ...,
16
+ is_coerce: bool = ...,
17
+ ) -> np.ndarray: ...
18
+ def is_unitless(dtype: np.dtype) -> bool: ...
19
+ def compare_mismatched_resolutions(
20
+ left: np.ndarray, right: np.ndarray, op
21
+ ) -> npt.NDArray[np.bool_]: ...
22
+ def add_overflowsafe(
23
+ left: npt.NDArray[np.int64],
24
+ right: npt.NDArray[np.int64],
25
+ ) -> npt.NDArray[np.int64]: ...
26
+ def get_supported_dtype(dtype: np.dtype) -> np.dtype: ...
27
+ def is_supported_dtype(dtype: np.dtype) -> bool: ...