xiaoanyu123 commited on
Commit
94c0edb
·
verified ·
1 Parent(s): 9c5fbee

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +1 -0
  2. pythonProject/.venv/Lib/site-packages/Jinja2-3.1.3.dist-info/INSTALLER +1 -0
  3. pythonProject/.venv/Lib/site-packages/Jinja2-3.1.3.dist-info/LICENSE.rst +28 -0
  4. pythonProject/.venv/Lib/site-packages/Jinja2-3.1.3.dist-info/METADATA +105 -0
  5. pythonProject/.venv/Lib/site-packages/Jinja2-3.1.3.dist-info/RECORD +58 -0
  6. pythonProject/.venv/Lib/site-packages/Jinja2-3.1.3.dist-info/WHEEL +5 -0
  7. pythonProject/.venv/Lib/site-packages/Jinja2-3.1.3.dist-info/entry_points.txt +2 -0
  8. pythonProject/.venv/Lib/site-packages/Jinja2-3.1.3.dist-info/top_level.txt +1 -0
  9. pythonProject/.venv/Lib/site-packages/huggingface_hub-0.34.4.dist-info/INSTALLER +1 -0
  10. pythonProject/.venv/Lib/site-packages/huggingface_hub-0.34.4.dist-info/LICENSE +201 -0
  11. pythonProject/.venv/Lib/site-packages/huggingface_hub-0.34.4.dist-info/METADATA +331 -0
  12. pythonProject/.venv/Lib/site-packages/huggingface_hub-0.34.4.dist-info/RECORD +327 -0
  13. pythonProject/.venv/Lib/site-packages/huggingface_hub-0.34.4.dist-info/WHEEL +5 -0
  14. pythonProject/.venv/Lib/site-packages/huggingface_hub-0.34.4.dist-info/entry_points.txt +8 -0
  15. pythonProject/.venv/Lib/site-packages/huggingface_hub-0.34.4.dist-info/top_level.txt +1 -0
  16. pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/__init__.cpython-310.pyc +0 -0
  17. pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/_commit_api.cpython-310.pyc +0 -0
  18. pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/_commit_scheduler.cpython-310.pyc +0 -0
  19. pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/_inference_endpoints.cpython-310.pyc +0 -0
  20. pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/_jobs_api.cpython-310.pyc +0 -0
  21. pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/_tensorboard_logger.cpython-310.pyc +0 -0
  22. pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/_upload_large_folder.cpython-310.pyc +0 -0
  23. pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/_webhooks_payload.cpython-310.pyc +0 -0
  24. pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/_webhooks_server.cpython-310.pyc +0 -0
  25. pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/community.cpython-310.pyc +0 -0
  26. pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/constants.cpython-310.pyc +0 -0
  27. pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/dataclasses.cpython-310.pyc +0 -0
  28. pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/errors.cpython-310.pyc +0 -0
  29. pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/fastai_utils.cpython-310.pyc +0 -0
  30. pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/file_download.cpython-310.pyc +0 -0
  31. pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/hf_file_system.cpython-310.pyc +0 -0
  32. pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/hub_mixin.cpython-310.pyc +0 -0
  33. pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/inference_api.cpython-310.pyc +0 -0
  34. pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/keras_mixin.cpython-310.pyc +0 -0
  35. pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/lfs.cpython-310.pyc +0 -0
  36. pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/repocard.cpython-310.pyc +0 -0
  37. pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/repocard_data.cpython-310.pyc +0 -0
  38. pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/repository.cpython-310.pyc +0 -0
  39. pythonProject/.venv/Lib/site-packages/huggingface_hub/inference/__pycache__/__init__.cpython-310.pyc +0 -0
  40. pythonProject/.venv/Lib/site-packages/huggingface_hub/inference/__pycache__/_client.cpython-310.pyc +3 -0
  41. pythonProject/.venv/Lib/site-packages/huggingface_hub/inference/__pycache__/_common.cpython-310.pyc +0 -0
  42. pythonProject/.venv/Lib/site-packages/huggingface_hub/inference/_providers/__pycache__/__init__.cpython-310.pyc +0 -0
  43. pythonProject/.venv/Lib/site-packages/huggingface_hub/inference/_providers/__pycache__/_common.cpython-310.pyc +0 -0
  44. pythonProject/.venv/Lib/site-packages/huggingface_hub/inference/_providers/__pycache__/replicate.cpython-310.pyc +0 -0
  45. pythonProject/.venv/Lib/site-packages/huggingface_hub/inference/_providers/__pycache__/sambanova.cpython-310.pyc +0 -0
  46. pythonProject/.venv/Lib/site-packages/huggingface_hub/inference/_providers/__pycache__/together.cpython-310.pyc +0 -0
  47. pythonProject/.venv/Lib/site-packages/huggingface_hub/serialization/__init__.py +27 -0
  48. pythonProject/.venv/Lib/site-packages/huggingface_hub/serialization/_base.py +210 -0
  49. pythonProject/.venv/Lib/site-packages/huggingface_hub/serialization/_dduf.py +387 -0
  50. pythonProject/.venv/Lib/site-packages/huggingface_hub/serialization/_tensorflow.py +95 -0
.gitattributes CHANGED
@@ -50,3 +50,4 @@ cudnn-windows-x86_64-8.9.5.30_cuda11-archive/bin/cudnn_ops_train64_8.dll filter=
50
  pythonProject/.venv/Lib/site-packages/diffusers/loaders/__pycache__/lora_pipeline.cpython-310.pyc filter=lfs diff=lfs merge=lfs -text
51
  pythonProject/.venv/Lib/site-packages/diffusers/utils/__pycache__/dummy_torch_and_transformers_objects.cpython-310.pyc filter=lfs diff=lfs merge=lfs -text
52
  pythonProject/.venv/Lib/site-packages/functorch/_C.cp310-win_amd64.pyd filter=lfs diff=lfs merge=lfs -text
 
 
50
  pythonProject/.venv/Lib/site-packages/diffusers/loaders/__pycache__/lora_pipeline.cpython-310.pyc filter=lfs diff=lfs merge=lfs -text
51
  pythonProject/.venv/Lib/site-packages/diffusers/utils/__pycache__/dummy_torch_and_transformers_objects.cpython-310.pyc filter=lfs diff=lfs merge=lfs -text
52
  pythonProject/.venv/Lib/site-packages/functorch/_C.cp310-win_amd64.pyd filter=lfs diff=lfs merge=lfs -text
53
+ pythonProject/.venv/Lib/site-packages/huggingface_hub/inference/__pycache__/_client.cpython-310.pyc filter=lfs diff=lfs merge=lfs -text
pythonProject/.venv/Lib/site-packages/Jinja2-3.1.3.dist-info/INSTALLER ADDED
@@ -0,0 +1 @@
 
 
1
+ pip
pythonProject/.venv/Lib/site-packages/Jinja2-3.1.3.dist-info/LICENSE.rst ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Copyright 2007 Pallets
2
+
3
+ Redistribution and use in source and binary forms, with or without
4
+ modification, are permitted provided that the following conditions are
5
+ met:
6
+
7
+ 1. Redistributions of source code must retain the above copyright
8
+ notice, this list of conditions and the following disclaimer.
9
+
10
+ 2. Redistributions in binary form must reproduce the above copyright
11
+ notice, this list of conditions and the following disclaimer in the
12
+ documentation and/or other materials provided with the distribution.
13
+
14
+ 3. Neither the name of the copyright holder nor the names of its
15
+ contributors may be used to endorse or promote products derived from
16
+ this software without specific prior written permission.
17
+
18
+ THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
19
+ "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
20
+ LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A
21
+ PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
22
+ HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
23
+ SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED
24
+ TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
25
+ PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF
26
+ LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING
27
+ NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
28
+ SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
pythonProject/.venv/Lib/site-packages/Jinja2-3.1.3.dist-info/METADATA ADDED
@@ -0,0 +1,105 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metadata-Version: 2.1
2
+ Name: Jinja2
3
+ Version: 3.1.3
4
+ Summary: A very fast and expressive template engine.
5
+ Home-page: https://palletsprojects.com/p/jinja/
6
+ Maintainer: Pallets
7
+ Maintainer-email: contact@palletsprojects.com
8
+ License: BSD-3-Clause
9
+ Project-URL: Donate, https://palletsprojects.com/donate
10
+ Project-URL: Documentation, https://jinja.palletsprojects.com/
11
+ Project-URL: Changes, https://jinja.palletsprojects.com/changes/
12
+ Project-URL: Source Code, https://github.com/pallets/jinja/
13
+ Project-URL: Issue Tracker, https://github.com/pallets/jinja/issues/
14
+ Project-URL: Chat, https://discord.gg/pallets
15
+ Classifier: Development Status :: 5 - Production/Stable
16
+ Classifier: Environment :: Web Environment
17
+ Classifier: Intended Audience :: Developers
18
+ Classifier: License :: OSI Approved :: BSD License
19
+ Classifier: Operating System :: OS Independent
20
+ Classifier: Programming Language :: Python
21
+ Classifier: Topic :: Internet :: WWW/HTTP :: Dynamic Content
22
+ Classifier: Topic :: Text Processing :: Markup :: HTML
23
+ Requires-Python: >=3.7
24
+ Description-Content-Type: text/x-rst
25
+ License-File: LICENSE.rst
26
+ Requires-Dist: MarkupSafe >=2.0
27
+ Provides-Extra: i18n
28
+ Requires-Dist: Babel >=2.7 ; extra == 'i18n'
29
+
30
+ Jinja
31
+ =====
32
+
33
+ Jinja is a fast, expressive, extensible templating engine. Special
34
+ placeholders in the template allow writing code similar to Python
35
+ syntax. Then the template is passed data to render the final document.
36
+
37
+ It includes:
38
+
39
+ - Template inheritance and inclusion.
40
+ - Define and import macros within templates.
41
+ - HTML templates can use autoescaping to prevent XSS from untrusted
42
+ user input.
43
+ - A sandboxed environment can safely render untrusted templates.
44
+ - AsyncIO support for generating templates and calling async
45
+ functions.
46
+ - I18N support with Babel.
47
+ - Templates are compiled to optimized Python code just-in-time and
48
+ cached, or can be compiled ahead-of-time.
49
+ - Exceptions point to the correct line in templates to make debugging
50
+ easier.
51
+ - Extensible filters, tests, functions, and even syntax.
52
+
53
+ Jinja's philosophy is that while application logic belongs in Python if
54
+ possible, it shouldn't make the template designer's job difficult by
55
+ restricting functionality too much.
56
+
57
+
58
+ Installing
59
+ ----------
60
+
61
+ Install and update using `pip`_:
62
+
63
+ .. code-block:: text
64
+
65
+ $ pip install -U Jinja2
66
+
67
+ .. _pip: https://pip.pypa.io/en/stable/getting-started/
68
+
69
+
70
+ In A Nutshell
71
+ -------------
72
+
73
+ .. code-block:: jinja
74
+
75
+ {% extends "base.html" %}
76
+ {% block title %}Members{% endblock %}
77
+ {% block content %}
78
+ <ul>
79
+ {% for user in users %}
80
+ <li><a href="{{ user.url }}">{{ user.username }}</a></li>
81
+ {% endfor %}
82
+ </ul>
83
+ {% endblock %}
84
+
85
+
86
+ Donate
87
+ ------
88
+
89
+ The Pallets organization develops and supports Jinja and other popular
90
+ packages. In order to grow the community of contributors and users, and
91
+ allow the maintainers to devote more time to the projects, `please
92
+ donate today`_.
93
+
94
+ .. _please donate today: https://palletsprojects.com/donate
95
+
96
+
97
+ Links
98
+ -----
99
+
100
+ - Documentation: https://jinja.palletsprojects.com/
101
+ - Changes: https://jinja.palletsprojects.com/changes/
102
+ - PyPI Releases: https://pypi.org/project/Jinja2/
103
+ - Source Code: https://github.com/pallets/jinja/
104
+ - Issue Tracker: https://github.com/pallets/jinja/issues/
105
+ - Chat: https://discord.gg/pallets
pythonProject/.venv/Lib/site-packages/Jinja2-3.1.3.dist-info/RECORD ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Jinja2-3.1.3.dist-info/INSTALLER,sha256=zuuue4knoyJ-UwPPXg8fezS7VCrXJQrAP7zeNuwvFQg,4
2
+ Jinja2-3.1.3.dist-info/LICENSE.rst,sha256=O0nc7kEF6ze6wQ-vG-JgQI_oXSUrjp3y4JefweCUQ3s,1475
3
+ Jinja2-3.1.3.dist-info/METADATA,sha256=0cLNbRCI91jytc7Bzv3XAQfZzFDF2gxkJuH46eF5vew,3301
4
+ Jinja2-3.1.3.dist-info/RECORD,,
5
+ Jinja2-3.1.3.dist-info/WHEEL,sha256=oiQVh_5PnQM0E3gPdiz09WCNmwiHDMaGer_elqB3coM,92
6
+ Jinja2-3.1.3.dist-info/entry_points.txt,sha256=zRd62fbqIyfUpsRtU7EVIFyiu1tPwfgO7EvPErnxgTE,59
7
+ Jinja2-3.1.3.dist-info/top_level.txt,sha256=PkeVWtLb3-CqjWi1fO29OCbj55EhX_chhKrCdrVe_zs,7
8
+ jinja2/__init__.py,sha256=NTBwMwsECrdHmxeXF7seusHLzrh6Ldn1A9qhS5cDuf0,1927
9
+ jinja2/__pycache__/__init__.cpython-310.pyc,,
10
+ jinja2/__pycache__/_identifier.cpython-310.pyc,,
11
+ jinja2/__pycache__/async_utils.cpython-310.pyc,,
12
+ jinja2/__pycache__/bccache.cpython-310.pyc,,
13
+ jinja2/__pycache__/compiler.cpython-310.pyc,,
14
+ jinja2/__pycache__/constants.cpython-310.pyc,,
15
+ jinja2/__pycache__/debug.cpython-310.pyc,,
16
+ jinja2/__pycache__/defaults.cpython-310.pyc,,
17
+ jinja2/__pycache__/environment.cpython-310.pyc,,
18
+ jinja2/__pycache__/exceptions.cpython-310.pyc,,
19
+ jinja2/__pycache__/ext.cpython-310.pyc,,
20
+ jinja2/__pycache__/filters.cpython-310.pyc,,
21
+ jinja2/__pycache__/idtracking.cpython-310.pyc,,
22
+ jinja2/__pycache__/lexer.cpython-310.pyc,,
23
+ jinja2/__pycache__/loaders.cpython-310.pyc,,
24
+ jinja2/__pycache__/meta.cpython-310.pyc,,
25
+ jinja2/__pycache__/nativetypes.cpython-310.pyc,,
26
+ jinja2/__pycache__/nodes.cpython-310.pyc,,
27
+ jinja2/__pycache__/optimizer.cpython-310.pyc,,
28
+ jinja2/__pycache__/parser.cpython-310.pyc,,
29
+ jinja2/__pycache__/runtime.cpython-310.pyc,,
30
+ jinja2/__pycache__/sandbox.cpython-310.pyc,,
31
+ jinja2/__pycache__/tests.cpython-310.pyc,,
32
+ jinja2/__pycache__/utils.cpython-310.pyc,,
33
+ jinja2/__pycache__/visitor.cpython-310.pyc,,
34
+ jinja2/_identifier.py,sha256=_zYctNKzRqlk_murTNlzrju1FFJL7Va_Ijqqd7ii2lU,1958
35
+ jinja2/async_utils.py,sha256=dFcmh6lMNfbh7eLKrBio8JqAKLHdZbpCuurFN4OERtY,2447
36
+ jinja2/bccache.py,sha256=mhz5xtLxCcHRAa56azOhphIAe19u1we0ojifNMClDio,14061
37
+ jinja2/compiler.py,sha256=PJzYdRLStlEOqmnQs1YxlizPrJoj3jTZuUleREn6AIQ,72199
38
+ jinja2/constants.py,sha256=GMoFydBF_kdpaRKPoM5cl5MviquVRLVyZtfp5-16jg0,1433
39
+ jinja2/debug.py,sha256=iWJ432RadxJNnaMOPrjIDInz50UEgni3_HKuFXi2vuQ,6299
40
+ jinja2/defaults.py,sha256=boBcSw78h-lp20YbaXSJsqkAI2uN_mD_TtCydpeq5wU,1267
41
+ jinja2/environment.py,sha256=0qldX3VQKZcm6lgn7zHz94oRFow7YPYERiqkquomNjU,61253
42
+ jinja2/exceptions.py,sha256=ioHeHrWwCWNaXX1inHmHVblvc4haO7AXsjCp3GfWvx0,5071
43
+ jinja2/ext.py,sha256=5fnMpllaXkfm2P_93RIvi-OnK7Tk8mCW8Du-GcD12Hc,31844
44
+ jinja2/filters.py,sha256=vYjKb2zaPShvYtn_LpSmqfS8SScbrA_KOanNibsMDIE,53862
45
+ jinja2/idtracking.py,sha256=GfNmadir4oDALVxzn3DL9YInhJDr69ebXeA2ygfuCGA,10704
46
+ jinja2/lexer.py,sha256=DW2nX9zk-6MWp65YR2bqqj0xqCvLtD-u9NWT8AnFRxQ,29726
47
+ jinja2/loaders.py,sha256=ayAwxfrA1SAffQta0nwSDm3TDT4KYiIGN_D9Z45B310,23085
48
+ jinja2/meta.py,sha256=GNPEvifmSaU3CMxlbheBOZjeZ277HThOPUTf1RkppKQ,4396
49
+ jinja2/nativetypes.py,sha256=7GIGALVJgdyL80oZJdQUaUfwSt5q2lSSZbXt0dNf_M4,4210
50
+ jinja2/nodes.py,sha256=i34GPRAZexXMT6bwuf5SEyvdmS-bRCy9KMjwN5O6pjk,34550
51
+ jinja2/optimizer.py,sha256=tHkMwXxfZkbfA1KmLcqmBMSaz7RLIvvItrJcPoXTyD8,1650
52
+ jinja2/parser.py,sha256=Y199wPL-G67gJoi5G_5sHuu9uEP1PJkjjLEW_xTH8-k,39736
53
+ jinja2/py.typed,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
54
+ jinja2/runtime.py,sha256=_6LkKIWFJjQdqlrgA3K39zBFQ-7Orm3wGDm96RwxQoE,33406
55
+ jinja2/sandbox.py,sha256=Y0xZeXQnH6EX5VjaV2YixESxoepnRbW_3UeQosaBU3M,14584
56
+ jinja2/tests.py,sha256=Am5Z6Lmfr2XaH_npIfJJ8MdXtWsbLjMULZJulTAj30E,5905
57
+ jinja2/utils.py,sha256=IMwRIcN1SsTw2-jdQtlH2KzNABsXZBW_-tnFXafQBvY,23933
58
+ jinja2/visitor.py,sha256=MH14C6yq24G_KVtWzjwaI7Wg14PCJIYlWW1kpkxYak0,3568
pythonProject/.venv/Lib/site-packages/Jinja2-3.1.3.dist-info/WHEEL ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ Wheel-Version: 1.0
2
+ Generator: bdist_wheel (0.42.0)
3
+ Root-Is-Purelib: true
4
+ Tag: py3-none-any
5
+
pythonProject/.venv/Lib/site-packages/Jinja2-3.1.3.dist-info/entry_points.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ [babel.extractors]
2
+ jinja2 = jinja2.ext:babel_extract[i18n]
pythonProject/.venv/Lib/site-packages/Jinja2-3.1.3.dist-info/top_level.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ jinja2
pythonProject/.venv/Lib/site-packages/huggingface_hub-0.34.4.dist-info/INSTALLER ADDED
@@ -0,0 +1 @@
 
 
1
+ pip
pythonProject/.venv/Lib/site-packages/huggingface_hub-0.34.4.dist-info/LICENSE ADDED
@@ -0,0 +1,201 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Apache License
2
+ Version 2.0, January 2004
3
+ http://www.apache.org/licenses/
4
+
5
+ TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION
6
+
7
+ 1. Definitions.
8
+
9
+ "License" shall mean the terms and conditions for use, reproduction,
10
+ and distribution as defined by Sections 1 through 9 of this document.
11
+
12
+ "Licensor" shall mean the copyright owner or entity authorized by
13
+ the copyright owner that is granting the License.
14
+
15
+ "Legal Entity" shall mean the union of the acting entity and all
16
+ other entities that control, are controlled by, or are under common
17
+ control with that entity. For the purposes of this definition,
18
+ "control" means (i) the power, direct or indirect, to cause the
19
+ direction or management of such entity, whether by contract or
20
+ otherwise, or (ii) ownership of fifty percent (50%) or more of the
21
+ outstanding shares, or (iii) beneficial ownership of such entity.
22
+
23
+ "You" (or "Your") shall mean an individual or Legal Entity
24
+ exercising permissions granted by this License.
25
+
26
+ "Source" form shall mean the preferred form for making modifications,
27
+ including but not limited to software source code, documentation
28
+ source, and configuration files.
29
+
30
+ "Object" form shall mean any form resulting from mechanical
31
+ transformation or translation of a Source form, including but
32
+ not limited to compiled object code, generated documentation,
33
+ and conversions to other media types.
34
+
35
+ "Work" shall mean the work of authorship, whether in Source or
36
+ Object form, made available under the License, as indicated by a
37
+ copyright notice that is included in or attached to the work
38
+ (an example is provided in the Appendix below).
39
+
40
+ "Derivative Works" shall mean any work, whether in Source or Object
41
+ form, that is based on (or derived from) the Work and for which the
42
+ editorial revisions, annotations, elaborations, or other modifications
43
+ represent, as a whole, an original work of authorship. For the purposes
44
+ of this License, Derivative Works shall not include works that remain
45
+ separable from, or merely link (or bind by name) to the interfaces of,
46
+ the Work and Derivative Works thereof.
47
+
48
+ "Contribution" shall mean any work of authorship, including
49
+ the original version of the Work and any modifications or additions
50
+ to that Work or Derivative Works thereof, that is intentionally
51
+ submitted to Licensor for inclusion in the Work by the copyright owner
52
+ or by an individual or Legal Entity authorized to submit on behalf of
53
+ the copyright owner. For the purposes of this definition, "submitted"
54
+ means any form of electronic, verbal, or written communication sent
55
+ to the Licensor or its representatives, including but not limited to
56
+ communication on electronic mailing lists, source code control systems,
57
+ and issue tracking systems that are managed by, or on behalf of, the
58
+ Licensor for the purpose of discussing and improving the Work, but
59
+ excluding communication that is conspicuously marked or otherwise
60
+ designated in writing by the copyright owner as "Not a Contribution."
61
+
62
+ "Contributor" shall mean Licensor and any individual or Legal Entity
63
+ on behalf of whom a Contribution has been received by Licensor and
64
+ subsequently incorporated within the Work.
65
+
66
+ 2. Grant of Copyright License. Subject to the terms and conditions of
67
+ this License, each Contributor hereby grants to You a perpetual,
68
+ worldwide, non-exclusive, no-charge, royalty-free, irrevocable
69
+ copyright license to reproduce, prepare Derivative Works of,
70
+ publicly display, publicly perform, sublicense, and distribute the
71
+ Work and such Derivative Works in Source or Object form.
72
+
73
+ 3. Grant of Patent License. Subject to the terms and conditions of
74
+ this License, each Contributor hereby grants to You a perpetual,
75
+ worldwide, non-exclusive, no-charge, royalty-free, irrevocable
76
+ (except as stated in this section) patent license to make, have made,
77
+ use, offer to sell, sell, import, and otherwise transfer the Work,
78
+ where such license applies only to those patent claims licensable
79
+ by such Contributor that are necessarily infringed by their
80
+ Contribution(s) alone or by combination of their Contribution(s)
81
+ with the Work to which such Contribution(s) was submitted. If You
82
+ institute patent litigation against any entity (including a
83
+ cross-claim or counterclaim in a lawsuit) alleging that the Work
84
+ or a Contribution incorporated within the Work constitutes direct
85
+ or contributory patent infringement, then any patent licenses
86
+ granted to You under this License for that Work shall terminate
87
+ as of the date such litigation is filed.
88
+
89
+ 4. Redistribution. You may reproduce and distribute copies of the
90
+ Work or Derivative Works thereof in any medium, with or without
91
+ modifications, and in Source or Object form, provided that You
92
+ meet the following conditions:
93
+
94
+ (a) You must give any other recipients of the Work or
95
+ Derivative Works a copy of this License; and
96
+
97
+ (b) You must cause any modified files to carry prominent notices
98
+ stating that You changed the files; and
99
+
100
+ (c) You must retain, in the Source form of any Derivative Works
101
+ that You distribute, all copyright, patent, trademark, and
102
+ attribution notices from the Source form of the Work,
103
+ excluding those notices that do not pertain to any part of
104
+ the Derivative Works; and
105
+
106
+ (d) If the Work includes a "NOTICE" text file as part of its
107
+ distribution, then any Derivative Works that You distribute must
108
+ include a readable copy of the attribution notices contained
109
+ within such NOTICE file, excluding those notices that do not
110
+ pertain to any part of the Derivative Works, in at least one
111
+ of the following places: within a NOTICE text file distributed
112
+ as part of the Derivative Works; within the Source form or
113
+ documentation, if provided along with the Derivative Works; or,
114
+ within a display generated by the Derivative Works, if and
115
+ wherever such third-party notices normally appear. The contents
116
+ of the NOTICE file are for informational purposes only and
117
+ do not modify the License. You may add Your own attribution
118
+ notices within Derivative Works that You distribute, alongside
119
+ or as an addendum to the NOTICE text from the Work, provided
120
+ that such additional attribution notices cannot be construed
121
+ as modifying the License.
122
+
123
+ You may add Your own copyright statement to Your modifications and
124
+ may provide additional or different license terms and conditions
125
+ for use, reproduction, or distribution of Your modifications, or
126
+ for any such Derivative Works as a whole, provided Your use,
127
+ reproduction, and distribution of the Work otherwise complies with
128
+ the conditions stated in this License.
129
+
130
+ 5. Submission of Contributions. Unless You explicitly state otherwise,
131
+ any Contribution intentionally submitted for inclusion in the Work
132
+ by You to the Licensor shall be under the terms and conditions of
133
+ this License, without any additional terms or conditions.
134
+ Notwithstanding the above, nothing herein shall supersede or modify
135
+ the terms of any separate license agreement you may have executed
136
+ with Licensor regarding such Contributions.
137
+
138
+ 6. Trademarks. This License does not grant permission to use the trade
139
+ names, trademarks, service marks, or product names of the Licensor,
140
+ except as required for reasonable and customary use in describing the
141
+ origin of the Work and reproducing the content of the NOTICE file.
142
+
143
+ 7. Disclaimer of Warranty. Unless required by applicable law or
144
+ agreed to in writing, Licensor provides the Work (and each
145
+ Contributor provides its Contributions) on an "AS IS" BASIS,
146
+ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
147
+ implied, including, without limitation, any warranties or conditions
148
+ of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A
149
+ PARTICULAR PURPOSE. You are solely responsible for determining the
150
+ appropriateness of using or redistributing the Work and assume any
151
+ risks associated with Your exercise of permissions under this License.
152
+
153
+ 8. Limitation of Liability. In no event and under no legal theory,
154
+ whether in tort (including negligence), contract, or otherwise,
155
+ unless required by applicable law (such as deliberate and grossly
156
+ negligent acts) or agreed to in writing, shall any Contributor be
157
+ liable to You for damages, including any direct, indirect, special,
158
+ incidental, or consequential damages of any character arising as a
159
+ result of this License or out of the use or inability to use the
160
+ Work (including but not limited to damages for loss of goodwill,
161
+ work stoppage, computer failure or malfunction, or any and all
162
+ other commercial damages or losses), even if such Contributor
163
+ has been advised of the possibility of such damages.
164
+
165
+ 9. Accepting Warranty or Additional Liability. While redistributing
166
+ the Work or Derivative Works thereof, You may choose to offer,
167
+ and charge a fee for, acceptance of support, warranty, indemnity,
168
+ or other liability obligations and/or rights consistent with this
169
+ License. However, in accepting such obligations, You may act only
170
+ on Your own behalf and on Your sole responsibility, not on behalf
171
+ of any other Contributor, and only if You agree to indemnify,
172
+ defend, and hold each Contributor harmless for any liability
173
+ incurred by, or claims asserted against, such Contributor by reason
174
+ of your accepting any such warranty or additional liability.
175
+
176
+ END OF TERMS AND CONDITIONS
177
+
178
+ APPENDIX: How to apply the Apache License to your work.
179
+
180
+ To apply the Apache License to your work, attach the following
181
+ boilerplate notice, with the fields enclosed by brackets "[]"
182
+ replaced with your own identifying information. (Don't include
183
+ the brackets!) The text should be enclosed in the appropriate
184
+ comment syntax for the file format. We also recommend that a
185
+ file or class name and description of purpose be included on the
186
+ same "printed page" as the copyright notice for easier
187
+ identification within third-party archives.
188
+
189
+ Copyright [yyyy] [name of copyright owner]
190
+
191
+ Licensed under the Apache License, Version 2.0 (the "License");
192
+ you may not use this file except in compliance with the License.
193
+ You may obtain a copy of the License at
194
+
195
+ http://www.apache.org/licenses/LICENSE-2.0
196
+
197
+ Unless required by applicable law or agreed to in writing, software
198
+ distributed under the License is distributed on an "AS IS" BASIS,
199
+ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
200
+ See the License for the specific language governing permissions and
201
+ limitations under the License.
pythonProject/.venv/Lib/site-packages/huggingface_hub-0.34.4.dist-info/METADATA ADDED
@@ -0,0 +1,331 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metadata-Version: 2.1
2
+ Name: huggingface-hub
3
+ Version: 0.34.4
4
+ Summary: Client library to download and publish models, datasets and other repos on the huggingface.co hub
5
+ Home-page: https://github.com/huggingface/huggingface_hub
6
+ Author: Hugging Face, Inc.
7
+ Author-email: julien@huggingface.co
8
+ License: Apache
9
+ Keywords: model-hub machine-learning models natural-language-processing deep-learning pytorch pretrained-models
10
+ Platform: UNKNOWN
11
+ Classifier: Intended Audience :: Developers
12
+ Classifier: Intended Audience :: Education
13
+ Classifier: Intended Audience :: Science/Research
14
+ Classifier: License :: OSI Approved :: Apache Software License
15
+ Classifier: Operating System :: OS Independent
16
+ Classifier: Programming Language :: Python :: 3
17
+ Classifier: Programming Language :: Python :: 3 :: Only
18
+ Classifier: Programming Language :: Python :: 3.8
19
+ Classifier: Programming Language :: Python :: 3.9
20
+ Classifier: Programming Language :: Python :: 3.10
21
+ Classifier: Programming Language :: Python :: 3.11
22
+ Classifier: Programming Language :: Python :: 3.12
23
+ Classifier: Programming Language :: Python :: 3.13
24
+ Classifier: Topic :: Scientific/Engineering :: Artificial Intelligence
25
+ Requires-Python: >=3.8.0
26
+ Description-Content-Type: text/markdown
27
+ License-File: LICENSE
28
+ Requires-Dist: filelock
29
+ Requires-Dist: fsspec>=2023.5.0
30
+ Requires-Dist: packaging>=20.9
31
+ Requires-Dist: pyyaml>=5.1
32
+ Requires-Dist: requests
33
+ Requires-Dist: tqdm>=4.42.1
34
+ Requires-Dist: typing-extensions>=3.7.4.3
35
+ Requires-Dist: hf-xet<2.0.0,>=1.1.3; platform_machine == "x86_64" or platform_machine == "amd64" or platform_machine == "arm64" or platform_machine == "aarch64"
36
+ Provides-Extra: all
37
+ Requires-Dist: InquirerPy==0.3.4; extra == "all"
38
+ Requires-Dist: aiohttp; extra == "all"
39
+ Requires-Dist: authlib>=1.3.2; extra == "all"
40
+ Requires-Dist: fastapi; extra == "all"
41
+ Requires-Dist: httpx; extra == "all"
42
+ Requires-Dist: itsdangerous; extra == "all"
43
+ Requires-Dist: jedi; extra == "all"
44
+ Requires-Dist: Jinja2; extra == "all"
45
+ Requires-Dist: pytest<8.2.2,>=8.1.1; extra == "all"
46
+ Requires-Dist: pytest-cov; extra == "all"
47
+ Requires-Dist: pytest-env; extra == "all"
48
+ Requires-Dist: pytest-xdist; extra == "all"
49
+ Requires-Dist: pytest-vcr; extra == "all"
50
+ Requires-Dist: pytest-asyncio; extra == "all"
51
+ Requires-Dist: pytest-rerunfailures; extra == "all"
52
+ Requires-Dist: pytest-mock; extra == "all"
53
+ Requires-Dist: urllib3<2.0; extra == "all"
54
+ Requires-Dist: soundfile; extra == "all"
55
+ Requires-Dist: Pillow; extra == "all"
56
+ Requires-Dist: gradio>=4.0.0; extra == "all"
57
+ Requires-Dist: numpy; extra == "all"
58
+ Requires-Dist: ruff>=0.9.0; extra == "all"
59
+ Requires-Dist: libcst>=1.4.0; extra == "all"
60
+ Requires-Dist: typing-extensions>=4.8.0; extra == "all"
61
+ Requires-Dist: types-PyYAML; extra == "all"
62
+ Requires-Dist: types-requests; extra == "all"
63
+ Requires-Dist: types-simplejson; extra == "all"
64
+ Requires-Dist: types-toml; extra == "all"
65
+ Requires-Dist: types-tqdm; extra == "all"
66
+ Requires-Dist: types-urllib3; extra == "all"
67
+ Requires-Dist: mypy<1.15.0,>=1.14.1; python_version == "3.8" and extra == "all"
68
+ Requires-Dist: mypy==1.15.0; python_version >= "3.9" and extra == "all"
69
+ Provides-Extra: cli
70
+ Requires-Dist: InquirerPy==0.3.4; extra == "cli"
71
+ Provides-Extra: dev
72
+ Requires-Dist: InquirerPy==0.3.4; extra == "dev"
73
+ Requires-Dist: aiohttp; extra == "dev"
74
+ Requires-Dist: authlib>=1.3.2; extra == "dev"
75
+ Requires-Dist: fastapi; extra == "dev"
76
+ Requires-Dist: httpx; extra == "dev"
77
+ Requires-Dist: itsdangerous; extra == "dev"
78
+ Requires-Dist: jedi; extra == "dev"
79
+ Requires-Dist: Jinja2; extra == "dev"
80
+ Requires-Dist: pytest<8.2.2,>=8.1.1; extra == "dev"
81
+ Requires-Dist: pytest-cov; extra == "dev"
82
+ Requires-Dist: pytest-env; extra == "dev"
83
+ Requires-Dist: pytest-xdist; extra == "dev"
84
+ Requires-Dist: pytest-vcr; extra == "dev"
85
+ Requires-Dist: pytest-asyncio; extra == "dev"
86
+ Requires-Dist: pytest-rerunfailures; extra == "dev"
87
+ Requires-Dist: pytest-mock; extra == "dev"
88
+ Requires-Dist: urllib3<2.0; extra == "dev"
89
+ Requires-Dist: soundfile; extra == "dev"
90
+ Requires-Dist: Pillow; extra == "dev"
91
+ Requires-Dist: gradio>=4.0.0; extra == "dev"
92
+ Requires-Dist: numpy; extra == "dev"
93
+ Requires-Dist: ruff>=0.9.0; extra == "dev"
94
+ Requires-Dist: libcst>=1.4.0; extra == "dev"
95
+ Requires-Dist: typing-extensions>=4.8.0; extra == "dev"
96
+ Requires-Dist: types-PyYAML; extra == "dev"
97
+ Requires-Dist: types-requests; extra == "dev"
98
+ Requires-Dist: types-simplejson; extra == "dev"
99
+ Requires-Dist: types-toml; extra == "dev"
100
+ Requires-Dist: types-tqdm; extra == "dev"
101
+ Requires-Dist: types-urllib3; extra == "dev"
102
+ Requires-Dist: mypy<1.15.0,>=1.14.1; python_version == "3.8" and extra == "dev"
103
+ Requires-Dist: mypy==1.15.0; python_version >= "3.9" and extra == "dev"
104
+ Provides-Extra: fastai
105
+ Requires-Dist: toml; extra == "fastai"
106
+ Requires-Dist: fastai>=2.4; extra == "fastai"
107
+ Requires-Dist: fastcore>=1.3.27; extra == "fastai"
108
+ Provides-Extra: hf_transfer
109
+ Requires-Dist: hf-transfer>=0.1.4; extra == "hf-transfer"
110
+ Provides-Extra: hf_xet
111
+ Requires-Dist: hf-xet<2.0.0,>=1.1.2; extra == "hf-xet"
112
+ Provides-Extra: inference
113
+ Requires-Dist: aiohttp; extra == "inference"
114
+ Provides-Extra: mcp
115
+ Requires-Dist: mcp>=1.8.0; extra == "mcp"
116
+ Requires-Dist: typer; extra == "mcp"
117
+ Requires-Dist: aiohttp; extra == "mcp"
118
+ Provides-Extra: oauth
119
+ Requires-Dist: authlib>=1.3.2; extra == "oauth"
120
+ Requires-Dist: fastapi; extra == "oauth"
121
+ Requires-Dist: httpx; extra == "oauth"
122
+ Requires-Dist: itsdangerous; extra == "oauth"
123
+ Provides-Extra: quality
124
+ Requires-Dist: ruff>=0.9.0; extra == "quality"
125
+ Requires-Dist: libcst>=1.4.0; extra == "quality"
126
+ Requires-Dist: mypy<1.15.0,>=1.14.1; python_version == "3.8" and extra == "quality"
127
+ Requires-Dist: mypy==1.15.0; python_version >= "3.9" and extra == "quality"
128
+ Provides-Extra: tensorflow
129
+ Requires-Dist: tensorflow; extra == "tensorflow"
130
+ Requires-Dist: pydot; extra == "tensorflow"
131
+ Requires-Dist: graphviz; extra == "tensorflow"
132
+ Provides-Extra: tensorflow-testing
133
+ Requires-Dist: tensorflow; extra == "tensorflow-testing"
134
+ Requires-Dist: keras<3.0; extra == "tensorflow-testing"
135
+ Provides-Extra: testing
136
+ Requires-Dist: InquirerPy==0.3.4; extra == "testing"
137
+ Requires-Dist: aiohttp; extra == "testing"
138
+ Requires-Dist: authlib>=1.3.2; extra == "testing"
139
+ Requires-Dist: fastapi; extra == "testing"
140
+ Requires-Dist: httpx; extra == "testing"
141
+ Requires-Dist: itsdangerous; extra == "testing"
142
+ Requires-Dist: jedi; extra == "testing"
143
+ Requires-Dist: Jinja2; extra == "testing"
144
+ Requires-Dist: pytest<8.2.2,>=8.1.1; extra == "testing"
145
+ Requires-Dist: pytest-cov; extra == "testing"
146
+ Requires-Dist: pytest-env; extra == "testing"
147
+ Requires-Dist: pytest-xdist; extra == "testing"
148
+ Requires-Dist: pytest-vcr; extra == "testing"
149
+ Requires-Dist: pytest-asyncio; extra == "testing"
150
+ Requires-Dist: pytest-rerunfailures; extra == "testing"
151
+ Requires-Dist: pytest-mock; extra == "testing"
152
+ Requires-Dist: urllib3<2.0; extra == "testing"
153
+ Requires-Dist: soundfile; extra == "testing"
154
+ Requires-Dist: Pillow; extra == "testing"
155
+ Requires-Dist: gradio>=4.0.0; extra == "testing"
156
+ Requires-Dist: numpy; extra == "testing"
157
+ Provides-Extra: torch
158
+ Requires-Dist: torch; extra == "torch"
159
+ Requires-Dist: safetensors[torch]; extra == "torch"
160
+ Provides-Extra: typing
161
+ Requires-Dist: typing-extensions>=4.8.0; extra == "typing"
162
+ Requires-Dist: types-PyYAML; extra == "typing"
163
+ Requires-Dist: types-requests; extra == "typing"
164
+ Requires-Dist: types-simplejson; extra == "typing"
165
+ Requires-Dist: types-toml; extra == "typing"
166
+ Requires-Dist: types-tqdm; extra == "typing"
167
+ Requires-Dist: types-urllib3; extra == "typing"
168
+
169
+ <p align="center">
170
+ <picture>
171
+ <source media="(prefers-color-scheme: dark)" srcset="https://huggingface.co/datasets/huggingface/documentation-images/raw/main/huggingface_hub-dark.svg">
172
+ <source media="(prefers-color-scheme: light)" srcset="https://huggingface.co/datasets/huggingface/documentation-images/raw/main/huggingface_hub.svg">
173
+ <img alt="huggingface_hub library logo" src="https://huggingface.co/datasets/huggingface/documentation-images/raw/main/huggingface_hub.svg" width="352" height="59" style="max-width: 100%;">
174
+ </picture>
175
+ <br/>
176
+ <br/>
177
+ </p>
178
+
179
+ <p align="center">
180
+ <i>The official Python client for the Huggingface Hub.</i>
181
+ </p>
182
+
183
+ <p align="center">
184
+ <a href="https://huggingface.co/docs/huggingface_hub/en/index"><img alt="Documentation" src="https://img.shields.io/website/http/huggingface.co/docs/huggingface_hub/index.svg?down_color=red&down_message=offline&up_message=online&label=doc"></a>
185
+ <a href="https://github.com/huggingface/huggingface_hub/releases"><img alt="GitHub release" src="https://img.shields.io/github/release/huggingface/huggingface_hub.svg"></a>
186
+ <a href="https://github.com/huggingface/huggingface_hub"><img alt="PyPi version" src="https://img.shields.io/pypi/pyversions/huggingface_hub.svg"></a>
187
+ <a href="https://pypi.org/project/huggingface-hub"><img alt="PyPI - Downloads" src="https://img.shields.io/pypi/dm/huggingface_hub"></a>
188
+ <a href="https://codecov.io/gh/huggingface/huggingface_hub"><img alt="Code coverage" src="https://codecov.io/gh/huggingface/huggingface_hub/branch/main/graph/badge.svg?token=RXP95LE2XL"></a>
189
+ </p>
190
+
191
+ <h4 align="center">
192
+ <p>
193
+ <b>English</b> |
194
+ <a href="https://github.com/huggingface/huggingface_hub/blob/main/i18n/README_de.md">Deutsch</a> |
195
+ <a href="https://github.com/huggingface/huggingface_hub/blob/main/i18n/README_hi.md">हिंदी</a> |
196
+ <a href="https://github.com/huggingface/huggingface_hub/blob/main/i18n/README_ko.md">한국어</a> |
197
+ <a href="https://github.com/huggingface/huggingface_hub/blob/main/i18n/README_cn.md">中文(简体)</a>
198
+ <p>
199
+ </h4>
200
+
201
+ ---
202
+
203
+ **Documentation**: <a href="https://hf.co/docs/huggingface_hub" target="_blank">https://hf.co/docs/huggingface_hub</a>
204
+
205
+ **Source Code**: <a href="https://github.com/huggingface/huggingface_hub" target="_blank">https://github.com/huggingface/huggingface_hub</a>
206
+
207
+ ---
208
+
209
+ ## Welcome to the huggingface_hub library
210
+
211
+ The `huggingface_hub` library allows you to interact with the [Hugging Face Hub](https://huggingface.co/), a platform democratizing open-source Machine Learning for creators and collaborators. Discover pre-trained models and datasets for your projects or play with the thousands of machine learning apps hosted on the Hub. You can also create and share your own models, datasets and demos with the community. The `huggingface_hub` library provides a simple way to do all these things with Python.
212
+
213
+ ## Key features
214
+
215
+ - [Download files](https://huggingface.co/docs/huggingface_hub/en/guides/download) from the Hub.
216
+ - [Upload files](https://huggingface.co/docs/huggingface_hub/en/guides/upload) to the Hub.
217
+ - [Manage your repositories](https://huggingface.co/docs/huggingface_hub/en/guides/repository).
218
+ - [Run Inference](https://huggingface.co/docs/huggingface_hub/en/guides/inference) on deployed models.
219
+ - [Search](https://huggingface.co/docs/huggingface_hub/en/guides/search) for models, datasets and Spaces.
220
+ - [Share Model Cards](https://huggingface.co/docs/huggingface_hub/en/guides/model-cards) to document your models.
221
+ - [Engage with the community](https://huggingface.co/docs/huggingface_hub/en/guides/community) through PRs and comments.
222
+
223
+ ## Installation
224
+
225
+ Install the `huggingface_hub` package with [pip](https://pypi.org/project/huggingface-hub/):
226
+
227
+ ```bash
228
+ pip install huggingface_hub
229
+ ```
230
+
231
+ If you prefer, you can also install it with [conda](https://huggingface.co/docs/huggingface_hub/en/installation#install-with-conda).
232
+
233
+ In order to keep the package minimal by default, `huggingface_hub` comes with optional dependencies useful for some use cases. For example, if you want have a complete experience for Inference, run:
234
+
235
+ ```bash
236
+ pip install huggingface_hub[inference]
237
+ ```
238
+
239
+ To learn more installation and optional dependencies, check out the [installation guide](https://huggingface.co/docs/huggingface_hub/en/installation).
240
+
241
+ ## Quick start
242
+
243
+ ### Download files
244
+
245
+ Download a single file
246
+
247
+ ```py
248
+ from huggingface_hub import hf_hub_download
249
+
250
+ hf_hub_download(repo_id="tiiuae/falcon-7b-instruct", filename="config.json")
251
+ ```
252
+
253
+ Or an entire repository
254
+
255
+ ```py
256
+ from huggingface_hub import snapshot_download
257
+
258
+ snapshot_download("stabilityai/stable-diffusion-2-1")
259
+ ```
260
+
261
+ Files will be downloaded in a local cache folder. More details in [this guide](https://huggingface.co/docs/huggingface_hub/en/guides/manage-cache).
262
+
263
+ ### Login
264
+
265
+ The Hugging Face Hub uses tokens to authenticate applications (see [docs](https://huggingface.co/docs/hub/security-tokens)). To log in your machine, run the following CLI:
266
+
267
+ ```bash
268
+ hf auth login
269
+ # or using an environment variable
270
+ hf auth login --token $HUGGINGFACE_TOKEN
271
+ ```
272
+
273
+ ### Create a repository
274
+
275
+ ```py
276
+ from huggingface_hub import create_repo
277
+
278
+ create_repo(repo_id="super-cool-model")
279
+ ```
280
+
281
+ ### Upload files
282
+
283
+ Upload a single file
284
+
285
+ ```py
286
+ from huggingface_hub import upload_file
287
+
288
+ upload_file(
289
+ path_or_fileobj="/home/lysandre/dummy-test/README.md",
290
+ path_in_repo="README.md",
291
+ repo_id="lysandre/test-model",
292
+ )
293
+ ```
294
+
295
+ Or an entire folder
296
+
297
+ ```py
298
+ from huggingface_hub import upload_folder
299
+
300
+ upload_folder(
301
+ folder_path="/path/to/local/space",
302
+ repo_id="username/my-cool-space",
303
+ repo_type="space",
304
+ )
305
+ ```
306
+
307
+ For details in the [upload guide](https://huggingface.co/docs/huggingface_hub/en/guides/upload).
308
+
309
+ ## Integrating to the Hub.
310
+
311
+ We're partnering with cool open source ML libraries to provide free model hosting and versioning. You can find the existing integrations [here](https://huggingface.co/docs/hub/libraries).
312
+
313
+ The advantages are:
314
+
315
+ - Free model or dataset hosting for libraries and their users.
316
+ - Built-in file versioning, even with very large files, thanks to a git-based approach.
317
+ - In-browser widgets to play with the uploaded models.
318
+ - Anyone can upload a new model for your library, they just need to add the corresponding tag for the model to be discoverable.
319
+ - Fast downloads! We use Cloudfront (a CDN) to geo-replicate downloads so they're blazing fast from anywhere on the globe.
320
+ - Usage stats and more features to come.
321
+
322
+ If you would like to integrate your library, feel free to open an issue to begin the discussion. We wrote a [step-by-step guide](https://huggingface.co/docs/hub/adding-a-library) with ❤️ showing how to do this integration.
323
+
324
+ ## Contributions (feature requests, bugs, etc.) are super welcome 💙💚💛💜🧡❤️
325
+
326
+ Everyone is welcome to contribute, and we value everybody's contribution. Code is not the only way to help the community.
327
+ Answering questions, helping others, reaching out and improving the documentations are immensely valuable to the community.
328
+ We wrote a [contribution guide](https://github.com/huggingface/huggingface_hub/blob/main/CONTRIBUTING.md) to summarize
329
+ how to get started to contribute to this repository.
330
+
331
+
pythonProject/.venv/Lib/site-packages/huggingface_hub-0.34.4.dist-info/RECORD ADDED
@@ -0,0 +1,327 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ../../Scripts/hf.exe,sha256=StSqlEsqGFWFge63Op8-BJoAHylmTBIDpSpvw0PHNhM,108411
2
+ ../../Scripts/huggingface-cli.exe,sha256=j98QwPmWkWMT3e4FJeVmL4Q23uJQPEkPlL6wJhvNaqI,108429
3
+ ../../Scripts/tiny-agents.exe,sha256=z7vLUru4du9IUGHzHWLsdzpRcm0wVAmgFHqra06vKp8,108421
4
+ huggingface_hub-0.34.4.dist-info/INSTALLER,sha256=zuuue4knoyJ-UwPPXg8fezS7VCrXJQrAP7zeNuwvFQg,4
5
+ huggingface_hub-0.34.4.dist-info/LICENSE,sha256=xx0jnfkXJvxRnG63LTGOxlggYnIysveWIZ6H3PNdCrQ,11357
6
+ huggingface_hub-0.34.4.dist-info/METADATA,sha256=sEy_vJu9uxm2AvxvoG49eqdJxzxDYRZPOpeBr4jKhkM,14699
7
+ huggingface_hub-0.34.4.dist-info/RECORD,,
8
+ huggingface_hub-0.34.4.dist-info/WHEEL,sha256=tZoeGjtWxWRfdplE7E3d45VPlLNQnvbKiYnx7gwAy8A,92
9
+ huggingface_hub-0.34.4.dist-info/entry_points.txt,sha256=HIzLhjwPTO7U_ncpW4AkmzAuaadr1ajmYagW5mdb5TM,217
10
+ huggingface_hub-0.34.4.dist-info/top_level.txt,sha256=8KzlQJAY4miUvjAssOAJodqKOw3harNzuiwGQ9qLSSk,16
11
+ huggingface_hub/__init__.py,sha256=MsUqTc5DczrwOwtEhkM9Y7tkGAlHvDtNTDbWQcEEqTA,51837
12
+ huggingface_hub/__pycache__/__init__.cpython-310.pyc,,
13
+ huggingface_hub/__pycache__/_commit_api.cpython-310.pyc,,
14
+ huggingface_hub/__pycache__/_commit_scheduler.cpython-310.pyc,,
15
+ huggingface_hub/__pycache__/_inference_endpoints.cpython-310.pyc,,
16
+ huggingface_hub/__pycache__/_jobs_api.cpython-310.pyc,,
17
+ huggingface_hub/__pycache__/_local_folder.cpython-310.pyc,,
18
+ huggingface_hub/__pycache__/_login.cpython-310.pyc,,
19
+ huggingface_hub/__pycache__/_oauth.cpython-310.pyc,,
20
+ huggingface_hub/__pycache__/_snapshot_download.cpython-310.pyc,,
21
+ huggingface_hub/__pycache__/_space_api.cpython-310.pyc,,
22
+ huggingface_hub/__pycache__/_tensorboard_logger.cpython-310.pyc,,
23
+ huggingface_hub/__pycache__/_upload_large_folder.cpython-310.pyc,,
24
+ huggingface_hub/__pycache__/_webhooks_payload.cpython-310.pyc,,
25
+ huggingface_hub/__pycache__/_webhooks_server.cpython-310.pyc,,
26
+ huggingface_hub/__pycache__/community.cpython-310.pyc,,
27
+ huggingface_hub/__pycache__/constants.cpython-310.pyc,,
28
+ huggingface_hub/__pycache__/dataclasses.cpython-310.pyc,,
29
+ huggingface_hub/__pycache__/errors.cpython-310.pyc,,
30
+ huggingface_hub/__pycache__/fastai_utils.cpython-310.pyc,,
31
+ huggingface_hub/__pycache__/file_download.cpython-310.pyc,,
32
+ huggingface_hub/__pycache__/hf_api.cpython-310.pyc,,
33
+ huggingface_hub/__pycache__/hf_file_system.cpython-310.pyc,,
34
+ huggingface_hub/__pycache__/hub_mixin.cpython-310.pyc,,
35
+ huggingface_hub/__pycache__/inference_api.cpython-310.pyc,,
36
+ huggingface_hub/__pycache__/keras_mixin.cpython-310.pyc,,
37
+ huggingface_hub/__pycache__/lfs.cpython-310.pyc,,
38
+ huggingface_hub/__pycache__/repocard.cpython-310.pyc,,
39
+ huggingface_hub/__pycache__/repocard_data.cpython-310.pyc,,
40
+ huggingface_hub/__pycache__/repository.cpython-310.pyc,,
41
+ huggingface_hub/_commit_api.py,sha256=68HxFnJE2s-QmGZRHQav5kOMTseYV_ZQi04ADaQmZUk,38979
42
+ huggingface_hub/_commit_scheduler.py,sha256=tfIoO1xWHjTJ6qy6VS6HIoymDycFPg0d6pBSZprrU2U,14679
43
+ huggingface_hub/_inference_endpoints.py,sha256=ahmbPcEXsJ_JcMb9TDgdkD8Z2z9uytkFG3_1o6dTm8g,17598
44
+ huggingface_hub/_jobs_api.py,sha256=Fm59WJ-P3E5mkCthQWgGugTlbXnQe4nxDKIwZ_jtpt8,5344
45
+ huggingface_hub/_local_folder.py,sha256=2iHXNgIT3UdSt2PvCovd0NzgVxTRypKb-rvAFLK-gZU,17305
46
+ huggingface_hub/_login.py,sha256=rcwx9EZdFUB3vuowC5QBiSYS4ImUnBzo04igR1Z8l40,20256
47
+ huggingface_hub/_oauth.py,sha256=75ya9toHxC0WRKsLOAI212CrssRjTSxs16mHWWNMb3w,18714
48
+ huggingface_hub/_snapshot_download.py,sha256=b-NzYQcvktsAirIfGQKgzQwu8w0S6lhBTvnJ5S6saw8,16166
49
+ huggingface_hub/_space_api.py,sha256=jb6rF8qLtjaNU12D-8ygAPM26xDiHCu8CHXHowhGTmg,5470
50
+ huggingface_hub/_tensorboard_logger.py,sha256=ZkYcAUiRC8RGL214QUYtp58O8G5tn-HF6DCWha9imcA,8358
51
+ huggingface_hub/_upload_large_folder.py,sha256=l2YWLZttOw69EGdihT3y_Nhr5mweLGooZG9L8smNoHY,30066
52
+ huggingface_hub/_webhooks_payload.py,sha256=Xm3KaK7tCOGBlXkuZvbym6zjHXrT1XCrbUFWuXiBmNY,3617
53
+ huggingface_hub/_webhooks_server.py,sha256=5J63wk9MUGKBNJVsOD9i60mJ-VMp0YYmlf87vQsl-L8,15767
54
+ huggingface_hub/cli/__init__.py,sha256=xzX1qgAvrtAX4gP59WrPlvOZFLuzuTgcjvanQvcpgHc,928
55
+ huggingface_hub/cli/__pycache__/__init__.cpython-310.pyc,,
56
+ huggingface_hub/cli/__pycache__/_cli_utils.cpython-310.pyc,,
57
+ huggingface_hub/cli/__pycache__/auth.cpython-310.pyc,,
58
+ huggingface_hub/cli/__pycache__/cache.cpython-310.pyc,,
59
+ huggingface_hub/cli/__pycache__/download.cpython-310.pyc,,
60
+ huggingface_hub/cli/__pycache__/hf.cpython-310.pyc,,
61
+ huggingface_hub/cli/__pycache__/jobs.cpython-310.pyc,,
62
+ huggingface_hub/cli/__pycache__/lfs.cpython-310.pyc,,
63
+ huggingface_hub/cli/__pycache__/repo.cpython-310.pyc,,
64
+ huggingface_hub/cli/__pycache__/repo_files.cpython-310.pyc,,
65
+ huggingface_hub/cli/__pycache__/system.cpython-310.pyc,,
66
+ huggingface_hub/cli/__pycache__/upload.cpython-310.pyc,,
67
+ huggingface_hub/cli/__pycache__/upload_large_folder.cpython-310.pyc,,
68
+ huggingface_hub/cli/_cli_utils.py,sha256=Nt6CjbkYqQQRuh70bUXVA6rZpbZt_Sa1WqBUxjQLu6g,2095
69
+ huggingface_hub/cli/auth.py,sha256=BO6sJJcdHhjouMEH5JpUmC0qg3vaukX4I5DtA7ohLes,7296
70
+ huggingface_hub/cli/cache.py,sha256=zPLEWP5MidLElo0FeLvkUjCT0EYlX1pmC-2TkCa8kD0,15871
71
+ huggingface_hub/cli/download.py,sha256=PUpW-nbu6ZAP6P9DpVhliAKSSlxvXWkVh0U2KZoukhQ,7115
72
+ huggingface_hub/cli/hf.py,sha256=SQ73_SXEQnWVJkhKT_6bwNQBHQXGOdI5qqlTTtI0XH0,2328
73
+ huggingface_hub/cli/jobs.py,sha256=QPF8YT31hoyjXXp_pD-LK_VHNFXF09su-nLeP4b9WGQ,21538
74
+ huggingface_hub/cli/lfs.py,sha256=J9MkKOGUW6GjBrKs2zZUCOaAGxpatxsEoSbBjuhDJV8,7230
75
+ huggingface_hub/cli/repo.py,sha256=lNDEZbOpLW8SQVBYDQ1xofw9nJ7M8AUsd2kBIV_m_do,10576
76
+ huggingface_hub/cli/repo_files.py,sha256=L-Ku52l2vZ04GCabp_OhVXqLzE9dsKQqaQKudGzjWg4,4831
77
+ huggingface_hub/cli/system.py,sha256=eLSYME7ywt5Ae3tYQnS43Tai2pR2JLtA1KGImzPt5pM,1707
78
+ huggingface_hub/cli/upload.py,sha256=qOGccIcBYJtodmlQlFyGfV_ZGHYhWlAARr3fxgYLDnE,14349
79
+ huggingface_hub/cli/upload_large_folder.py,sha256=dEb1EKbPi2nRpFS-gz4P0D49-Rlf26MAs6kgl7l2vJk,6151
80
+ huggingface_hub/commands/__init__.py,sha256=AkbM2a-iGh0Vq_xAWhK3mu3uZ44km8-X5uWjKcvcrUQ,928
81
+ huggingface_hub/commands/__pycache__/__init__.cpython-310.pyc,,
82
+ huggingface_hub/commands/__pycache__/_cli_utils.cpython-310.pyc,,
83
+ huggingface_hub/commands/__pycache__/delete_cache.cpython-310.pyc,,
84
+ huggingface_hub/commands/__pycache__/download.cpython-310.pyc,,
85
+ huggingface_hub/commands/__pycache__/env.cpython-310.pyc,,
86
+ huggingface_hub/commands/__pycache__/huggingface_cli.cpython-310.pyc,,
87
+ huggingface_hub/commands/__pycache__/lfs.cpython-310.pyc,,
88
+ huggingface_hub/commands/__pycache__/repo.cpython-310.pyc,,
89
+ huggingface_hub/commands/__pycache__/repo_files.cpython-310.pyc,,
90
+ huggingface_hub/commands/__pycache__/scan_cache.cpython-310.pyc,,
91
+ huggingface_hub/commands/__pycache__/tag.cpython-310.pyc,,
92
+ huggingface_hub/commands/__pycache__/upload.cpython-310.pyc,,
93
+ huggingface_hub/commands/__pycache__/upload_large_folder.cpython-310.pyc,,
94
+ huggingface_hub/commands/__pycache__/user.cpython-310.pyc,,
95
+ huggingface_hub/commands/__pycache__/version.cpython-310.pyc,,
96
+ huggingface_hub/commands/_cli_utils.py,sha256=ePYTIEWnU677nPvdNC5AdYcEB1400L6qYEUxMkVUzME,2329
97
+ huggingface_hub/commands/delete_cache.py,sha256=GpeDKs2I16rfxGnGX1widy9TBZSZM8PQ1qmHHRcwd-o,17734
98
+ huggingface_hub/commands/download.py,sha256=0QY9ho7eiAPvFndBPttGtH6vXNk3r9AioltNwc8h1Z4,8310
99
+ huggingface_hub/commands/env.py,sha256=qv4SmjuzUz9exo4RDMY2HqabLCKE1oRb55cBA6LN9R4,1342
100
+ huggingface_hub/commands/huggingface_cli.py,sha256=gDi7JueyiLD0bGclTEYfHPQWpAY_WBdPfHT7vkqa5v0,2654
101
+ huggingface_hub/commands/lfs.py,sha256=xdbnNRO04UuQemEhUGT809jFgQn9Rj-SnyT_0Ph-VYg,7342
102
+ huggingface_hub/commands/repo.py,sha256=WcRDFqUYKB0Kz0zFopegiG614ot6VOYTAf6jht0BMss,6042
103
+ huggingface_hub/commands/repo_files.py,sha256=ftjLCC3XCY-AMmiYiZPIdRMmIqZbqVZw-BSjBLcZup4,5054
104
+ huggingface_hub/commands/scan_cache.py,sha256=Er5y0paRv0feXnwxe-eAwuJA6xlZ-OdEQ1FbB80BqNU,8670
105
+ huggingface_hub/commands/tag.py,sha256=4fgQuXJHG59lTVyOjIUZjxdJDL4JZW4q10XDPSo-gss,6382
106
+ huggingface_hub/commands/upload.py,sha256=eAJIig4ljtO9FRyGjiz6HbHS-Q4MOQziRgzjQrl5Koo,14576
107
+ huggingface_hub/commands/upload_large_folder.py,sha256=_1id84BFtbL8HgFRKZ-el_uPrijamz1qWlzO16KbUAc,6254
108
+ huggingface_hub/commands/user.py,sha256=MjG1lwMq1p5QAlBolFnRX_pUxE3Kd3UiPl-nEEQSgXg,7537
109
+ huggingface_hub/commands/version.py,sha256=rGpCbvxImY9eQqXrshYt609Iws27R75WARmKQrIo6Ok,1390
110
+ huggingface_hub/community.py,sha256=4MtcoxEI9_0lmmilBEnvUEi8_O1Ivfa8p6eKxYU5-ts,12198
111
+ huggingface_hub/constants.py,sha256=nILseAp4rqLu_KQTZDpPGOhepVAPanD7azbomAvovj0,10313
112
+ huggingface_hub/dataclasses.py,sha256=sgPdEi2UDprhNPP2PPkiSlzsHdC1WcpwVTLwlHAEcr0,17224
113
+ huggingface_hub/errors.py,sha256=D7Lw0Jjrf8vfmD0B26LEvg-JWkU8Zq0KDPJOzFY4QLw,11201
114
+ huggingface_hub/fastai_utils.py,sha256=DpeH9d-6ut2k_nCAAwglM51XmRmgfbRe2SPifpVL5Yk,16745
115
+ huggingface_hub/file_download.py,sha256=E-NWON01pprbAsw7Kz477JX6f8HTWsdpEdQAtA37t5c,78974
116
+ huggingface_hub/hf_api.py,sha256=a12ywN6Lc43A12B36xxK4dJzXsgtmjQTZqYAC4gjTzY,465077
117
+ huggingface_hub/hf_file_system.py,sha256=nrNOoNHRwf1swebtQvZExSblRjQg9rHKxL7Cslk72uw,46899
118
+ huggingface_hub/hub_mixin.py,sha256=MArtbUxjXiYeOvOmNBG9I_j5t02m2xCVAcge4waip1w,38091
119
+ huggingface_hub/inference/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
120
+ huggingface_hub/inference/__pycache__/__init__.cpython-310.pyc,,
121
+ huggingface_hub/inference/__pycache__/_client.cpython-310.pyc,,
122
+ huggingface_hub/inference/__pycache__/_common.cpython-310.pyc,,
123
+ huggingface_hub/inference/_client.py,sha256=ggY3ebg_xjC9zuHUTpzG_rghFr65vDt7Btg3pATfJjE,165039
124
+ huggingface_hub/inference/_common.py,sha256=6qAIauugyl1eHk0FhWdjBNEXBNF33_VXC8lc1GR8t7s,15874
125
+ huggingface_hub/inference/_generated/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
126
+ huggingface_hub/inference/_generated/__pycache__/__init__.cpython-310.pyc,,
127
+ huggingface_hub/inference/_generated/__pycache__/_async_client.cpython-310.pyc,,
128
+ huggingface_hub/inference/_generated/_async_client.py,sha256=oklZPp7TBmCkUB5B3UpuivEXQWGgUxXJwA-Zs4YFDsM,171274
129
+ huggingface_hub/inference/_generated/types/__init__.py,sha256=9WvrGQ8aThtKSNzZF06j-CIE2ZuItne8FFnea1p1u38,6557
130
+ huggingface_hub/inference/_generated/types/__pycache__/__init__.cpython-310.pyc,,
131
+ huggingface_hub/inference/_generated/types/__pycache__/audio_classification.cpython-310.pyc,,
132
+ huggingface_hub/inference/_generated/types/__pycache__/audio_to_audio.cpython-310.pyc,,
133
+ huggingface_hub/inference/_generated/types/__pycache__/automatic_speech_recognition.cpython-310.pyc,,
134
+ huggingface_hub/inference/_generated/types/__pycache__/base.cpython-310.pyc,,
135
+ huggingface_hub/inference/_generated/types/__pycache__/chat_completion.cpython-310.pyc,,
136
+ huggingface_hub/inference/_generated/types/__pycache__/depth_estimation.cpython-310.pyc,,
137
+ huggingface_hub/inference/_generated/types/__pycache__/document_question_answering.cpython-310.pyc,,
138
+ huggingface_hub/inference/_generated/types/__pycache__/feature_extraction.cpython-310.pyc,,
139
+ huggingface_hub/inference/_generated/types/__pycache__/fill_mask.cpython-310.pyc,,
140
+ huggingface_hub/inference/_generated/types/__pycache__/image_classification.cpython-310.pyc,,
141
+ huggingface_hub/inference/_generated/types/__pycache__/image_segmentation.cpython-310.pyc,,
142
+ huggingface_hub/inference/_generated/types/__pycache__/image_to_image.cpython-310.pyc,,
143
+ huggingface_hub/inference/_generated/types/__pycache__/image_to_text.cpython-310.pyc,,
144
+ huggingface_hub/inference/_generated/types/__pycache__/image_to_video.cpython-310.pyc,,
145
+ huggingface_hub/inference/_generated/types/__pycache__/object_detection.cpython-310.pyc,,
146
+ huggingface_hub/inference/_generated/types/__pycache__/question_answering.cpython-310.pyc,,
147
+ huggingface_hub/inference/_generated/types/__pycache__/sentence_similarity.cpython-310.pyc,,
148
+ huggingface_hub/inference/_generated/types/__pycache__/summarization.cpython-310.pyc,,
149
+ huggingface_hub/inference/_generated/types/__pycache__/table_question_answering.cpython-310.pyc,,
150
+ huggingface_hub/inference/_generated/types/__pycache__/text2text_generation.cpython-310.pyc,,
151
+ huggingface_hub/inference/_generated/types/__pycache__/text_classification.cpython-310.pyc,,
152
+ huggingface_hub/inference/_generated/types/__pycache__/text_generation.cpython-310.pyc,,
153
+ huggingface_hub/inference/_generated/types/__pycache__/text_to_audio.cpython-310.pyc,,
154
+ huggingface_hub/inference/_generated/types/__pycache__/text_to_image.cpython-310.pyc,,
155
+ huggingface_hub/inference/_generated/types/__pycache__/text_to_speech.cpython-310.pyc,,
156
+ huggingface_hub/inference/_generated/types/__pycache__/text_to_video.cpython-310.pyc,,
157
+ huggingface_hub/inference/_generated/types/__pycache__/token_classification.cpython-310.pyc,,
158
+ huggingface_hub/inference/_generated/types/__pycache__/translation.cpython-310.pyc,,
159
+ huggingface_hub/inference/_generated/types/__pycache__/video_classification.cpython-310.pyc,,
160
+ huggingface_hub/inference/_generated/types/__pycache__/visual_question_answering.cpython-310.pyc,,
161
+ huggingface_hub/inference/_generated/types/__pycache__/zero_shot_classification.cpython-310.pyc,,
162
+ huggingface_hub/inference/_generated/types/__pycache__/zero_shot_image_classification.cpython-310.pyc,,
163
+ huggingface_hub/inference/_generated/types/__pycache__/zero_shot_object_detection.cpython-310.pyc,,
164
+ huggingface_hub/inference/_generated/types/audio_classification.py,sha256=Jg3mzfGhCSH6CfvVvgJSiFpkz6v4nNA0G4LJXacEgNc,1573
165
+ huggingface_hub/inference/_generated/types/audio_to_audio.py,sha256=2Ep4WkePL7oJwcp5nRJqApwviumGHbft9HhXE9XLHj4,891
166
+ huggingface_hub/inference/_generated/types/automatic_speech_recognition.py,sha256=8CEphr6rvRHgq1L5Md3tq14V0tEAmzJkemh1_7gSswo,5515
167
+ huggingface_hub/inference/_generated/types/base.py,sha256=4XG49q0-2SOftYQ8HXQnWLxiJktou-a7IoG3kdOv-kg,6751
168
+ huggingface_hub/inference/_generated/types/chat_completion.py,sha256=6EqWnpbeT0zsiLNZjoJDzrmZ34M2j01S99oMMayZg9Y,11182
169
+ huggingface_hub/inference/_generated/types/depth_estimation.py,sha256=rcpe9MhYMeLjflOwBs3KMZPr6WjOH3FYEThStG-FJ3M,929
170
+ huggingface_hub/inference/_generated/types/document_question_answering.py,sha256=6BEYGwJcqGlah4RBJDAvWFTEXkO0mosBiMy82432nAM,3202
171
+ huggingface_hub/inference/_generated/types/feature_extraction.py,sha256=NMWVL_TLSG5SS5bdt1-fflkZ75UMlMKeTMtmdnUTADc,1537
172
+ huggingface_hub/inference/_generated/types/fill_mask.py,sha256=OrTgQ7Ndn0_dWK5thQhZwTOHbQni8j0iJcx9llyhRds,1708
173
+ huggingface_hub/inference/_generated/types/image_classification.py,sha256=A-Y024o8723_n8mGVos4TwdAkVL62McGeL1iIo4VzNs,1585
174
+ huggingface_hub/inference/_generated/types/image_segmentation.py,sha256=vrkI4SuP1Iq_iLXc-2pQhYY3SHN4gzvFBoZqbUHxU7o,1950
175
+ huggingface_hub/inference/_generated/types/image_to_image.py,sha256=HPz1uKXk_9xvgNUi3GV6n4lw-J3G6cdGTcW3Ou_N0l8,2044
176
+ huggingface_hub/inference/_generated/types/image_to_text.py,sha256=OaFEBAfgT-fOVzJ7xVermGf7VODhrc9-Jg38WrM7-2o,4810
177
+ huggingface_hub/inference/_generated/types/image_to_video.py,sha256=bC-L_cNsDhk4s_IdSiprJ9d1NeMGePLcUp7UPpco21w,2240
178
+ huggingface_hub/inference/_generated/types/object_detection.py,sha256=VuFlb1281qTXoSgJDmquGz-VNfEZLo2H0Rh_F6MF6ts,2000
179
+ huggingface_hub/inference/_generated/types/question_answering.py,sha256=zw38a9_9l2k1ifYZefjkioqZ4asfSRM9M4nU3gSCmAQ,2898
180
+ huggingface_hub/inference/_generated/types/sentence_similarity.py,sha256=w5Nj1g18eBzopZwxuDLI-fEsyaCK2KrHA5yf_XfSjgo,1052
181
+ huggingface_hub/inference/_generated/types/summarization.py,sha256=WGGr8uDLrZg8JQgF9ZMUP9euw6uZo6zwkVZ-IfvCFI0,1487
182
+ huggingface_hub/inference/_generated/types/table_question_answering.py,sha256=cJnIPA2fIbQP2Ejn7X_esY48qGWoXg30fnNOqCXiOVQ,2293
183
+ huggingface_hub/inference/_generated/types/text2text_generation.py,sha256=v-418w1JNNSZ2tuW9DUl6a36TQQCADa438A3ufvcbOw,1609
184
+ huggingface_hub/inference/_generated/types/text_classification.py,sha256=FarAjygLEfPofLfKeabzJ7PKEBItlHGoUNUOzyLRpL4,1445
185
+ huggingface_hub/inference/_generated/types/text_generation.py,sha256=28u-1zU7elk2teP3y4u1VAtDDHzY0JZ2KEEJe5d5uvg,5922
186
+ huggingface_hub/inference/_generated/types/text_to_audio.py,sha256=1HR9Q6s9MXqtKGTvHPLGVMum5-eg7O-Pgv6Nd0v8_HU,4741
187
+ huggingface_hub/inference/_generated/types/text_to_image.py,sha256=sGGi1Fa0n5Pmd6G3I-F2SBJcJ1M7Gmqnng6sfi0AVzs,1903
188
+ huggingface_hub/inference/_generated/types/text_to_speech.py,sha256=ROFuR32ijROCeqbv81Jos0lmaA8SRWyIUsWrdD4yWow,4760
189
+ huggingface_hub/inference/_generated/types/text_to_video.py,sha256=yHXVNs3t6aYO7visrBlB5cH7kjoysxF9510aofcf_18,1790
190
+ huggingface_hub/inference/_generated/types/token_classification.py,sha256=iblAcgfxXeaLYJ14NdiiCMIQuBlarUknLkXUklhvcLI,1915
191
+ huggingface_hub/inference/_generated/types/translation.py,sha256=xww4X5cfCYv_F0oINWLwqJRPCT6SV3VBAJuPjTs_j7o,1763
192
+ huggingface_hub/inference/_generated/types/video_classification.py,sha256=TyydjQw2NRLK9sDGzJUVnkDeo848ebmCx588Ur8I9q0,1680
193
+ huggingface_hub/inference/_generated/types/visual_question_answering.py,sha256=AWrQ6qo4gZa3PGedaNpzDFqx5yOYyjhnUB6iuZEj_uo,1673
194
+ huggingface_hub/inference/_generated/types/zero_shot_classification.py,sha256=BAiebPjsqoNa8EU35Dx0pfIv8W2c4GSl-TJckV1MaxQ,1738
195
+ huggingface_hub/inference/_generated/types/zero_shot_image_classification.py,sha256=8J9n6VqFARkWvPfAZNWEG70AlrMGldU95EGQQwn06zI,1487
196
+ huggingface_hub/inference/_generated/types/zero_shot_object_detection.py,sha256=GUd81LIV7oEbRWayDlAVgyLmY596r1M3AW0jXDp1yTA,1630
197
+ huggingface_hub/inference/_mcp/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
198
+ huggingface_hub/inference/_mcp/__pycache__/__init__.cpython-310.pyc,,
199
+ huggingface_hub/inference/_mcp/__pycache__/_cli_hacks.cpython-310.pyc,,
200
+ huggingface_hub/inference/_mcp/__pycache__/agent.cpython-310.pyc,,
201
+ huggingface_hub/inference/_mcp/__pycache__/cli.cpython-310.pyc,,
202
+ huggingface_hub/inference/_mcp/__pycache__/constants.cpython-310.pyc,,
203
+ huggingface_hub/inference/_mcp/__pycache__/mcp_client.cpython-310.pyc,,
204
+ huggingface_hub/inference/_mcp/__pycache__/types.cpython-310.pyc,,
205
+ huggingface_hub/inference/_mcp/__pycache__/utils.cpython-310.pyc,,
206
+ huggingface_hub/inference/_mcp/_cli_hacks.py,sha256=cMZirVFe4N0EM9Nzzs9aEmzUBUEBYR4oYZpByTWlZCM,3182
207
+ huggingface_hub/inference/_mcp/agent.py,sha256=VahvSqldiC1R72CFH4T05l80uEXl5OjLwboWQFUJbsw,4281
208
+ huggingface_hub/inference/_mcp/cli.py,sha256=UY4K20zp5ydGx7Gzp4goVXVIDV2gB2arKFeGWK1AzwE,9625
209
+ huggingface_hub/inference/_mcp/constants.py,sha256=AnOp_oR5Vty0d5J3AynGGNK9i1I9KgGhCd9nPoZcD1M,2495
210
+ huggingface_hub/inference/_mcp/mcp_client.py,sha256=ndaTcZZPbU1ZTNUeB9-WdaOx7bHD3lsrXnKxCeiwpUg,15788
211
+ huggingface_hub/inference/_mcp/types.py,sha256=ic8VSR9JY1d-vPWsBVXYtXtIU669-HbGQ3m12Szs7BQ,815
212
+ huggingface_hub/inference/_mcp/utils.py,sha256=VsRWl0fuSZDS0zNT9n7FOMSlzA0UBbP8p8xWKWDt2Pc,4093
213
+ huggingface_hub/inference/_providers/__init__.py,sha256=8i8Y2X_hEhm20ndFjEQxgddHVAEQpYn38UfLeH4V0Rg,8354
214
+ huggingface_hub/inference/_providers/__pycache__/__init__.cpython-310.pyc,,
215
+ huggingface_hub/inference/_providers/__pycache__/_common.cpython-310.pyc,,
216
+ huggingface_hub/inference/_providers/__pycache__/black_forest_labs.cpython-310.pyc,,
217
+ huggingface_hub/inference/_providers/__pycache__/cerebras.cpython-310.pyc,,
218
+ huggingface_hub/inference/_providers/__pycache__/cohere.cpython-310.pyc,,
219
+ huggingface_hub/inference/_providers/__pycache__/fal_ai.cpython-310.pyc,,
220
+ huggingface_hub/inference/_providers/__pycache__/featherless_ai.cpython-310.pyc,,
221
+ huggingface_hub/inference/_providers/__pycache__/fireworks_ai.cpython-310.pyc,,
222
+ huggingface_hub/inference/_providers/__pycache__/groq.cpython-310.pyc,,
223
+ huggingface_hub/inference/_providers/__pycache__/hf_inference.cpython-310.pyc,,
224
+ huggingface_hub/inference/_providers/__pycache__/hyperbolic.cpython-310.pyc,,
225
+ huggingface_hub/inference/_providers/__pycache__/nebius.cpython-310.pyc,,
226
+ huggingface_hub/inference/_providers/__pycache__/novita.cpython-310.pyc,,
227
+ huggingface_hub/inference/_providers/__pycache__/nscale.cpython-310.pyc,,
228
+ huggingface_hub/inference/_providers/__pycache__/openai.cpython-310.pyc,,
229
+ huggingface_hub/inference/_providers/__pycache__/replicate.cpython-310.pyc,,
230
+ huggingface_hub/inference/_providers/__pycache__/sambanova.cpython-310.pyc,,
231
+ huggingface_hub/inference/_providers/__pycache__/together.cpython-310.pyc,,
232
+ huggingface_hub/inference/_providers/_common.py,sha256=dDtKWEUTcvO0llx96x5E3RwOSzkCIL8cKbiFk9RRh8o,11330
233
+ huggingface_hub/inference/_providers/black_forest_labs.py,sha256=wO7qgRyNyrIKlZtvL3vJEbS4-D19kfoXZk6PDh1dTis,2842
234
+ huggingface_hub/inference/_providers/cerebras.py,sha256=QOJ-1U-os7uE7p6eUnn_P_APq-yQhx28be7c3Tq2EuA,210
235
+ huggingface_hub/inference/_providers/cohere.py,sha256=O3tC-qIUL91mx_mE8bOHCtDWcQuKOUauhUoXSUBUCZ8,1253
236
+ huggingface_hub/inference/_providers/fal_ai.py,sha256=Z3-MPddtsfjWT1Zmvni0m8LPBSFjOa_wyv8xV9FNtoA,9860
237
+ huggingface_hub/inference/_providers/featherless_ai.py,sha256=QxBz-32O4PztxixrIjrfKuTOzvfqyUi-cVsw0Hf_zlY,1382
238
+ huggingface_hub/inference/_providers/fireworks_ai.py,sha256=Id226ITfPkOcFMFzly3MW9l-dZl9l4qizL4JEHWkBFk,1215
239
+ huggingface_hub/inference/_providers/groq.py,sha256=JTk2JV4ZOlaohho7zLAFQtk92kGVsPmLJ1hmzcwsqvQ,315
240
+ huggingface_hub/inference/_providers/hf_inference.py,sha256=P4FJdzk3M2T6GdSTumBvpcr8NNRixWLXymv4ptpHroI,9508
241
+ huggingface_hub/inference/_providers/hyperbolic.py,sha256=OQIBi2j3aNvuaSQ8BUK1K1PVeRXdrxc80G-6YmBa-ns,1985
242
+ huggingface_hub/inference/_providers/nebius.py,sha256=VJpTF2JZ58rznc9wxdk-57vwF8sV2vESw_WkXjXqCho,3580
243
+ huggingface_hub/inference/_providers/novita.py,sha256=HGVC8wPraRQUuI5uBoye1Y4Wqe4X116B71GhhbWy5yM,2514
244
+ huggingface_hub/inference/_providers/nscale.py,sha256=qWUsWinQmUbNUqehyKn34tVoWehu8gd-OZ2F4uj2SWM,1802
245
+ huggingface_hub/inference/_providers/openai.py,sha256=GCVYeNdjWIgpQQ7E_Xv8IebmdhTi0S6WfFosz3nLtps,1089
246
+ huggingface_hub/inference/_providers/replicate.py,sha256=BuLb1x4nUlH5SfazBwvMiFwwcs-OS99U87m3QWdx2is,3810
247
+ huggingface_hub/inference/_providers/sambanova.py,sha256=Unt3H3jr_kgI9vzRjmmW1DFyoEuPkKCcgIIloiOj3j8,2037
248
+ huggingface_hub/inference/_providers/together.py,sha256=KHF19CS3qXS7G1-CwcMiD8Z5wzPKEKi4F2DzqAthbBE,3439
249
+ huggingface_hub/inference_api.py,sha256=b4-NhPSn9b44nYKV8tDKXodmE4JVdEymMWL4CVGkzlE,8323
250
+ huggingface_hub/keras_mixin.py,sha256=WGNQZROdw6yjJ1DGTPZPwKAxf1UbkzAx1dRidkeb2fk,19553
251
+ huggingface_hub/lfs.py,sha256=n-TIjK7J7aXG3zi__0nkd6aNkE4djOf9CD6dYQOQ5P8,16649
252
+ huggingface_hub/py.typed,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
253
+ huggingface_hub/repocard.py,sha256=KxQoJTvnMVA6mKpi7660O6dKtbw4wRJMBxyXuP7vvzU,34751
254
+ huggingface_hub/repocard_data.py,sha256=hr4ReFpEQMNdh_9Dx-L-IJoI1ElHyk-h-8ZRqwVYYOE,34082
255
+ huggingface_hub/repository.py,sha256=Lerq3kr7tC-oUdZk5i1CdhAA84ZvYiqjaGR77j2iOyk,54536
256
+ huggingface_hub/serialization/__init__.py,sha256=kn-Fa-m4FzMnN8lNsF-SwFcfzug4CucexybGKyvZ8S0,1041
257
+ huggingface_hub/serialization/__pycache__/__init__.cpython-310.pyc,,
258
+ huggingface_hub/serialization/__pycache__/_base.cpython-310.pyc,,
259
+ huggingface_hub/serialization/__pycache__/_dduf.cpython-310.pyc,,
260
+ huggingface_hub/serialization/__pycache__/_tensorflow.cpython-310.pyc,,
261
+ huggingface_hub/serialization/__pycache__/_torch.cpython-310.pyc,,
262
+ huggingface_hub/serialization/_base.py,sha256=Df3GwGR9NzeK_SD75prXLucJAzPiNPgHbgXSw-_LTk8,8126
263
+ huggingface_hub/serialization/_dduf.py,sha256=s42239rLiHwaJE36QDEmS5GH7DSmQ__BffiHJO5RjIg,15424
264
+ huggingface_hub/serialization/_tensorflow.py,sha256=zHOvEMg-JHC55Fm4roDT3LUCDO5zB9qtXZffG065RAM,3625
265
+ huggingface_hub/serialization/_torch.py,sha256=jpBmuSZJymMpvLcDcMaNxDu_fE5VkY_pAVH8e8stYIo,45201
266
+ huggingface_hub/templates/datasetcard_template.md,sha256=W-EMqR6wndbrnZorkVv56URWPG49l7MATGeI015kTvs,5503
267
+ huggingface_hub/templates/modelcard_template.md,sha256=4AqArS3cqdtbit5Bo-DhjcnDFR-pza5hErLLTPM4Yuc,6870
268
+ huggingface_hub/utils/__init__.py,sha256=ORfVkn5D0wuLIq12jjhTzn5_c4F8fRPxB7TG-iednuQ,3722
269
+ huggingface_hub/utils/__pycache__/__init__.cpython-310.pyc,,
270
+ huggingface_hub/utils/__pycache__/_auth.cpython-310.pyc,,
271
+ huggingface_hub/utils/__pycache__/_cache_assets.cpython-310.pyc,,
272
+ huggingface_hub/utils/__pycache__/_cache_manager.cpython-310.pyc,,
273
+ huggingface_hub/utils/__pycache__/_chunk_utils.cpython-310.pyc,,
274
+ huggingface_hub/utils/__pycache__/_datetime.cpython-310.pyc,,
275
+ huggingface_hub/utils/__pycache__/_deprecation.cpython-310.pyc,,
276
+ huggingface_hub/utils/__pycache__/_dotenv.cpython-310.pyc,,
277
+ huggingface_hub/utils/__pycache__/_experimental.cpython-310.pyc,,
278
+ huggingface_hub/utils/__pycache__/_fixes.cpython-310.pyc,,
279
+ huggingface_hub/utils/__pycache__/_git_credential.cpython-310.pyc,,
280
+ huggingface_hub/utils/__pycache__/_headers.cpython-310.pyc,,
281
+ huggingface_hub/utils/__pycache__/_hf_folder.cpython-310.pyc,,
282
+ huggingface_hub/utils/__pycache__/_http.cpython-310.pyc,,
283
+ huggingface_hub/utils/__pycache__/_lfs.cpython-310.pyc,,
284
+ huggingface_hub/utils/__pycache__/_pagination.cpython-310.pyc,,
285
+ huggingface_hub/utils/__pycache__/_paths.cpython-310.pyc,,
286
+ huggingface_hub/utils/__pycache__/_runtime.cpython-310.pyc,,
287
+ huggingface_hub/utils/__pycache__/_safetensors.cpython-310.pyc,,
288
+ huggingface_hub/utils/__pycache__/_subprocess.cpython-310.pyc,,
289
+ huggingface_hub/utils/__pycache__/_telemetry.cpython-310.pyc,,
290
+ huggingface_hub/utils/__pycache__/_typing.cpython-310.pyc,,
291
+ huggingface_hub/utils/__pycache__/_validators.cpython-310.pyc,,
292
+ huggingface_hub/utils/__pycache__/_xet.cpython-310.pyc,,
293
+ huggingface_hub/utils/__pycache__/_xet_progress_reporting.cpython-310.pyc,,
294
+ huggingface_hub/utils/__pycache__/endpoint_helpers.cpython-310.pyc,,
295
+ huggingface_hub/utils/__pycache__/insecure_hashlib.cpython-310.pyc,,
296
+ huggingface_hub/utils/__pycache__/logging.cpython-310.pyc,,
297
+ huggingface_hub/utils/__pycache__/sha.cpython-310.pyc,,
298
+ huggingface_hub/utils/__pycache__/tqdm.cpython-310.pyc,,
299
+ huggingface_hub/utils/_auth.py,sha256=Ixve2vxdftHXXk2R2vfyLzlVoDT39Tkq-Hrou9KCUvw,8286
300
+ huggingface_hub/utils/_cache_assets.py,sha256=kai77HPQMfYpROouMBQCr_gdBCaeTm996Sqj0dExbNg,5728
301
+ huggingface_hub/utils/_cache_manager.py,sha256=osqV4gM5Mx293Kw1POuJ5uKIcoIG57gY0k6BU6vdjPA,34518
302
+ huggingface_hub/utils/_chunk_utils.py,sha256=kRCaj5228_vKcyLWspd8Xq01f17Jz6ds5Sr9ed5d_RU,2130
303
+ huggingface_hub/utils/_datetime.py,sha256=kCS5jaKV25kOncX1xujbXsz5iDLcjLcLw85semGNzxQ,2770
304
+ huggingface_hub/utils/_deprecation.py,sha256=HZhRGGUX_QMKBBBwHHlffLtmCSK01TOpeXHefZbPfwI,4872
305
+ huggingface_hub/utils/_dotenv.py,sha256=RzHqC8HgzVxE-N4DFBcnemvX0NHmXcV0My2ASK0U1OQ,2017
306
+ huggingface_hub/utils/_experimental.py,sha256=3-c8irbn9sJr2CwWbzhGkIrdXKg8_x7BifhHFy32ei8,2470
307
+ huggingface_hub/utils/_fixes.py,sha256=xQV1QkUn2WpLqLjtXNiyn9gh-454K6AF-Q3kwkYAQD8,4437
308
+ huggingface_hub/utils/_git_credential.py,sha256=SDdsiREr1TcAR2Ze2TB0E5cYzVJgvDZrs60od9lAsMc,4596
309
+ huggingface_hub/utils/_headers.py,sha256=w4ayq4hLGaZ3B7nwdEi5Zu23SmmDuOwv58It78wkakk,8868
310
+ huggingface_hub/utils/_hf_folder.py,sha256=WNjTnu0Q7tqcSS9EsP4ssCJrrJMcCvAt8P_-LEtmOU8,2487
311
+ huggingface_hub/utils/_http.py,sha256=her7UZ0KRo9WYDArpqVFyEXTusOGUECj5HNS8Eahqm8,25531
312
+ huggingface_hub/utils/_lfs.py,sha256=EC0Oz6Wiwl8foRNkUOzrETXzAWlbgpnpxo5a410ovFY,3957
313
+ huggingface_hub/utils/_pagination.py,sha256=EX5tRasSuQDaKbXuGYbInBK2odnSWNHgzw2tSgqeBRI,1906
314
+ huggingface_hub/utils/_paths.py,sha256=w1ZhFmmD5ykWjp_hAvhjtOoa2ZUcOXJrF4a6O3QpAWo,5042
315
+ huggingface_hub/utils/_runtime.py,sha256=L7SOYezdxKcwd4DovAY0UGY3qt27toXO-QjceIDwExk,11634
316
+ huggingface_hub/utils/_safetensors.py,sha256=GW3nyv7xQcuwObKYeYoT9VhURVzG1DZTbKBKho8Bbos,4458
317
+ huggingface_hub/utils/_subprocess.py,sha256=u9FFUDE7TrzQTiuEzlUnHx7S2P57GbYRV8u16GJwrFw,4625
318
+ huggingface_hub/utils/_telemetry.py,sha256=54LXeIJU5pEGghPAh06gqNAR-UoxOjVLvKqAQscwqZs,4890
319
+ huggingface_hub/utils/_typing.py,sha256=Dgp6TQUlpzStfVLoSvXHCBP4b3NzHZ8E0Gg9mYAoDS4,2903
320
+ huggingface_hub/utils/_validators.py,sha256=dDsVG31iooTYrIyi5Vwr1DukL0fEmJwu3ceVNduhsuE,9204
321
+ huggingface_hub/utils/_xet.py,sha256=f8qfk8YKePAeGUL6lQiQ1w_3bcs78oWwbeACYdUeg5k,7312
322
+ huggingface_hub/utils/_xet_progress_reporting.py,sha256=i1nOjoUtaUpbWV7Tkn5H9cEJZ0CduVz3Q6eIjNvyBhk,5450
323
+ huggingface_hub/utils/endpoint_helpers.py,sha256=9VtIAlxQ5H_4y30sjCAgbu7XCqAtNLC7aRYxaNn0hLI,2366
324
+ huggingface_hub/utils/insecure_hashlib.py,sha256=iAaepavFZ5Dhfa5n8KozRfQprKmvcjSnt3X58OUl9fQ,1142
325
+ huggingface_hub/utils/logging.py,sha256=0A8fF1yh3L9Ka_bCDX2ml4U5Ht0tY8Dr3JcbRvWFuwo,4909
326
+ huggingface_hub/utils/sha.py,sha256=OFnNGCba0sNcT2gUwaVCJnldxlltrHHe0DS_PCpV3C4,2134
327
+ huggingface_hub/utils/tqdm.py,sha256=xAKcyfnNHsZ7L09WuEM5Ew5-MDhiahLACbbN2zMmcLs,10671
pythonProject/.venv/Lib/site-packages/huggingface_hub-0.34.4.dist-info/WHEEL ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ Wheel-Version: 1.0
2
+ Generator: bdist_wheel (0.45.1)
3
+ Root-Is-Purelib: true
4
+ Tag: py3-none-any
5
+
pythonProject/.venv/Lib/site-packages/huggingface_hub-0.34.4.dist-info/entry_points.txt ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ [console_scripts]
2
+ hf = huggingface_hub.cli.hf:main
3
+ huggingface-cli = huggingface_hub.commands.huggingface_cli:main
4
+ tiny-agents = huggingface_hub.inference._mcp.cli:app
5
+
6
+ [fsspec.specs]
7
+ hf=huggingface_hub.HfFileSystem
8
+
pythonProject/.venv/Lib/site-packages/huggingface_hub-0.34.4.dist-info/top_level.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ huggingface_hub
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/__init__.cpython-310.pyc ADDED
Binary file (20.6 kB). View file
 
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/_commit_api.cpython-310.pyc ADDED
Binary file (30.1 kB). View file
 
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/_commit_scheduler.cpython-310.pyc ADDED
Binary file (13.8 kB). View file
 
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/_inference_endpoints.cpython-310.pyc ADDED
Binary file (15.8 kB). View file
 
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/_jobs_api.cpython-310.pyc ADDED
Binary file (5.1 kB). View file
 
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/_tensorboard_logger.cpython-310.pyc ADDED
Binary file (6.86 kB). View file
 
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/_upload_large_folder.cpython-310.pyc ADDED
Binary file (20.7 kB). View file
 
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/_webhooks_payload.cpython-310.pyc ADDED
Binary file (3.93 kB). View file
 
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/_webhooks_server.cpython-310.pyc ADDED
Binary file (14 kB). View file
 
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/community.cpython-310.pyc ADDED
Binary file (13 kB). View file
 
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/constants.cpython-310.pyc ADDED
Binary file (6.02 kB). View file
 
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/dataclasses.cpython-310.pyc ADDED
Binary file (13.6 kB). View file
 
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/errors.cpython-310.pyc ADDED
Binary file (14.1 kB). View file
 
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/fastai_utils.cpython-310.pyc ADDED
Binary file (14.3 kB). View file
 
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/file_download.cpython-310.pyc ADDED
Binary file (50.8 kB). View file
 
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/hf_file_system.cpython-310.pyc ADDED
Binary file (33.1 kB). View file
 
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/hub_mixin.cpython-310.pyc ADDED
Binary file (29.1 kB). View file
 
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/inference_api.cpython-310.pyc ADDED
Binary file (7.57 kB). View file
 
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/keras_mixin.cpython-310.pyc ADDED
Binary file (17 kB). View file
 
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/lfs.cpython-310.pyc ADDED
Binary file (13.8 kB). View file
 
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/repocard.cpython-310.pyc ADDED
Binary file (29.6 kB). View file
 
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/repocard_data.cpython-310.pyc ADDED
Binary file (29.4 kB). View file
 
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/repository.cpython-310.pyc ADDED
Binary file (45.6 kB). View file
 
pythonProject/.venv/Lib/site-packages/huggingface_hub/inference/__pycache__/__init__.cpython-310.pyc ADDED
Binary file (182 Bytes). View file
 
pythonProject/.venv/Lib/site-packages/huggingface_hub/inference/__pycache__/_client.cpython-310.pyc ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0fcedb6fe5ff0f30068953849983fc2d2bc0163e249770936db741589d32fc62
3
+ size 137799
pythonProject/.venv/Lib/site-packages/huggingface_hub/inference/__pycache__/_common.cpython-310.pyc ADDED
Binary file (12 kB). View file
 
pythonProject/.venv/Lib/site-packages/huggingface_hub/inference/_providers/__pycache__/__init__.cpython-310.pyc ADDED
Binary file (5.64 kB). View file
 
pythonProject/.venv/Lib/site-packages/huggingface_hub/inference/_providers/__pycache__/_common.cpython-310.pyc ADDED
Binary file (10.3 kB). View file
 
pythonProject/.venv/Lib/site-packages/huggingface_hub/inference/_providers/__pycache__/replicate.cpython-310.pyc ADDED
Binary file (4.45 kB). View file
 
pythonProject/.venv/Lib/site-packages/huggingface_hub/inference/_providers/__pycache__/sambanova.cpython-310.pyc ADDED
Binary file (2.92 kB). View file
 
pythonProject/.venv/Lib/site-packages/huggingface_hub/inference/_providers/__pycache__/together.cpython-310.pyc ADDED
Binary file (4.2 kB). View file
 
pythonProject/.venv/Lib/site-packages/huggingface_hub/serialization/__init__.py ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright 2024 The HuggingFace Team. All rights reserved.
2
+ #
3
+ # Licensed under the Apache License, Version 2.0 (the "License");
4
+ # you may not use this file except in compliance with the License.
5
+ # You may obtain a copy of the License at
6
+ #
7
+ # http://www.apache.org/licenses/LICENSE-2.0
8
+ #
9
+ # Unless required by applicable law or agreed to in writing, software
10
+ # distributed under the License is distributed on an "AS IS" BASIS,
11
+ # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12
+ # See the License for the specific language governing permissions and
13
+ # limitations under the License.
14
+ # ruff: noqa: F401
15
+ """Contains helpers to serialize tensors."""
16
+
17
+ from ._base import StateDictSplit, split_state_dict_into_shards_factory
18
+ from ._tensorflow import get_tf_storage_size, split_tf_state_dict_into_shards
19
+ from ._torch import (
20
+ get_torch_storage_id,
21
+ get_torch_storage_size,
22
+ load_state_dict_from_file,
23
+ load_torch_model,
24
+ save_torch_model,
25
+ save_torch_state_dict,
26
+ split_torch_state_dict_into_shards,
27
+ )
pythonProject/.venv/Lib/site-packages/huggingface_hub/serialization/_base.py ADDED
@@ -0,0 +1,210 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright 2024 The HuggingFace Team. All rights reserved.
2
+ #
3
+ # Licensed under the Apache License, Version 2.0 (the "License");
4
+ # you may not use this file except in compliance with the License.
5
+ # You may obtain a copy of the License at
6
+ #
7
+ # http://www.apache.org/licenses/LICENSE-2.0
8
+ #
9
+ # Unless required by applicable law or agreed to in writing, software
10
+ # distributed under the License is distributed on an "AS IS" BASIS,
11
+ # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12
+ # See the License for the specific language governing permissions and
13
+ # limitations under the License.
14
+ """Contains helpers to split tensors into shards."""
15
+
16
+ from dataclasses import dataclass, field
17
+ from typing import Any, Callable, Dict, List, Optional, TypeVar, Union
18
+
19
+ from .. import logging
20
+
21
+
22
+ TensorT = TypeVar("TensorT")
23
+ TensorSizeFn_T = Callable[[TensorT], int]
24
+ StorageIDFn_T = Callable[[TensorT], Optional[Any]]
25
+
26
+ MAX_SHARD_SIZE = "5GB"
27
+ SIZE_UNITS = {
28
+ "TB": 10**12,
29
+ "GB": 10**9,
30
+ "MB": 10**6,
31
+ "KB": 10**3,
32
+ }
33
+
34
+
35
+ logger = logging.get_logger(__file__)
36
+
37
+
38
+ @dataclass
39
+ class StateDictSplit:
40
+ is_sharded: bool = field(init=False)
41
+ metadata: Dict[str, Any]
42
+ filename_to_tensors: Dict[str, List[str]]
43
+ tensor_to_filename: Dict[str, str]
44
+
45
+ def __post_init__(self):
46
+ self.is_sharded = len(self.filename_to_tensors) > 1
47
+
48
+
49
+ def split_state_dict_into_shards_factory(
50
+ state_dict: Dict[str, TensorT],
51
+ *,
52
+ get_storage_size: TensorSizeFn_T,
53
+ filename_pattern: str,
54
+ get_storage_id: StorageIDFn_T = lambda tensor: None,
55
+ max_shard_size: Union[int, str] = MAX_SHARD_SIZE,
56
+ ) -> StateDictSplit:
57
+ """
58
+ Split a model state dictionary in shards so that each shard is smaller than a given size.
59
+
60
+ The shards are determined by iterating through the `state_dict` in the order of its keys. There is no optimization
61
+ made to make each shard as close as possible to the maximum size passed. For example, if the limit is 10GB and we
62
+ have tensors of sizes [6GB, 6GB, 2GB, 6GB, 2GB, 2GB] they will get sharded as [6GB], [6+2GB], [6+2+2GB] and not
63
+ [6+2+2GB], [6+2GB], [6GB].
64
+
65
+ <Tip warning={true}>
66
+
67
+ If one of the model's tensor is bigger than `max_shard_size`, it will end up in its own shard which will have a
68
+ size greater than `max_shard_size`.
69
+
70
+ </Tip>
71
+
72
+ Args:
73
+ state_dict (`Dict[str, Tensor]`):
74
+ The state dictionary to save.
75
+ get_storage_size (`Callable[[Tensor], int]`):
76
+ A function that returns the size of a tensor when saved on disk in bytes.
77
+ get_storage_id (`Callable[[Tensor], Optional[Any]]`, *optional*):
78
+ A function that returns a unique identifier to a tensor storage. Multiple different tensors can share the
79
+ same underlying storage. This identifier is guaranteed to be unique and constant for this tensor's storage
80
+ during its lifetime. Two tensor storages with non-overlapping lifetimes may have the same id.
81
+ filename_pattern (`str`, *optional*):
82
+ The pattern to generate the files names in which the model will be saved. Pattern must be a string that
83
+ can be formatted with `filename_pattern.format(suffix=...)` and must contain the keyword `suffix`
84
+ max_shard_size (`int` or `str`, *optional*):
85
+ The maximum size of each shard, in bytes. Defaults to 5GB.
86
+
87
+ Returns:
88
+ [`StateDictSplit`]: A `StateDictSplit` object containing the shards and the index to retrieve them.
89
+ """
90
+ storage_id_to_tensors: Dict[Any, List[str]] = {}
91
+
92
+ shard_list: List[Dict[str, TensorT]] = []
93
+ current_shard: Dict[str, TensorT] = {}
94
+ current_shard_size = 0
95
+ total_size = 0
96
+
97
+ if isinstance(max_shard_size, str):
98
+ max_shard_size = parse_size_to_int(max_shard_size)
99
+
100
+ for key, tensor in state_dict.items():
101
+ # when bnb serialization is used the weights in the state dict can be strings
102
+ # check: https://github.com/huggingface/transformers/pull/24416 for more details
103
+ if isinstance(tensor, str):
104
+ logger.info("Skipping tensor %s as it is a string (bnb serialization)", key)
105
+ continue
106
+
107
+ # If a `tensor` shares the same underlying storage as another tensor, we put `tensor` in the same `block`
108
+ storage_id = get_storage_id(tensor)
109
+ if storage_id is not None:
110
+ if storage_id in storage_id_to_tensors:
111
+ # We skip this tensor for now and will reassign to correct shard later
112
+ storage_id_to_tensors[storage_id].append(key)
113
+ continue
114
+ else:
115
+ # This is the first tensor with this storage_id, we create a new entry
116
+ # in the storage_id_to_tensors dict => we will assign the shard id later
117
+ storage_id_to_tensors[storage_id] = [key]
118
+
119
+ # Compute tensor size
120
+ tensor_size = get_storage_size(tensor)
121
+
122
+ # If this tensor is bigger than the maximal size, we put it in its own shard
123
+ if tensor_size > max_shard_size:
124
+ total_size += tensor_size
125
+ shard_list.append({key: tensor})
126
+ continue
127
+
128
+ # If this tensor is going to tip up over the maximal size, we split.
129
+ # Current shard already has some tensors, we add it to the list of shards and create a new one.
130
+ if current_shard_size + tensor_size > max_shard_size:
131
+ shard_list.append(current_shard)
132
+ current_shard = {}
133
+ current_shard_size = 0
134
+
135
+ # Add the tensor to the current shard
136
+ current_shard[key] = tensor
137
+ current_shard_size += tensor_size
138
+ total_size += tensor_size
139
+
140
+ # Add the last shard
141
+ if len(current_shard) > 0:
142
+ shard_list.append(current_shard)
143
+ nb_shards = len(shard_list)
144
+
145
+ # Loop over the tensors that share the same storage and assign them together
146
+ for storage_id, keys in storage_id_to_tensors.items():
147
+ # Let's try to find the shard where the first tensor of this storage is and put all tensors in the same shard
148
+ for shard in shard_list:
149
+ if keys[0] in shard:
150
+ for key in keys:
151
+ shard[key] = state_dict[key]
152
+ break
153
+
154
+ # If we only have one shard, we return it => no need to build the index
155
+ if nb_shards == 1:
156
+ filename = filename_pattern.format(suffix="")
157
+ return StateDictSplit(
158
+ metadata={"total_size": total_size},
159
+ filename_to_tensors={filename: list(state_dict.keys())},
160
+ tensor_to_filename={key: filename for key in state_dict.keys()},
161
+ )
162
+
163
+ # Now that each tensor is assigned to a shard, let's assign a filename to each shard
164
+ tensor_name_to_filename = {}
165
+ filename_to_tensors = {}
166
+ for idx, shard in enumerate(shard_list):
167
+ filename = filename_pattern.format(suffix=f"-{idx + 1:05d}-of-{nb_shards:05d}")
168
+ for key in shard:
169
+ tensor_name_to_filename[key] = filename
170
+ filename_to_tensors[filename] = list(shard.keys())
171
+
172
+ # Build the index and return
173
+ return StateDictSplit(
174
+ metadata={"total_size": total_size},
175
+ filename_to_tensors=filename_to_tensors,
176
+ tensor_to_filename=tensor_name_to_filename,
177
+ )
178
+
179
+
180
+ def parse_size_to_int(size_as_str: str) -> int:
181
+ """
182
+ Parse a size expressed as a string with digits and unit (like `"5MB"`) to an integer (in bytes).
183
+
184
+ Supported units are "TB", "GB", "MB", "KB".
185
+
186
+ Args:
187
+ size_as_str (`str`): The size to convert. Will be directly returned if an `int`.
188
+
189
+ Example:
190
+
191
+ ```py
192
+ >>> parse_size_to_int("5MB")
193
+ 5000000
194
+ ```
195
+ """
196
+ size_as_str = size_as_str.strip()
197
+
198
+ # Parse unit
199
+ unit = size_as_str[-2:].upper()
200
+ if unit not in SIZE_UNITS:
201
+ raise ValueError(f"Unit '{unit}' not supported. Supported units are TB, GB, MB, KB. Got '{size_as_str}'.")
202
+ multiplier = SIZE_UNITS[unit]
203
+
204
+ # Parse value
205
+ try:
206
+ value = float(size_as_str[:-2].strip())
207
+ except ValueError as e:
208
+ raise ValueError(f"Could not parse the size value from '{size_as_str}': {e}") from e
209
+
210
+ return int(value * multiplier)
pythonProject/.venv/Lib/site-packages/huggingface_hub/serialization/_dduf.py ADDED
@@ -0,0 +1,387 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import json
2
+ import logging
3
+ import mmap
4
+ import os
5
+ import shutil
6
+ import zipfile
7
+ from contextlib import contextmanager
8
+ from dataclasses import dataclass, field
9
+ from pathlib import Path
10
+ from typing import Any, Dict, Generator, Iterable, Tuple, Union
11
+
12
+ from ..errors import DDUFCorruptedFileError, DDUFExportError, DDUFInvalidEntryNameError
13
+
14
+
15
+ logger = logging.getLogger(__name__)
16
+
17
+ DDUF_ALLOWED_ENTRIES = {
18
+ # Allowed file extensions in a DDUF file
19
+ ".json",
20
+ ".model",
21
+ ".safetensors",
22
+ ".txt",
23
+ }
24
+
25
+ DDUF_FOLDER_REQUIRED_ENTRIES = {
26
+ # Each folder must contain at least one of these entries
27
+ "config.json",
28
+ "tokenizer_config.json",
29
+ "preprocessor_config.json",
30
+ "scheduler_config.json",
31
+ }
32
+
33
+
34
+ @dataclass
35
+ class DDUFEntry:
36
+ """Object representing a file entry in a DDUF file.
37
+
38
+ See [`read_dduf_file`] for how to read a DDUF file.
39
+
40
+ Attributes:
41
+ filename (str):
42
+ The name of the file in the DDUF archive.
43
+ offset (int):
44
+ The offset of the file in the DDUF archive.
45
+ length (int):
46
+ The length of the file in the DDUF archive.
47
+ dduf_path (str):
48
+ The path to the DDUF archive (for internal use).
49
+ """
50
+
51
+ filename: str
52
+ length: int
53
+ offset: int
54
+
55
+ dduf_path: Path = field(repr=False)
56
+
57
+ @contextmanager
58
+ def as_mmap(self) -> Generator[bytes, None, None]:
59
+ """Open the file as a memory-mapped file.
60
+
61
+ Useful to load safetensors directly from the file.
62
+
63
+ Example:
64
+ ```py
65
+ >>> import safetensors.torch
66
+ >>> with entry.as_mmap() as mm:
67
+ ... tensors = safetensors.torch.load(mm)
68
+ ```
69
+ """
70
+ with self.dduf_path.open("rb") as f:
71
+ with mmap.mmap(f.fileno(), length=0, access=mmap.ACCESS_READ) as mm:
72
+ yield mm[self.offset : self.offset + self.length]
73
+
74
+ def read_text(self, encoding: str = "utf-8") -> str:
75
+ """Read the file as text.
76
+
77
+ Useful for '.txt' and '.json' entries.
78
+
79
+ Example:
80
+ ```py
81
+ >>> import json
82
+ >>> index = json.loads(entry.read_text())
83
+ ```
84
+ """
85
+ with self.dduf_path.open("rb") as f:
86
+ f.seek(self.offset)
87
+ return f.read(self.length).decode(encoding=encoding)
88
+
89
+
90
+ def read_dduf_file(dduf_path: Union[os.PathLike, str]) -> Dict[str, DDUFEntry]:
91
+ """
92
+ Read a DDUF file and return a dictionary of entries.
93
+
94
+ Only the metadata is read, the data is not loaded in memory.
95
+
96
+ Args:
97
+ dduf_path (`str` or `os.PathLike`):
98
+ The path to the DDUF file to read.
99
+
100
+ Returns:
101
+ `Dict[str, DDUFEntry]`:
102
+ A dictionary of [`DDUFEntry`] indexed by filename.
103
+
104
+ Raises:
105
+ - [`DDUFCorruptedFileError`]: If the DDUF file is corrupted (i.e. doesn't follow the DDUF format).
106
+
107
+ Example:
108
+ ```python
109
+ >>> import json
110
+ >>> import safetensors.torch
111
+ >>> from huggingface_hub import read_dduf_file
112
+
113
+ # Read DDUF metadata
114
+ >>> dduf_entries = read_dduf_file("FLUX.1-dev.dduf")
115
+
116
+ # Returns a mapping filename <> DDUFEntry
117
+ >>> dduf_entries["model_index.json"]
118
+ DDUFEntry(filename='model_index.json', offset=66, length=587)
119
+
120
+ # Load model index as JSON
121
+ >>> json.loads(dduf_entries["model_index.json"].read_text())
122
+ {'_class_name': 'FluxPipeline', '_diffusers_version': '0.32.0.dev0', '_name_or_path': 'black-forest-labs/FLUX.1-dev', ...
123
+
124
+ # Load VAE weights using safetensors
125
+ >>> with dduf_entries["vae/diffusion_pytorch_model.safetensors"].as_mmap() as mm:
126
+ ... state_dict = safetensors.torch.load(mm)
127
+ ```
128
+ """
129
+ entries = {}
130
+ dduf_path = Path(dduf_path)
131
+ logger.info(f"Reading DDUF file {dduf_path}")
132
+ with zipfile.ZipFile(str(dduf_path), "r") as zf:
133
+ for info in zf.infolist():
134
+ logger.debug(f"Reading entry {info.filename}")
135
+ if info.compress_type != zipfile.ZIP_STORED:
136
+ raise DDUFCorruptedFileError("Data must not be compressed in DDUF file.")
137
+
138
+ try:
139
+ _validate_dduf_entry_name(info.filename)
140
+ except DDUFInvalidEntryNameError as e:
141
+ raise DDUFCorruptedFileError(f"Invalid entry name in DDUF file: {info.filename}") from e
142
+
143
+ offset = _get_data_offset(zf, info)
144
+
145
+ entries[info.filename] = DDUFEntry(
146
+ filename=info.filename, offset=offset, length=info.file_size, dduf_path=dduf_path
147
+ )
148
+
149
+ # Consistency checks on the DDUF file
150
+ if "model_index.json" not in entries:
151
+ raise DDUFCorruptedFileError("Missing required 'model_index.json' entry in DDUF file.")
152
+ index = json.loads(entries["model_index.json"].read_text())
153
+ _validate_dduf_structure(index, entries.keys())
154
+
155
+ logger.info(f"Done reading DDUF file {dduf_path}. Found {len(entries)} entries")
156
+ return entries
157
+
158
+
159
+ def export_entries_as_dduf(
160
+ dduf_path: Union[str, os.PathLike], entries: Iterable[Tuple[str, Union[str, Path, bytes]]]
161
+ ) -> None:
162
+ """Write a DDUF file from an iterable of entries.
163
+
164
+ This is a lower-level helper than [`export_folder_as_dduf`] that allows more flexibility when serializing data.
165
+ In particular, you don't need to save the data on disk before exporting it in the DDUF file.
166
+
167
+ Args:
168
+ dduf_path (`str` or `os.PathLike`):
169
+ The path to the DDUF file to write.
170
+ entries (`Iterable[Tuple[str, Union[str, Path, bytes]]]`):
171
+ An iterable of entries to write in the DDUF file. Each entry is a tuple with the filename and the content.
172
+ The filename should be the path to the file in the DDUF archive.
173
+ The content can be a string or a pathlib.Path representing a path to a file on the local disk or directly the content as bytes.
174
+
175
+ Raises:
176
+ - [`DDUFExportError`]: If anything goes wrong during the export (e.g. invalid entry name, missing 'model_index.json', etc.).
177
+
178
+ Example:
179
+ ```python
180
+ # Export specific files from the local disk.
181
+ >>> from huggingface_hub import export_entries_as_dduf
182
+ >>> export_entries_as_dduf(
183
+ ... dduf_path="stable-diffusion-v1-4-FP16.dduf",
184
+ ... entries=[ # List entries to add to the DDUF file (here, only FP16 weights)
185
+ ... ("model_index.json", "path/to/model_index.json"),
186
+ ... ("vae/config.json", "path/to/vae/config.json"),
187
+ ... ("vae/diffusion_pytorch_model.fp16.safetensors", "path/to/vae/diffusion_pytorch_model.fp16.safetensors"),
188
+ ... ("text_encoder/config.json", "path/to/text_encoder/config.json"),
189
+ ... ("text_encoder/model.fp16.safetensors", "path/to/text_encoder/model.fp16.safetensors"),
190
+ ... # ... add more entries here
191
+ ... ]
192
+ ... )
193
+ ```
194
+
195
+ ```python
196
+ # Export state_dicts one by one from a loaded pipeline
197
+ >>> from diffusers import DiffusionPipeline
198
+ >>> from typing import Generator, Tuple
199
+ >>> import safetensors.torch
200
+ >>> from huggingface_hub import export_entries_as_dduf
201
+ >>> pipe = DiffusionPipeline.from_pretrained("CompVis/stable-diffusion-v1-4")
202
+ ... # ... do some work with the pipeline
203
+
204
+ >>> def as_entries(pipe: DiffusionPipeline) -> Generator[Tuple[str, bytes], None, None]:
205
+ ... # Build an generator that yields the entries to add to the DDUF file.
206
+ ... # The first element of the tuple is the filename in the DDUF archive (must use UNIX separator!). The second element is the content of the file.
207
+ ... # Entries will be evaluated lazily when the DDUF file is created (only 1 entry is loaded in memory at a time)
208
+ ... yield "vae/config.json", pipe.vae.to_json_string().encode()
209
+ ... yield "vae/diffusion_pytorch_model.safetensors", safetensors.torch.save(pipe.vae.state_dict())
210
+ ... yield "text_encoder/config.json", pipe.text_encoder.config.to_json_string().encode()
211
+ ... yield "text_encoder/model.safetensors", safetensors.torch.save(pipe.text_encoder.state_dict())
212
+ ... # ... add more entries here
213
+
214
+ >>> export_entries_as_dduf(dduf_path="stable-diffusion-v1-4.dduf", entries=as_entries(pipe))
215
+ ```
216
+ """
217
+ logger.info(f"Exporting DDUF file '{dduf_path}'")
218
+ filenames = set()
219
+ index = None
220
+ with zipfile.ZipFile(str(dduf_path), "w", zipfile.ZIP_STORED) as archive:
221
+ for filename, content in entries:
222
+ if filename in filenames:
223
+ raise DDUFExportError(f"Can't add duplicate entry: {filename}")
224
+ filenames.add(filename)
225
+
226
+ if filename == "model_index.json":
227
+ try:
228
+ index = json.loads(_load_content(content).decode())
229
+ except json.JSONDecodeError as e:
230
+ raise DDUFExportError("Failed to parse 'model_index.json'.") from e
231
+
232
+ try:
233
+ filename = _validate_dduf_entry_name(filename)
234
+ except DDUFInvalidEntryNameError as e:
235
+ raise DDUFExportError(f"Invalid entry name: {filename}") from e
236
+ logger.debug(f"Adding entry '{filename}' to DDUF file")
237
+ _dump_content_in_archive(archive, filename, content)
238
+
239
+ # Consistency checks on the DDUF file
240
+ if index is None:
241
+ raise DDUFExportError("Missing required 'model_index.json' entry in DDUF file.")
242
+ try:
243
+ _validate_dduf_structure(index, filenames)
244
+ except DDUFCorruptedFileError as e:
245
+ raise DDUFExportError("Invalid DDUF file structure.") from e
246
+
247
+ logger.info(f"Done writing DDUF file {dduf_path}")
248
+
249
+
250
+ def export_folder_as_dduf(dduf_path: Union[str, os.PathLike], folder_path: Union[str, os.PathLike]) -> None:
251
+ """
252
+ Export a folder as a DDUF file.
253
+
254
+ AUses [`export_entries_as_dduf`] under the hood.
255
+
256
+ Args:
257
+ dduf_path (`str` or `os.PathLike`):
258
+ The path to the DDUF file to write.
259
+ folder_path (`str` or `os.PathLike`):
260
+ The path to the folder containing the diffusion model.
261
+
262
+ Example:
263
+ ```python
264
+ >>> from huggingface_hub import export_folder_as_dduf
265
+ >>> export_folder_as_dduf(dduf_path="FLUX.1-dev.dduf", folder_path="path/to/FLUX.1-dev")
266
+ ```
267
+ """
268
+ folder_path = Path(folder_path)
269
+
270
+ def _iterate_over_folder() -> Iterable[Tuple[str, Path]]:
271
+ for path in Path(folder_path).glob("**/*"):
272
+ if not path.is_file():
273
+ continue
274
+ if path.suffix not in DDUF_ALLOWED_ENTRIES:
275
+ logger.debug(f"Skipping file '{path}' (file type not allowed)")
276
+ continue
277
+ path_in_archive = path.relative_to(folder_path)
278
+ if len(path_in_archive.parts) >= 3:
279
+ logger.debug(f"Skipping file '{path}' (nested directories not allowed)")
280
+ continue
281
+ yield path_in_archive.as_posix(), path
282
+
283
+ export_entries_as_dduf(dduf_path, _iterate_over_folder())
284
+
285
+
286
+ def _dump_content_in_archive(archive: zipfile.ZipFile, filename: str, content: Union[str, os.PathLike, bytes]) -> None:
287
+ with archive.open(filename, "w", force_zip64=True) as archive_fh:
288
+ if isinstance(content, (str, Path)):
289
+ content_path = Path(content)
290
+ with content_path.open("rb") as content_fh:
291
+ shutil.copyfileobj(content_fh, archive_fh, 1024 * 1024 * 8) # type: ignore[misc]
292
+ elif isinstance(content, bytes):
293
+ archive_fh.write(content)
294
+ else:
295
+ raise DDUFExportError(f"Invalid content type for {filename}. Must be str, Path or bytes.")
296
+
297
+
298
+ def _load_content(content: Union[str, Path, bytes]) -> bytes:
299
+ """Load the content of an entry as bytes.
300
+
301
+ Used only for small checks (not to dump content into archive).
302
+ """
303
+ if isinstance(content, (str, Path)):
304
+ return Path(content).read_bytes()
305
+ elif isinstance(content, bytes):
306
+ return content
307
+ else:
308
+ raise DDUFExportError(f"Invalid content type. Must be str, Path or bytes. Got {type(content)}.")
309
+
310
+
311
+ def _validate_dduf_entry_name(entry_name: str) -> str:
312
+ if "." + entry_name.split(".")[-1] not in DDUF_ALLOWED_ENTRIES:
313
+ raise DDUFInvalidEntryNameError(f"File type not allowed: {entry_name}")
314
+ if "\\" in entry_name:
315
+ raise DDUFInvalidEntryNameError(f"Entry names must use UNIX separators ('/'). Got {entry_name}.")
316
+ entry_name = entry_name.strip("/")
317
+ if entry_name.count("/") > 1:
318
+ raise DDUFInvalidEntryNameError(f"DDUF only supports 1 level of directory. Got {entry_name}.")
319
+ return entry_name
320
+
321
+
322
+ def _validate_dduf_structure(index: Any, entry_names: Iterable[str]) -> None:
323
+ """
324
+ Consistency checks on the DDUF file structure.
325
+
326
+ Rules:
327
+ - The 'model_index.json' entry is required and must contain a dictionary.
328
+ - Each folder name must correspond to an entry in 'model_index.json'.
329
+ - Each folder must contain at least a config file ('config.json', 'tokenizer_config.json', 'preprocessor_config.json', 'scheduler_config.json').
330
+
331
+ Args:
332
+ index (Any):
333
+ The content of the 'model_index.json' entry.
334
+ entry_names (Iterable[str]):
335
+ The list of entry names in the DDUF file.
336
+
337
+ Raises:
338
+ - [`DDUFCorruptedFileError`]: If the DDUF file is corrupted (i.e. doesn't follow the DDUF format).
339
+ """
340
+ if not isinstance(index, dict):
341
+ raise DDUFCorruptedFileError(f"Invalid 'model_index.json' content. Must be a dictionary. Got {type(index)}.")
342
+
343
+ dduf_folders = {entry.split("/")[0] for entry in entry_names if "/" in entry}
344
+ for folder in dduf_folders:
345
+ if folder not in index:
346
+ raise DDUFCorruptedFileError(f"Missing required entry '{folder}' in 'model_index.json'.")
347
+ if not any(f"{folder}/{required_entry}" in entry_names for required_entry in DDUF_FOLDER_REQUIRED_ENTRIES):
348
+ raise DDUFCorruptedFileError(
349
+ f"Missing required file in folder '{folder}'. Must contains at least one of {DDUF_FOLDER_REQUIRED_ENTRIES}."
350
+ )
351
+
352
+
353
+ def _get_data_offset(zf: zipfile.ZipFile, info: zipfile.ZipInfo) -> int:
354
+ """
355
+ Calculate the data offset for a file in a ZIP archive.
356
+
357
+ Args:
358
+ zf (`zipfile.ZipFile`):
359
+ The opened ZIP file. Must be opened in read mode.
360
+ info (`zipfile.ZipInfo`):
361
+ The file info.
362
+
363
+ Returns:
364
+ int: The offset of the file data in the ZIP archive.
365
+ """
366
+ if zf.fp is None:
367
+ raise DDUFCorruptedFileError("ZipFile object must be opened in read mode.")
368
+
369
+ # Step 1: Get the local file header offset
370
+ header_offset = info.header_offset
371
+
372
+ # Step 2: Read the local file header
373
+ zf.fp.seek(header_offset)
374
+ local_file_header = zf.fp.read(30) # Fixed-size part of the local header
375
+
376
+ if len(local_file_header) < 30:
377
+ raise DDUFCorruptedFileError("Incomplete local file header.")
378
+
379
+ # Step 3: Parse the header fields to calculate the start of file data
380
+ # Local file header: https://en.wikipedia.org/wiki/ZIP_(file_format)#File_headers
381
+ filename_len = int.from_bytes(local_file_header[26:28], "little")
382
+ extra_field_len = int.from_bytes(local_file_header[28:30], "little")
383
+
384
+ # Data offset is after the fixed header, filename, and extra fields
385
+ data_offset = header_offset + 30 + filename_len + extra_field_len
386
+
387
+ return data_offset
pythonProject/.venv/Lib/site-packages/huggingface_hub/serialization/_tensorflow.py ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright 2024 The HuggingFace Team. All rights reserved.
2
+ #
3
+ # Licensed under the Apache License, Version 2.0 (the "License");
4
+ # you may not use this file except in compliance with the License.
5
+ # You may obtain a copy of the License at
6
+ #
7
+ # http://www.apache.org/licenses/LICENSE-2.0
8
+ #
9
+ # Unless required by applicable law or agreed to in writing, software
10
+ # distributed under the License is distributed on an "AS IS" BASIS,
11
+ # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12
+ # See the License for the specific language governing permissions and
13
+ # limitations under the License.
14
+ """Contains tensorflow-specific helpers."""
15
+
16
+ import math
17
+ import re
18
+ from typing import TYPE_CHECKING, Dict, Union
19
+
20
+ from .. import constants
21
+ from ._base import MAX_SHARD_SIZE, StateDictSplit, split_state_dict_into_shards_factory
22
+
23
+
24
+ if TYPE_CHECKING:
25
+ import tensorflow as tf
26
+
27
+
28
+ def split_tf_state_dict_into_shards(
29
+ state_dict: Dict[str, "tf.Tensor"],
30
+ *,
31
+ filename_pattern: str = constants.TF2_WEIGHTS_FILE_PATTERN,
32
+ max_shard_size: Union[int, str] = MAX_SHARD_SIZE,
33
+ ) -> StateDictSplit:
34
+ """
35
+ Split a model state dictionary in shards so that each shard is smaller than a given size.
36
+
37
+ The shards are determined by iterating through the `state_dict` in the order of its keys. There is no optimization
38
+ made to make each shard as close as possible to the maximum size passed. For example, if the limit is 10GB and we
39
+ have tensors of sizes [6GB, 6GB, 2GB, 6GB, 2GB, 2GB] they will get sharded as [6GB], [6+2GB], [6+2+2GB] and not
40
+ [6+2+2GB], [6+2GB], [6GB].
41
+
42
+ <Tip warning={true}>
43
+
44
+ If one of the model's tensor is bigger than `max_shard_size`, it will end up in its own shard which will have a
45
+ size greater than `max_shard_size`.
46
+
47
+ </Tip>
48
+
49
+ Args:
50
+ state_dict (`Dict[str, Tensor]`):
51
+ The state dictionary to save.
52
+ filename_pattern (`str`, *optional*):
53
+ The pattern to generate the files names in which the model will be saved. Pattern must be a string that
54
+ can be formatted with `filename_pattern.format(suffix=...)` and must contain the keyword `suffix`
55
+ Defaults to `"tf_model{suffix}.h5"`.
56
+ max_shard_size (`int` or `str`, *optional*):
57
+ The maximum size of each shard, in bytes. Defaults to 5GB.
58
+
59
+ Returns:
60
+ [`StateDictSplit`]: A `StateDictSplit` object containing the shards and the index to retrieve them.
61
+ """
62
+ return split_state_dict_into_shards_factory(
63
+ state_dict,
64
+ max_shard_size=max_shard_size,
65
+ filename_pattern=filename_pattern,
66
+ get_storage_size=get_tf_storage_size,
67
+ )
68
+
69
+
70
+ def get_tf_storage_size(tensor: "tf.Tensor") -> int:
71
+ # Return `math.ceil` since dtype byte size can be a float (e.g., 0.125 for tf.bool).
72
+ # Better to overestimate than underestimate.
73
+ return math.ceil(tensor.numpy().size * _dtype_byte_size_tf(tensor.dtype))
74
+
75
+
76
+ def _dtype_byte_size_tf(dtype) -> float:
77
+ """
78
+ Returns the size (in bytes) occupied by one parameter of type `dtype`.
79
+ Taken from https://github.com/huggingface/transformers/blob/74d9d0cebb0263a3f8ab9c280569170cc74651d0/src/transformers/modeling_tf_utils.py#L608.
80
+ NOTE: why not `tensor.numpy().nbytes`?
81
+ Example:
82
+ ```py
83
+ >>> _dtype_byte_size(tf.float32)
84
+ 4
85
+ ```
86
+ """
87
+ import tensorflow as tf
88
+
89
+ if dtype == tf.bool:
90
+ return 1 / 8
91
+ bit_search = re.search(r"[^\d](\d+)$", dtype.name)
92
+ if bit_search is None:
93
+ raise ValueError(f"`dtype` is not a valid dtype: {dtype}.")
94
+ bit_size = int(bit_search.groups()[0])
95
+ return bit_size // 8