Add files using upload-large-folder tool
Browse filesThis view is limited to 50 files because it contains too many changes. See raw diff
- .gitattributes +1 -0
- pythonProject/.venv/Lib/site-packages/Jinja2-3.1.3.dist-info/INSTALLER +1 -0
- pythonProject/.venv/Lib/site-packages/Jinja2-3.1.3.dist-info/LICENSE.rst +28 -0
- pythonProject/.venv/Lib/site-packages/Jinja2-3.1.3.dist-info/METADATA +105 -0
- pythonProject/.venv/Lib/site-packages/Jinja2-3.1.3.dist-info/RECORD +58 -0
- pythonProject/.venv/Lib/site-packages/Jinja2-3.1.3.dist-info/WHEEL +5 -0
- pythonProject/.venv/Lib/site-packages/Jinja2-3.1.3.dist-info/entry_points.txt +2 -0
- pythonProject/.venv/Lib/site-packages/Jinja2-3.1.3.dist-info/top_level.txt +1 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub-0.34.4.dist-info/INSTALLER +1 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub-0.34.4.dist-info/LICENSE +201 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub-0.34.4.dist-info/METADATA +331 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub-0.34.4.dist-info/RECORD +327 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub-0.34.4.dist-info/WHEEL +5 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub-0.34.4.dist-info/entry_points.txt +8 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub-0.34.4.dist-info/top_level.txt +1 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/__init__.cpython-310.pyc +0 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/_commit_api.cpython-310.pyc +0 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/_commit_scheduler.cpython-310.pyc +0 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/_inference_endpoints.cpython-310.pyc +0 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/_jobs_api.cpython-310.pyc +0 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/_tensorboard_logger.cpython-310.pyc +0 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/_upload_large_folder.cpython-310.pyc +0 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/_webhooks_payload.cpython-310.pyc +0 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/_webhooks_server.cpython-310.pyc +0 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/community.cpython-310.pyc +0 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/constants.cpython-310.pyc +0 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/dataclasses.cpython-310.pyc +0 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/errors.cpython-310.pyc +0 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/fastai_utils.cpython-310.pyc +0 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/file_download.cpython-310.pyc +0 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/hf_file_system.cpython-310.pyc +0 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/hub_mixin.cpython-310.pyc +0 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/inference_api.cpython-310.pyc +0 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/keras_mixin.cpython-310.pyc +0 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/lfs.cpython-310.pyc +0 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/repocard.cpython-310.pyc +0 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/repocard_data.cpython-310.pyc +0 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/repository.cpython-310.pyc +0 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub/inference/__pycache__/__init__.cpython-310.pyc +0 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub/inference/__pycache__/_client.cpython-310.pyc +3 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub/inference/__pycache__/_common.cpython-310.pyc +0 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub/inference/_providers/__pycache__/__init__.cpython-310.pyc +0 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub/inference/_providers/__pycache__/_common.cpython-310.pyc +0 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub/inference/_providers/__pycache__/replicate.cpython-310.pyc +0 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub/inference/_providers/__pycache__/sambanova.cpython-310.pyc +0 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub/inference/_providers/__pycache__/together.cpython-310.pyc +0 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub/serialization/__init__.py +27 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub/serialization/_base.py +210 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub/serialization/_dduf.py +387 -0
- pythonProject/.venv/Lib/site-packages/huggingface_hub/serialization/_tensorflow.py +95 -0
.gitattributes
CHANGED
|
@@ -50,3 +50,4 @@ cudnn-windows-x86_64-8.9.5.30_cuda11-archive/bin/cudnn_ops_train64_8.dll filter=
|
|
| 50 |
pythonProject/.venv/Lib/site-packages/diffusers/loaders/__pycache__/lora_pipeline.cpython-310.pyc filter=lfs diff=lfs merge=lfs -text
|
| 51 |
pythonProject/.venv/Lib/site-packages/diffusers/utils/__pycache__/dummy_torch_and_transformers_objects.cpython-310.pyc filter=lfs diff=lfs merge=lfs -text
|
| 52 |
pythonProject/.venv/Lib/site-packages/functorch/_C.cp310-win_amd64.pyd filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
| 50 |
pythonProject/.venv/Lib/site-packages/diffusers/loaders/__pycache__/lora_pipeline.cpython-310.pyc filter=lfs diff=lfs merge=lfs -text
|
| 51 |
pythonProject/.venv/Lib/site-packages/diffusers/utils/__pycache__/dummy_torch_and_transformers_objects.cpython-310.pyc filter=lfs diff=lfs merge=lfs -text
|
| 52 |
pythonProject/.venv/Lib/site-packages/functorch/_C.cp310-win_amd64.pyd filter=lfs diff=lfs merge=lfs -text
|
| 53 |
+
pythonProject/.venv/Lib/site-packages/huggingface_hub/inference/__pycache__/_client.cpython-310.pyc filter=lfs diff=lfs merge=lfs -text
|
pythonProject/.venv/Lib/site-packages/Jinja2-3.1.3.dist-info/INSTALLER
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
pip
|
pythonProject/.venv/Lib/site-packages/Jinja2-3.1.3.dist-info/LICENSE.rst
ADDED
|
@@ -0,0 +1,28 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Copyright 2007 Pallets
|
| 2 |
+
|
| 3 |
+
Redistribution and use in source and binary forms, with or without
|
| 4 |
+
modification, are permitted provided that the following conditions are
|
| 5 |
+
met:
|
| 6 |
+
|
| 7 |
+
1. Redistributions of source code must retain the above copyright
|
| 8 |
+
notice, this list of conditions and the following disclaimer.
|
| 9 |
+
|
| 10 |
+
2. Redistributions in binary form must reproduce the above copyright
|
| 11 |
+
notice, this list of conditions and the following disclaimer in the
|
| 12 |
+
documentation and/or other materials provided with the distribution.
|
| 13 |
+
|
| 14 |
+
3. Neither the name of the copyright holder nor the names of its
|
| 15 |
+
contributors may be used to endorse or promote products derived from
|
| 16 |
+
this software without specific prior written permission.
|
| 17 |
+
|
| 18 |
+
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
| 19 |
+
"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
| 20 |
+
LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A
|
| 21 |
+
PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
| 22 |
+
HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
| 23 |
+
SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED
|
| 24 |
+
TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
|
| 25 |
+
PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF
|
| 26 |
+
LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING
|
| 27 |
+
NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
|
| 28 |
+
SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
pythonProject/.venv/Lib/site-packages/Jinja2-3.1.3.dist-info/METADATA
ADDED
|
@@ -0,0 +1,105 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Metadata-Version: 2.1
|
| 2 |
+
Name: Jinja2
|
| 3 |
+
Version: 3.1.3
|
| 4 |
+
Summary: A very fast and expressive template engine.
|
| 5 |
+
Home-page: https://palletsprojects.com/p/jinja/
|
| 6 |
+
Maintainer: Pallets
|
| 7 |
+
Maintainer-email: contact@palletsprojects.com
|
| 8 |
+
License: BSD-3-Clause
|
| 9 |
+
Project-URL: Donate, https://palletsprojects.com/donate
|
| 10 |
+
Project-URL: Documentation, https://jinja.palletsprojects.com/
|
| 11 |
+
Project-URL: Changes, https://jinja.palletsprojects.com/changes/
|
| 12 |
+
Project-URL: Source Code, https://github.com/pallets/jinja/
|
| 13 |
+
Project-URL: Issue Tracker, https://github.com/pallets/jinja/issues/
|
| 14 |
+
Project-URL: Chat, https://discord.gg/pallets
|
| 15 |
+
Classifier: Development Status :: 5 - Production/Stable
|
| 16 |
+
Classifier: Environment :: Web Environment
|
| 17 |
+
Classifier: Intended Audience :: Developers
|
| 18 |
+
Classifier: License :: OSI Approved :: BSD License
|
| 19 |
+
Classifier: Operating System :: OS Independent
|
| 20 |
+
Classifier: Programming Language :: Python
|
| 21 |
+
Classifier: Topic :: Internet :: WWW/HTTP :: Dynamic Content
|
| 22 |
+
Classifier: Topic :: Text Processing :: Markup :: HTML
|
| 23 |
+
Requires-Python: >=3.7
|
| 24 |
+
Description-Content-Type: text/x-rst
|
| 25 |
+
License-File: LICENSE.rst
|
| 26 |
+
Requires-Dist: MarkupSafe >=2.0
|
| 27 |
+
Provides-Extra: i18n
|
| 28 |
+
Requires-Dist: Babel >=2.7 ; extra == 'i18n'
|
| 29 |
+
|
| 30 |
+
Jinja
|
| 31 |
+
=====
|
| 32 |
+
|
| 33 |
+
Jinja is a fast, expressive, extensible templating engine. Special
|
| 34 |
+
placeholders in the template allow writing code similar to Python
|
| 35 |
+
syntax. Then the template is passed data to render the final document.
|
| 36 |
+
|
| 37 |
+
It includes:
|
| 38 |
+
|
| 39 |
+
- Template inheritance and inclusion.
|
| 40 |
+
- Define and import macros within templates.
|
| 41 |
+
- HTML templates can use autoescaping to prevent XSS from untrusted
|
| 42 |
+
user input.
|
| 43 |
+
- A sandboxed environment can safely render untrusted templates.
|
| 44 |
+
- AsyncIO support for generating templates and calling async
|
| 45 |
+
functions.
|
| 46 |
+
- I18N support with Babel.
|
| 47 |
+
- Templates are compiled to optimized Python code just-in-time and
|
| 48 |
+
cached, or can be compiled ahead-of-time.
|
| 49 |
+
- Exceptions point to the correct line in templates to make debugging
|
| 50 |
+
easier.
|
| 51 |
+
- Extensible filters, tests, functions, and even syntax.
|
| 52 |
+
|
| 53 |
+
Jinja's philosophy is that while application logic belongs in Python if
|
| 54 |
+
possible, it shouldn't make the template designer's job difficult by
|
| 55 |
+
restricting functionality too much.
|
| 56 |
+
|
| 57 |
+
|
| 58 |
+
Installing
|
| 59 |
+
----------
|
| 60 |
+
|
| 61 |
+
Install and update using `pip`_:
|
| 62 |
+
|
| 63 |
+
.. code-block:: text
|
| 64 |
+
|
| 65 |
+
$ pip install -U Jinja2
|
| 66 |
+
|
| 67 |
+
.. _pip: https://pip.pypa.io/en/stable/getting-started/
|
| 68 |
+
|
| 69 |
+
|
| 70 |
+
In A Nutshell
|
| 71 |
+
-------------
|
| 72 |
+
|
| 73 |
+
.. code-block:: jinja
|
| 74 |
+
|
| 75 |
+
{% extends "base.html" %}
|
| 76 |
+
{% block title %}Members{% endblock %}
|
| 77 |
+
{% block content %}
|
| 78 |
+
<ul>
|
| 79 |
+
{% for user in users %}
|
| 80 |
+
<li><a href="{{ user.url }}">{{ user.username }}</a></li>
|
| 81 |
+
{% endfor %}
|
| 82 |
+
</ul>
|
| 83 |
+
{% endblock %}
|
| 84 |
+
|
| 85 |
+
|
| 86 |
+
Donate
|
| 87 |
+
------
|
| 88 |
+
|
| 89 |
+
The Pallets organization develops and supports Jinja and other popular
|
| 90 |
+
packages. In order to grow the community of contributors and users, and
|
| 91 |
+
allow the maintainers to devote more time to the projects, `please
|
| 92 |
+
donate today`_.
|
| 93 |
+
|
| 94 |
+
.. _please donate today: https://palletsprojects.com/donate
|
| 95 |
+
|
| 96 |
+
|
| 97 |
+
Links
|
| 98 |
+
-----
|
| 99 |
+
|
| 100 |
+
- Documentation: https://jinja.palletsprojects.com/
|
| 101 |
+
- Changes: https://jinja.palletsprojects.com/changes/
|
| 102 |
+
- PyPI Releases: https://pypi.org/project/Jinja2/
|
| 103 |
+
- Source Code: https://github.com/pallets/jinja/
|
| 104 |
+
- Issue Tracker: https://github.com/pallets/jinja/issues/
|
| 105 |
+
- Chat: https://discord.gg/pallets
|
pythonProject/.venv/Lib/site-packages/Jinja2-3.1.3.dist-info/RECORD
ADDED
|
@@ -0,0 +1,58 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Jinja2-3.1.3.dist-info/INSTALLER,sha256=zuuue4knoyJ-UwPPXg8fezS7VCrXJQrAP7zeNuwvFQg,4
|
| 2 |
+
Jinja2-3.1.3.dist-info/LICENSE.rst,sha256=O0nc7kEF6ze6wQ-vG-JgQI_oXSUrjp3y4JefweCUQ3s,1475
|
| 3 |
+
Jinja2-3.1.3.dist-info/METADATA,sha256=0cLNbRCI91jytc7Bzv3XAQfZzFDF2gxkJuH46eF5vew,3301
|
| 4 |
+
Jinja2-3.1.3.dist-info/RECORD,,
|
| 5 |
+
Jinja2-3.1.3.dist-info/WHEEL,sha256=oiQVh_5PnQM0E3gPdiz09WCNmwiHDMaGer_elqB3coM,92
|
| 6 |
+
Jinja2-3.1.3.dist-info/entry_points.txt,sha256=zRd62fbqIyfUpsRtU7EVIFyiu1tPwfgO7EvPErnxgTE,59
|
| 7 |
+
Jinja2-3.1.3.dist-info/top_level.txt,sha256=PkeVWtLb3-CqjWi1fO29OCbj55EhX_chhKrCdrVe_zs,7
|
| 8 |
+
jinja2/__init__.py,sha256=NTBwMwsECrdHmxeXF7seusHLzrh6Ldn1A9qhS5cDuf0,1927
|
| 9 |
+
jinja2/__pycache__/__init__.cpython-310.pyc,,
|
| 10 |
+
jinja2/__pycache__/_identifier.cpython-310.pyc,,
|
| 11 |
+
jinja2/__pycache__/async_utils.cpython-310.pyc,,
|
| 12 |
+
jinja2/__pycache__/bccache.cpython-310.pyc,,
|
| 13 |
+
jinja2/__pycache__/compiler.cpython-310.pyc,,
|
| 14 |
+
jinja2/__pycache__/constants.cpython-310.pyc,,
|
| 15 |
+
jinja2/__pycache__/debug.cpython-310.pyc,,
|
| 16 |
+
jinja2/__pycache__/defaults.cpython-310.pyc,,
|
| 17 |
+
jinja2/__pycache__/environment.cpython-310.pyc,,
|
| 18 |
+
jinja2/__pycache__/exceptions.cpython-310.pyc,,
|
| 19 |
+
jinja2/__pycache__/ext.cpython-310.pyc,,
|
| 20 |
+
jinja2/__pycache__/filters.cpython-310.pyc,,
|
| 21 |
+
jinja2/__pycache__/idtracking.cpython-310.pyc,,
|
| 22 |
+
jinja2/__pycache__/lexer.cpython-310.pyc,,
|
| 23 |
+
jinja2/__pycache__/loaders.cpython-310.pyc,,
|
| 24 |
+
jinja2/__pycache__/meta.cpython-310.pyc,,
|
| 25 |
+
jinja2/__pycache__/nativetypes.cpython-310.pyc,,
|
| 26 |
+
jinja2/__pycache__/nodes.cpython-310.pyc,,
|
| 27 |
+
jinja2/__pycache__/optimizer.cpython-310.pyc,,
|
| 28 |
+
jinja2/__pycache__/parser.cpython-310.pyc,,
|
| 29 |
+
jinja2/__pycache__/runtime.cpython-310.pyc,,
|
| 30 |
+
jinja2/__pycache__/sandbox.cpython-310.pyc,,
|
| 31 |
+
jinja2/__pycache__/tests.cpython-310.pyc,,
|
| 32 |
+
jinja2/__pycache__/utils.cpython-310.pyc,,
|
| 33 |
+
jinja2/__pycache__/visitor.cpython-310.pyc,,
|
| 34 |
+
jinja2/_identifier.py,sha256=_zYctNKzRqlk_murTNlzrju1FFJL7Va_Ijqqd7ii2lU,1958
|
| 35 |
+
jinja2/async_utils.py,sha256=dFcmh6lMNfbh7eLKrBio8JqAKLHdZbpCuurFN4OERtY,2447
|
| 36 |
+
jinja2/bccache.py,sha256=mhz5xtLxCcHRAa56azOhphIAe19u1we0ojifNMClDio,14061
|
| 37 |
+
jinja2/compiler.py,sha256=PJzYdRLStlEOqmnQs1YxlizPrJoj3jTZuUleREn6AIQ,72199
|
| 38 |
+
jinja2/constants.py,sha256=GMoFydBF_kdpaRKPoM5cl5MviquVRLVyZtfp5-16jg0,1433
|
| 39 |
+
jinja2/debug.py,sha256=iWJ432RadxJNnaMOPrjIDInz50UEgni3_HKuFXi2vuQ,6299
|
| 40 |
+
jinja2/defaults.py,sha256=boBcSw78h-lp20YbaXSJsqkAI2uN_mD_TtCydpeq5wU,1267
|
| 41 |
+
jinja2/environment.py,sha256=0qldX3VQKZcm6lgn7zHz94oRFow7YPYERiqkquomNjU,61253
|
| 42 |
+
jinja2/exceptions.py,sha256=ioHeHrWwCWNaXX1inHmHVblvc4haO7AXsjCp3GfWvx0,5071
|
| 43 |
+
jinja2/ext.py,sha256=5fnMpllaXkfm2P_93RIvi-OnK7Tk8mCW8Du-GcD12Hc,31844
|
| 44 |
+
jinja2/filters.py,sha256=vYjKb2zaPShvYtn_LpSmqfS8SScbrA_KOanNibsMDIE,53862
|
| 45 |
+
jinja2/idtracking.py,sha256=GfNmadir4oDALVxzn3DL9YInhJDr69ebXeA2ygfuCGA,10704
|
| 46 |
+
jinja2/lexer.py,sha256=DW2nX9zk-6MWp65YR2bqqj0xqCvLtD-u9NWT8AnFRxQ,29726
|
| 47 |
+
jinja2/loaders.py,sha256=ayAwxfrA1SAffQta0nwSDm3TDT4KYiIGN_D9Z45B310,23085
|
| 48 |
+
jinja2/meta.py,sha256=GNPEvifmSaU3CMxlbheBOZjeZ277HThOPUTf1RkppKQ,4396
|
| 49 |
+
jinja2/nativetypes.py,sha256=7GIGALVJgdyL80oZJdQUaUfwSt5q2lSSZbXt0dNf_M4,4210
|
| 50 |
+
jinja2/nodes.py,sha256=i34GPRAZexXMT6bwuf5SEyvdmS-bRCy9KMjwN5O6pjk,34550
|
| 51 |
+
jinja2/optimizer.py,sha256=tHkMwXxfZkbfA1KmLcqmBMSaz7RLIvvItrJcPoXTyD8,1650
|
| 52 |
+
jinja2/parser.py,sha256=Y199wPL-G67gJoi5G_5sHuu9uEP1PJkjjLEW_xTH8-k,39736
|
| 53 |
+
jinja2/py.typed,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
| 54 |
+
jinja2/runtime.py,sha256=_6LkKIWFJjQdqlrgA3K39zBFQ-7Orm3wGDm96RwxQoE,33406
|
| 55 |
+
jinja2/sandbox.py,sha256=Y0xZeXQnH6EX5VjaV2YixESxoepnRbW_3UeQosaBU3M,14584
|
| 56 |
+
jinja2/tests.py,sha256=Am5Z6Lmfr2XaH_npIfJJ8MdXtWsbLjMULZJulTAj30E,5905
|
| 57 |
+
jinja2/utils.py,sha256=IMwRIcN1SsTw2-jdQtlH2KzNABsXZBW_-tnFXafQBvY,23933
|
| 58 |
+
jinja2/visitor.py,sha256=MH14C6yq24G_KVtWzjwaI7Wg14PCJIYlWW1kpkxYak0,3568
|
pythonProject/.venv/Lib/site-packages/Jinja2-3.1.3.dist-info/WHEEL
ADDED
|
@@ -0,0 +1,5 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Wheel-Version: 1.0
|
| 2 |
+
Generator: bdist_wheel (0.42.0)
|
| 3 |
+
Root-Is-Purelib: true
|
| 4 |
+
Tag: py3-none-any
|
| 5 |
+
|
pythonProject/.venv/Lib/site-packages/Jinja2-3.1.3.dist-info/entry_points.txt
ADDED
|
@@ -0,0 +1,2 @@
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[babel.extractors]
|
| 2 |
+
jinja2 = jinja2.ext:babel_extract[i18n]
|
pythonProject/.venv/Lib/site-packages/Jinja2-3.1.3.dist-info/top_level.txt
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
jinja2
|
pythonProject/.venv/Lib/site-packages/huggingface_hub-0.34.4.dist-info/INSTALLER
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
pip
|
pythonProject/.venv/Lib/site-packages/huggingface_hub-0.34.4.dist-info/LICENSE
ADDED
|
@@ -0,0 +1,201 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Apache License
|
| 2 |
+
Version 2.0, January 2004
|
| 3 |
+
http://www.apache.org/licenses/
|
| 4 |
+
|
| 5 |
+
TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION
|
| 6 |
+
|
| 7 |
+
1. Definitions.
|
| 8 |
+
|
| 9 |
+
"License" shall mean the terms and conditions for use, reproduction,
|
| 10 |
+
and distribution as defined by Sections 1 through 9 of this document.
|
| 11 |
+
|
| 12 |
+
"Licensor" shall mean the copyright owner or entity authorized by
|
| 13 |
+
the copyright owner that is granting the License.
|
| 14 |
+
|
| 15 |
+
"Legal Entity" shall mean the union of the acting entity and all
|
| 16 |
+
other entities that control, are controlled by, or are under common
|
| 17 |
+
control with that entity. For the purposes of this definition,
|
| 18 |
+
"control" means (i) the power, direct or indirect, to cause the
|
| 19 |
+
direction or management of such entity, whether by contract or
|
| 20 |
+
otherwise, or (ii) ownership of fifty percent (50%) or more of the
|
| 21 |
+
outstanding shares, or (iii) beneficial ownership of such entity.
|
| 22 |
+
|
| 23 |
+
"You" (or "Your") shall mean an individual or Legal Entity
|
| 24 |
+
exercising permissions granted by this License.
|
| 25 |
+
|
| 26 |
+
"Source" form shall mean the preferred form for making modifications,
|
| 27 |
+
including but not limited to software source code, documentation
|
| 28 |
+
source, and configuration files.
|
| 29 |
+
|
| 30 |
+
"Object" form shall mean any form resulting from mechanical
|
| 31 |
+
transformation or translation of a Source form, including but
|
| 32 |
+
not limited to compiled object code, generated documentation,
|
| 33 |
+
and conversions to other media types.
|
| 34 |
+
|
| 35 |
+
"Work" shall mean the work of authorship, whether in Source or
|
| 36 |
+
Object form, made available under the License, as indicated by a
|
| 37 |
+
copyright notice that is included in or attached to the work
|
| 38 |
+
(an example is provided in the Appendix below).
|
| 39 |
+
|
| 40 |
+
"Derivative Works" shall mean any work, whether in Source or Object
|
| 41 |
+
form, that is based on (or derived from) the Work and for which the
|
| 42 |
+
editorial revisions, annotations, elaborations, or other modifications
|
| 43 |
+
represent, as a whole, an original work of authorship. For the purposes
|
| 44 |
+
of this License, Derivative Works shall not include works that remain
|
| 45 |
+
separable from, or merely link (or bind by name) to the interfaces of,
|
| 46 |
+
the Work and Derivative Works thereof.
|
| 47 |
+
|
| 48 |
+
"Contribution" shall mean any work of authorship, including
|
| 49 |
+
the original version of the Work and any modifications or additions
|
| 50 |
+
to that Work or Derivative Works thereof, that is intentionally
|
| 51 |
+
submitted to Licensor for inclusion in the Work by the copyright owner
|
| 52 |
+
or by an individual or Legal Entity authorized to submit on behalf of
|
| 53 |
+
the copyright owner. For the purposes of this definition, "submitted"
|
| 54 |
+
means any form of electronic, verbal, or written communication sent
|
| 55 |
+
to the Licensor or its representatives, including but not limited to
|
| 56 |
+
communication on electronic mailing lists, source code control systems,
|
| 57 |
+
and issue tracking systems that are managed by, or on behalf of, the
|
| 58 |
+
Licensor for the purpose of discussing and improving the Work, but
|
| 59 |
+
excluding communication that is conspicuously marked or otherwise
|
| 60 |
+
designated in writing by the copyright owner as "Not a Contribution."
|
| 61 |
+
|
| 62 |
+
"Contributor" shall mean Licensor and any individual or Legal Entity
|
| 63 |
+
on behalf of whom a Contribution has been received by Licensor and
|
| 64 |
+
subsequently incorporated within the Work.
|
| 65 |
+
|
| 66 |
+
2. Grant of Copyright License. Subject to the terms and conditions of
|
| 67 |
+
this License, each Contributor hereby grants to You a perpetual,
|
| 68 |
+
worldwide, non-exclusive, no-charge, royalty-free, irrevocable
|
| 69 |
+
copyright license to reproduce, prepare Derivative Works of,
|
| 70 |
+
publicly display, publicly perform, sublicense, and distribute the
|
| 71 |
+
Work and such Derivative Works in Source or Object form.
|
| 72 |
+
|
| 73 |
+
3. Grant of Patent License. Subject to the terms and conditions of
|
| 74 |
+
this License, each Contributor hereby grants to You a perpetual,
|
| 75 |
+
worldwide, non-exclusive, no-charge, royalty-free, irrevocable
|
| 76 |
+
(except as stated in this section) patent license to make, have made,
|
| 77 |
+
use, offer to sell, sell, import, and otherwise transfer the Work,
|
| 78 |
+
where such license applies only to those patent claims licensable
|
| 79 |
+
by such Contributor that are necessarily infringed by their
|
| 80 |
+
Contribution(s) alone or by combination of their Contribution(s)
|
| 81 |
+
with the Work to which such Contribution(s) was submitted. If You
|
| 82 |
+
institute patent litigation against any entity (including a
|
| 83 |
+
cross-claim or counterclaim in a lawsuit) alleging that the Work
|
| 84 |
+
or a Contribution incorporated within the Work constitutes direct
|
| 85 |
+
or contributory patent infringement, then any patent licenses
|
| 86 |
+
granted to You under this License for that Work shall terminate
|
| 87 |
+
as of the date such litigation is filed.
|
| 88 |
+
|
| 89 |
+
4. Redistribution. You may reproduce and distribute copies of the
|
| 90 |
+
Work or Derivative Works thereof in any medium, with or without
|
| 91 |
+
modifications, and in Source or Object form, provided that You
|
| 92 |
+
meet the following conditions:
|
| 93 |
+
|
| 94 |
+
(a) You must give any other recipients of the Work or
|
| 95 |
+
Derivative Works a copy of this License; and
|
| 96 |
+
|
| 97 |
+
(b) You must cause any modified files to carry prominent notices
|
| 98 |
+
stating that You changed the files; and
|
| 99 |
+
|
| 100 |
+
(c) You must retain, in the Source form of any Derivative Works
|
| 101 |
+
that You distribute, all copyright, patent, trademark, and
|
| 102 |
+
attribution notices from the Source form of the Work,
|
| 103 |
+
excluding those notices that do not pertain to any part of
|
| 104 |
+
the Derivative Works; and
|
| 105 |
+
|
| 106 |
+
(d) If the Work includes a "NOTICE" text file as part of its
|
| 107 |
+
distribution, then any Derivative Works that You distribute must
|
| 108 |
+
include a readable copy of the attribution notices contained
|
| 109 |
+
within such NOTICE file, excluding those notices that do not
|
| 110 |
+
pertain to any part of the Derivative Works, in at least one
|
| 111 |
+
of the following places: within a NOTICE text file distributed
|
| 112 |
+
as part of the Derivative Works; within the Source form or
|
| 113 |
+
documentation, if provided along with the Derivative Works; or,
|
| 114 |
+
within a display generated by the Derivative Works, if and
|
| 115 |
+
wherever such third-party notices normally appear. The contents
|
| 116 |
+
of the NOTICE file are for informational purposes only and
|
| 117 |
+
do not modify the License. You may add Your own attribution
|
| 118 |
+
notices within Derivative Works that You distribute, alongside
|
| 119 |
+
or as an addendum to the NOTICE text from the Work, provided
|
| 120 |
+
that such additional attribution notices cannot be construed
|
| 121 |
+
as modifying the License.
|
| 122 |
+
|
| 123 |
+
You may add Your own copyright statement to Your modifications and
|
| 124 |
+
may provide additional or different license terms and conditions
|
| 125 |
+
for use, reproduction, or distribution of Your modifications, or
|
| 126 |
+
for any such Derivative Works as a whole, provided Your use,
|
| 127 |
+
reproduction, and distribution of the Work otherwise complies with
|
| 128 |
+
the conditions stated in this License.
|
| 129 |
+
|
| 130 |
+
5. Submission of Contributions. Unless You explicitly state otherwise,
|
| 131 |
+
any Contribution intentionally submitted for inclusion in the Work
|
| 132 |
+
by You to the Licensor shall be under the terms and conditions of
|
| 133 |
+
this License, without any additional terms or conditions.
|
| 134 |
+
Notwithstanding the above, nothing herein shall supersede or modify
|
| 135 |
+
the terms of any separate license agreement you may have executed
|
| 136 |
+
with Licensor regarding such Contributions.
|
| 137 |
+
|
| 138 |
+
6. Trademarks. This License does not grant permission to use the trade
|
| 139 |
+
names, trademarks, service marks, or product names of the Licensor,
|
| 140 |
+
except as required for reasonable and customary use in describing the
|
| 141 |
+
origin of the Work and reproducing the content of the NOTICE file.
|
| 142 |
+
|
| 143 |
+
7. Disclaimer of Warranty. Unless required by applicable law or
|
| 144 |
+
agreed to in writing, Licensor provides the Work (and each
|
| 145 |
+
Contributor provides its Contributions) on an "AS IS" BASIS,
|
| 146 |
+
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
|
| 147 |
+
implied, including, without limitation, any warranties or conditions
|
| 148 |
+
of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A
|
| 149 |
+
PARTICULAR PURPOSE. You are solely responsible for determining the
|
| 150 |
+
appropriateness of using or redistributing the Work and assume any
|
| 151 |
+
risks associated with Your exercise of permissions under this License.
|
| 152 |
+
|
| 153 |
+
8. Limitation of Liability. In no event and under no legal theory,
|
| 154 |
+
whether in tort (including negligence), contract, or otherwise,
|
| 155 |
+
unless required by applicable law (such as deliberate and grossly
|
| 156 |
+
negligent acts) or agreed to in writing, shall any Contributor be
|
| 157 |
+
liable to You for damages, including any direct, indirect, special,
|
| 158 |
+
incidental, or consequential damages of any character arising as a
|
| 159 |
+
result of this License or out of the use or inability to use the
|
| 160 |
+
Work (including but not limited to damages for loss of goodwill,
|
| 161 |
+
work stoppage, computer failure or malfunction, or any and all
|
| 162 |
+
other commercial damages or losses), even if such Contributor
|
| 163 |
+
has been advised of the possibility of such damages.
|
| 164 |
+
|
| 165 |
+
9. Accepting Warranty or Additional Liability. While redistributing
|
| 166 |
+
the Work or Derivative Works thereof, You may choose to offer,
|
| 167 |
+
and charge a fee for, acceptance of support, warranty, indemnity,
|
| 168 |
+
or other liability obligations and/or rights consistent with this
|
| 169 |
+
License. However, in accepting such obligations, You may act only
|
| 170 |
+
on Your own behalf and on Your sole responsibility, not on behalf
|
| 171 |
+
of any other Contributor, and only if You agree to indemnify,
|
| 172 |
+
defend, and hold each Contributor harmless for any liability
|
| 173 |
+
incurred by, or claims asserted against, such Contributor by reason
|
| 174 |
+
of your accepting any such warranty or additional liability.
|
| 175 |
+
|
| 176 |
+
END OF TERMS AND CONDITIONS
|
| 177 |
+
|
| 178 |
+
APPENDIX: How to apply the Apache License to your work.
|
| 179 |
+
|
| 180 |
+
To apply the Apache License to your work, attach the following
|
| 181 |
+
boilerplate notice, with the fields enclosed by brackets "[]"
|
| 182 |
+
replaced with your own identifying information. (Don't include
|
| 183 |
+
the brackets!) The text should be enclosed in the appropriate
|
| 184 |
+
comment syntax for the file format. We also recommend that a
|
| 185 |
+
file or class name and description of purpose be included on the
|
| 186 |
+
same "printed page" as the copyright notice for easier
|
| 187 |
+
identification within third-party archives.
|
| 188 |
+
|
| 189 |
+
Copyright [yyyy] [name of copyright owner]
|
| 190 |
+
|
| 191 |
+
Licensed under the Apache License, Version 2.0 (the "License");
|
| 192 |
+
you may not use this file except in compliance with the License.
|
| 193 |
+
You may obtain a copy of the License at
|
| 194 |
+
|
| 195 |
+
http://www.apache.org/licenses/LICENSE-2.0
|
| 196 |
+
|
| 197 |
+
Unless required by applicable law or agreed to in writing, software
|
| 198 |
+
distributed under the License is distributed on an "AS IS" BASIS,
|
| 199 |
+
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
| 200 |
+
See the License for the specific language governing permissions and
|
| 201 |
+
limitations under the License.
|
pythonProject/.venv/Lib/site-packages/huggingface_hub-0.34.4.dist-info/METADATA
ADDED
|
@@ -0,0 +1,331 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Metadata-Version: 2.1
|
| 2 |
+
Name: huggingface-hub
|
| 3 |
+
Version: 0.34.4
|
| 4 |
+
Summary: Client library to download and publish models, datasets and other repos on the huggingface.co hub
|
| 5 |
+
Home-page: https://github.com/huggingface/huggingface_hub
|
| 6 |
+
Author: Hugging Face, Inc.
|
| 7 |
+
Author-email: julien@huggingface.co
|
| 8 |
+
License: Apache
|
| 9 |
+
Keywords: model-hub machine-learning models natural-language-processing deep-learning pytorch pretrained-models
|
| 10 |
+
Platform: UNKNOWN
|
| 11 |
+
Classifier: Intended Audience :: Developers
|
| 12 |
+
Classifier: Intended Audience :: Education
|
| 13 |
+
Classifier: Intended Audience :: Science/Research
|
| 14 |
+
Classifier: License :: OSI Approved :: Apache Software License
|
| 15 |
+
Classifier: Operating System :: OS Independent
|
| 16 |
+
Classifier: Programming Language :: Python :: 3
|
| 17 |
+
Classifier: Programming Language :: Python :: 3 :: Only
|
| 18 |
+
Classifier: Programming Language :: Python :: 3.8
|
| 19 |
+
Classifier: Programming Language :: Python :: 3.9
|
| 20 |
+
Classifier: Programming Language :: Python :: 3.10
|
| 21 |
+
Classifier: Programming Language :: Python :: 3.11
|
| 22 |
+
Classifier: Programming Language :: Python :: 3.12
|
| 23 |
+
Classifier: Programming Language :: Python :: 3.13
|
| 24 |
+
Classifier: Topic :: Scientific/Engineering :: Artificial Intelligence
|
| 25 |
+
Requires-Python: >=3.8.0
|
| 26 |
+
Description-Content-Type: text/markdown
|
| 27 |
+
License-File: LICENSE
|
| 28 |
+
Requires-Dist: filelock
|
| 29 |
+
Requires-Dist: fsspec>=2023.5.0
|
| 30 |
+
Requires-Dist: packaging>=20.9
|
| 31 |
+
Requires-Dist: pyyaml>=5.1
|
| 32 |
+
Requires-Dist: requests
|
| 33 |
+
Requires-Dist: tqdm>=4.42.1
|
| 34 |
+
Requires-Dist: typing-extensions>=3.7.4.3
|
| 35 |
+
Requires-Dist: hf-xet<2.0.0,>=1.1.3; platform_machine == "x86_64" or platform_machine == "amd64" or platform_machine == "arm64" or platform_machine == "aarch64"
|
| 36 |
+
Provides-Extra: all
|
| 37 |
+
Requires-Dist: InquirerPy==0.3.4; extra == "all"
|
| 38 |
+
Requires-Dist: aiohttp; extra == "all"
|
| 39 |
+
Requires-Dist: authlib>=1.3.2; extra == "all"
|
| 40 |
+
Requires-Dist: fastapi; extra == "all"
|
| 41 |
+
Requires-Dist: httpx; extra == "all"
|
| 42 |
+
Requires-Dist: itsdangerous; extra == "all"
|
| 43 |
+
Requires-Dist: jedi; extra == "all"
|
| 44 |
+
Requires-Dist: Jinja2; extra == "all"
|
| 45 |
+
Requires-Dist: pytest<8.2.2,>=8.1.1; extra == "all"
|
| 46 |
+
Requires-Dist: pytest-cov; extra == "all"
|
| 47 |
+
Requires-Dist: pytest-env; extra == "all"
|
| 48 |
+
Requires-Dist: pytest-xdist; extra == "all"
|
| 49 |
+
Requires-Dist: pytest-vcr; extra == "all"
|
| 50 |
+
Requires-Dist: pytest-asyncio; extra == "all"
|
| 51 |
+
Requires-Dist: pytest-rerunfailures; extra == "all"
|
| 52 |
+
Requires-Dist: pytest-mock; extra == "all"
|
| 53 |
+
Requires-Dist: urllib3<2.0; extra == "all"
|
| 54 |
+
Requires-Dist: soundfile; extra == "all"
|
| 55 |
+
Requires-Dist: Pillow; extra == "all"
|
| 56 |
+
Requires-Dist: gradio>=4.0.0; extra == "all"
|
| 57 |
+
Requires-Dist: numpy; extra == "all"
|
| 58 |
+
Requires-Dist: ruff>=0.9.0; extra == "all"
|
| 59 |
+
Requires-Dist: libcst>=1.4.0; extra == "all"
|
| 60 |
+
Requires-Dist: typing-extensions>=4.8.0; extra == "all"
|
| 61 |
+
Requires-Dist: types-PyYAML; extra == "all"
|
| 62 |
+
Requires-Dist: types-requests; extra == "all"
|
| 63 |
+
Requires-Dist: types-simplejson; extra == "all"
|
| 64 |
+
Requires-Dist: types-toml; extra == "all"
|
| 65 |
+
Requires-Dist: types-tqdm; extra == "all"
|
| 66 |
+
Requires-Dist: types-urllib3; extra == "all"
|
| 67 |
+
Requires-Dist: mypy<1.15.0,>=1.14.1; python_version == "3.8" and extra == "all"
|
| 68 |
+
Requires-Dist: mypy==1.15.0; python_version >= "3.9" and extra == "all"
|
| 69 |
+
Provides-Extra: cli
|
| 70 |
+
Requires-Dist: InquirerPy==0.3.4; extra == "cli"
|
| 71 |
+
Provides-Extra: dev
|
| 72 |
+
Requires-Dist: InquirerPy==0.3.4; extra == "dev"
|
| 73 |
+
Requires-Dist: aiohttp; extra == "dev"
|
| 74 |
+
Requires-Dist: authlib>=1.3.2; extra == "dev"
|
| 75 |
+
Requires-Dist: fastapi; extra == "dev"
|
| 76 |
+
Requires-Dist: httpx; extra == "dev"
|
| 77 |
+
Requires-Dist: itsdangerous; extra == "dev"
|
| 78 |
+
Requires-Dist: jedi; extra == "dev"
|
| 79 |
+
Requires-Dist: Jinja2; extra == "dev"
|
| 80 |
+
Requires-Dist: pytest<8.2.2,>=8.1.1; extra == "dev"
|
| 81 |
+
Requires-Dist: pytest-cov; extra == "dev"
|
| 82 |
+
Requires-Dist: pytest-env; extra == "dev"
|
| 83 |
+
Requires-Dist: pytest-xdist; extra == "dev"
|
| 84 |
+
Requires-Dist: pytest-vcr; extra == "dev"
|
| 85 |
+
Requires-Dist: pytest-asyncio; extra == "dev"
|
| 86 |
+
Requires-Dist: pytest-rerunfailures; extra == "dev"
|
| 87 |
+
Requires-Dist: pytest-mock; extra == "dev"
|
| 88 |
+
Requires-Dist: urllib3<2.0; extra == "dev"
|
| 89 |
+
Requires-Dist: soundfile; extra == "dev"
|
| 90 |
+
Requires-Dist: Pillow; extra == "dev"
|
| 91 |
+
Requires-Dist: gradio>=4.0.0; extra == "dev"
|
| 92 |
+
Requires-Dist: numpy; extra == "dev"
|
| 93 |
+
Requires-Dist: ruff>=0.9.0; extra == "dev"
|
| 94 |
+
Requires-Dist: libcst>=1.4.0; extra == "dev"
|
| 95 |
+
Requires-Dist: typing-extensions>=4.8.0; extra == "dev"
|
| 96 |
+
Requires-Dist: types-PyYAML; extra == "dev"
|
| 97 |
+
Requires-Dist: types-requests; extra == "dev"
|
| 98 |
+
Requires-Dist: types-simplejson; extra == "dev"
|
| 99 |
+
Requires-Dist: types-toml; extra == "dev"
|
| 100 |
+
Requires-Dist: types-tqdm; extra == "dev"
|
| 101 |
+
Requires-Dist: types-urllib3; extra == "dev"
|
| 102 |
+
Requires-Dist: mypy<1.15.0,>=1.14.1; python_version == "3.8" and extra == "dev"
|
| 103 |
+
Requires-Dist: mypy==1.15.0; python_version >= "3.9" and extra == "dev"
|
| 104 |
+
Provides-Extra: fastai
|
| 105 |
+
Requires-Dist: toml; extra == "fastai"
|
| 106 |
+
Requires-Dist: fastai>=2.4; extra == "fastai"
|
| 107 |
+
Requires-Dist: fastcore>=1.3.27; extra == "fastai"
|
| 108 |
+
Provides-Extra: hf_transfer
|
| 109 |
+
Requires-Dist: hf-transfer>=0.1.4; extra == "hf-transfer"
|
| 110 |
+
Provides-Extra: hf_xet
|
| 111 |
+
Requires-Dist: hf-xet<2.0.0,>=1.1.2; extra == "hf-xet"
|
| 112 |
+
Provides-Extra: inference
|
| 113 |
+
Requires-Dist: aiohttp; extra == "inference"
|
| 114 |
+
Provides-Extra: mcp
|
| 115 |
+
Requires-Dist: mcp>=1.8.0; extra == "mcp"
|
| 116 |
+
Requires-Dist: typer; extra == "mcp"
|
| 117 |
+
Requires-Dist: aiohttp; extra == "mcp"
|
| 118 |
+
Provides-Extra: oauth
|
| 119 |
+
Requires-Dist: authlib>=1.3.2; extra == "oauth"
|
| 120 |
+
Requires-Dist: fastapi; extra == "oauth"
|
| 121 |
+
Requires-Dist: httpx; extra == "oauth"
|
| 122 |
+
Requires-Dist: itsdangerous; extra == "oauth"
|
| 123 |
+
Provides-Extra: quality
|
| 124 |
+
Requires-Dist: ruff>=0.9.0; extra == "quality"
|
| 125 |
+
Requires-Dist: libcst>=1.4.0; extra == "quality"
|
| 126 |
+
Requires-Dist: mypy<1.15.0,>=1.14.1; python_version == "3.8" and extra == "quality"
|
| 127 |
+
Requires-Dist: mypy==1.15.0; python_version >= "3.9" and extra == "quality"
|
| 128 |
+
Provides-Extra: tensorflow
|
| 129 |
+
Requires-Dist: tensorflow; extra == "tensorflow"
|
| 130 |
+
Requires-Dist: pydot; extra == "tensorflow"
|
| 131 |
+
Requires-Dist: graphviz; extra == "tensorflow"
|
| 132 |
+
Provides-Extra: tensorflow-testing
|
| 133 |
+
Requires-Dist: tensorflow; extra == "tensorflow-testing"
|
| 134 |
+
Requires-Dist: keras<3.0; extra == "tensorflow-testing"
|
| 135 |
+
Provides-Extra: testing
|
| 136 |
+
Requires-Dist: InquirerPy==0.3.4; extra == "testing"
|
| 137 |
+
Requires-Dist: aiohttp; extra == "testing"
|
| 138 |
+
Requires-Dist: authlib>=1.3.2; extra == "testing"
|
| 139 |
+
Requires-Dist: fastapi; extra == "testing"
|
| 140 |
+
Requires-Dist: httpx; extra == "testing"
|
| 141 |
+
Requires-Dist: itsdangerous; extra == "testing"
|
| 142 |
+
Requires-Dist: jedi; extra == "testing"
|
| 143 |
+
Requires-Dist: Jinja2; extra == "testing"
|
| 144 |
+
Requires-Dist: pytest<8.2.2,>=8.1.1; extra == "testing"
|
| 145 |
+
Requires-Dist: pytest-cov; extra == "testing"
|
| 146 |
+
Requires-Dist: pytest-env; extra == "testing"
|
| 147 |
+
Requires-Dist: pytest-xdist; extra == "testing"
|
| 148 |
+
Requires-Dist: pytest-vcr; extra == "testing"
|
| 149 |
+
Requires-Dist: pytest-asyncio; extra == "testing"
|
| 150 |
+
Requires-Dist: pytest-rerunfailures; extra == "testing"
|
| 151 |
+
Requires-Dist: pytest-mock; extra == "testing"
|
| 152 |
+
Requires-Dist: urllib3<2.0; extra == "testing"
|
| 153 |
+
Requires-Dist: soundfile; extra == "testing"
|
| 154 |
+
Requires-Dist: Pillow; extra == "testing"
|
| 155 |
+
Requires-Dist: gradio>=4.0.0; extra == "testing"
|
| 156 |
+
Requires-Dist: numpy; extra == "testing"
|
| 157 |
+
Provides-Extra: torch
|
| 158 |
+
Requires-Dist: torch; extra == "torch"
|
| 159 |
+
Requires-Dist: safetensors[torch]; extra == "torch"
|
| 160 |
+
Provides-Extra: typing
|
| 161 |
+
Requires-Dist: typing-extensions>=4.8.0; extra == "typing"
|
| 162 |
+
Requires-Dist: types-PyYAML; extra == "typing"
|
| 163 |
+
Requires-Dist: types-requests; extra == "typing"
|
| 164 |
+
Requires-Dist: types-simplejson; extra == "typing"
|
| 165 |
+
Requires-Dist: types-toml; extra == "typing"
|
| 166 |
+
Requires-Dist: types-tqdm; extra == "typing"
|
| 167 |
+
Requires-Dist: types-urllib3; extra == "typing"
|
| 168 |
+
|
| 169 |
+
<p align="center">
|
| 170 |
+
<picture>
|
| 171 |
+
<source media="(prefers-color-scheme: dark)" srcset="https://huggingface.co/datasets/huggingface/documentation-images/raw/main/huggingface_hub-dark.svg">
|
| 172 |
+
<source media="(prefers-color-scheme: light)" srcset="https://huggingface.co/datasets/huggingface/documentation-images/raw/main/huggingface_hub.svg">
|
| 173 |
+
<img alt="huggingface_hub library logo" src="https://huggingface.co/datasets/huggingface/documentation-images/raw/main/huggingface_hub.svg" width="352" height="59" style="max-width: 100%;">
|
| 174 |
+
</picture>
|
| 175 |
+
<br/>
|
| 176 |
+
<br/>
|
| 177 |
+
</p>
|
| 178 |
+
|
| 179 |
+
<p align="center">
|
| 180 |
+
<i>The official Python client for the Huggingface Hub.</i>
|
| 181 |
+
</p>
|
| 182 |
+
|
| 183 |
+
<p align="center">
|
| 184 |
+
<a href="https://huggingface.co/docs/huggingface_hub/en/index"><img alt="Documentation" src="https://img.shields.io/website/http/huggingface.co/docs/huggingface_hub/index.svg?down_color=red&down_message=offline&up_message=online&label=doc"></a>
|
| 185 |
+
<a href="https://github.com/huggingface/huggingface_hub/releases"><img alt="GitHub release" src="https://img.shields.io/github/release/huggingface/huggingface_hub.svg"></a>
|
| 186 |
+
<a href="https://github.com/huggingface/huggingface_hub"><img alt="PyPi version" src="https://img.shields.io/pypi/pyversions/huggingface_hub.svg"></a>
|
| 187 |
+
<a href="https://pypi.org/project/huggingface-hub"><img alt="PyPI - Downloads" src="https://img.shields.io/pypi/dm/huggingface_hub"></a>
|
| 188 |
+
<a href="https://codecov.io/gh/huggingface/huggingface_hub"><img alt="Code coverage" src="https://codecov.io/gh/huggingface/huggingface_hub/branch/main/graph/badge.svg?token=RXP95LE2XL"></a>
|
| 189 |
+
</p>
|
| 190 |
+
|
| 191 |
+
<h4 align="center">
|
| 192 |
+
<p>
|
| 193 |
+
<b>English</b> |
|
| 194 |
+
<a href="https://github.com/huggingface/huggingface_hub/blob/main/i18n/README_de.md">Deutsch</a> |
|
| 195 |
+
<a href="https://github.com/huggingface/huggingface_hub/blob/main/i18n/README_hi.md">हिंदी</a> |
|
| 196 |
+
<a href="https://github.com/huggingface/huggingface_hub/blob/main/i18n/README_ko.md">한국어</a> |
|
| 197 |
+
<a href="https://github.com/huggingface/huggingface_hub/blob/main/i18n/README_cn.md">中文(简体)</a>
|
| 198 |
+
<p>
|
| 199 |
+
</h4>
|
| 200 |
+
|
| 201 |
+
---
|
| 202 |
+
|
| 203 |
+
**Documentation**: <a href="https://hf.co/docs/huggingface_hub" target="_blank">https://hf.co/docs/huggingface_hub</a>
|
| 204 |
+
|
| 205 |
+
**Source Code**: <a href="https://github.com/huggingface/huggingface_hub" target="_blank">https://github.com/huggingface/huggingface_hub</a>
|
| 206 |
+
|
| 207 |
+
---
|
| 208 |
+
|
| 209 |
+
## Welcome to the huggingface_hub library
|
| 210 |
+
|
| 211 |
+
The `huggingface_hub` library allows you to interact with the [Hugging Face Hub](https://huggingface.co/), a platform democratizing open-source Machine Learning for creators and collaborators. Discover pre-trained models and datasets for your projects or play with the thousands of machine learning apps hosted on the Hub. You can also create and share your own models, datasets and demos with the community. The `huggingface_hub` library provides a simple way to do all these things with Python.
|
| 212 |
+
|
| 213 |
+
## Key features
|
| 214 |
+
|
| 215 |
+
- [Download files](https://huggingface.co/docs/huggingface_hub/en/guides/download) from the Hub.
|
| 216 |
+
- [Upload files](https://huggingface.co/docs/huggingface_hub/en/guides/upload) to the Hub.
|
| 217 |
+
- [Manage your repositories](https://huggingface.co/docs/huggingface_hub/en/guides/repository).
|
| 218 |
+
- [Run Inference](https://huggingface.co/docs/huggingface_hub/en/guides/inference) on deployed models.
|
| 219 |
+
- [Search](https://huggingface.co/docs/huggingface_hub/en/guides/search) for models, datasets and Spaces.
|
| 220 |
+
- [Share Model Cards](https://huggingface.co/docs/huggingface_hub/en/guides/model-cards) to document your models.
|
| 221 |
+
- [Engage with the community](https://huggingface.co/docs/huggingface_hub/en/guides/community) through PRs and comments.
|
| 222 |
+
|
| 223 |
+
## Installation
|
| 224 |
+
|
| 225 |
+
Install the `huggingface_hub` package with [pip](https://pypi.org/project/huggingface-hub/):
|
| 226 |
+
|
| 227 |
+
```bash
|
| 228 |
+
pip install huggingface_hub
|
| 229 |
+
```
|
| 230 |
+
|
| 231 |
+
If you prefer, you can also install it with [conda](https://huggingface.co/docs/huggingface_hub/en/installation#install-with-conda).
|
| 232 |
+
|
| 233 |
+
In order to keep the package minimal by default, `huggingface_hub` comes with optional dependencies useful for some use cases. For example, if you want have a complete experience for Inference, run:
|
| 234 |
+
|
| 235 |
+
```bash
|
| 236 |
+
pip install huggingface_hub[inference]
|
| 237 |
+
```
|
| 238 |
+
|
| 239 |
+
To learn more installation and optional dependencies, check out the [installation guide](https://huggingface.co/docs/huggingface_hub/en/installation).
|
| 240 |
+
|
| 241 |
+
## Quick start
|
| 242 |
+
|
| 243 |
+
### Download files
|
| 244 |
+
|
| 245 |
+
Download a single file
|
| 246 |
+
|
| 247 |
+
```py
|
| 248 |
+
from huggingface_hub import hf_hub_download
|
| 249 |
+
|
| 250 |
+
hf_hub_download(repo_id="tiiuae/falcon-7b-instruct", filename="config.json")
|
| 251 |
+
```
|
| 252 |
+
|
| 253 |
+
Or an entire repository
|
| 254 |
+
|
| 255 |
+
```py
|
| 256 |
+
from huggingface_hub import snapshot_download
|
| 257 |
+
|
| 258 |
+
snapshot_download("stabilityai/stable-diffusion-2-1")
|
| 259 |
+
```
|
| 260 |
+
|
| 261 |
+
Files will be downloaded in a local cache folder. More details in [this guide](https://huggingface.co/docs/huggingface_hub/en/guides/manage-cache).
|
| 262 |
+
|
| 263 |
+
### Login
|
| 264 |
+
|
| 265 |
+
The Hugging Face Hub uses tokens to authenticate applications (see [docs](https://huggingface.co/docs/hub/security-tokens)). To log in your machine, run the following CLI:
|
| 266 |
+
|
| 267 |
+
```bash
|
| 268 |
+
hf auth login
|
| 269 |
+
# or using an environment variable
|
| 270 |
+
hf auth login --token $HUGGINGFACE_TOKEN
|
| 271 |
+
```
|
| 272 |
+
|
| 273 |
+
### Create a repository
|
| 274 |
+
|
| 275 |
+
```py
|
| 276 |
+
from huggingface_hub import create_repo
|
| 277 |
+
|
| 278 |
+
create_repo(repo_id="super-cool-model")
|
| 279 |
+
```
|
| 280 |
+
|
| 281 |
+
### Upload files
|
| 282 |
+
|
| 283 |
+
Upload a single file
|
| 284 |
+
|
| 285 |
+
```py
|
| 286 |
+
from huggingface_hub import upload_file
|
| 287 |
+
|
| 288 |
+
upload_file(
|
| 289 |
+
path_or_fileobj="/home/lysandre/dummy-test/README.md",
|
| 290 |
+
path_in_repo="README.md",
|
| 291 |
+
repo_id="lysandre/test-model",
|
| 292 |
+
)
|
| 293 |
+
```
|
| 294 |
+
|
| 295 |
+
Or an entire folder
|
| 296 |
+
|
| 297 |
+
```py
|
| 298 |
+
from huggingface_hub import upload_folder
|
| 299 |
+
|
| 300 |
+
upload_folder(
|
| 301 |
+
folder_path="/path/to/local/space",
|
| 302 |
+
repo_id="username/my-cool-space",
|
| 303 |
+
repo_type="space",
|
| 304 |
+
)
|
| 305 |
+
```
|
| 306 |
+
|
| 307 |
+
For details in the [upload guide](https://huggingface.co/docs/huggingface_hub/en/guides/upload).
|
| 308 |
+
|
| 309 |
+
## Integrating to the Hub.
|
| 310 |
+
|
| 311 |
+
We're partnering with cool open source ML libraries to provide free model hosting and versioning. You can find the existing integrations [here](https://huggingface.co/docs/hub/libraries).
|
| 312 |
+
|
| 313 |
+
The advantages are:
|
| 314 |
+
|
| 315 |
+
- Free model or dataset hosting for libraries and their users.
|
| 316 |
+
- Built-in file versioning, even with very large files, thanks to a git-based approach.
|
| 317 |
+
- In-browser widgets to play with the uploaded models.
|
| 318 |
+
- Anyone can upload a new model for your library, they just need to add the corresponding tag for the model to be discoverable.
|
| 319 |
+
- Fast downloads! We use Cloudfront (a CDN) to geo-replicate downloads so they're blazing fast from anywhere on the globe.
|
| 320 |
+
- Usage stats and more features to come.
|
| 321 |
+
|
| 322 |
+
If you would like to integrate your library, feel free to open an issue to begin the discussion. We wrote a [step-by-step guide](https://huggingface.co/docs/hub/adding-a-library) with ❤️ showing how to do this integration.
|
| 323 |
+
|
| 324 |
+
## Contributions (feature requests, bugs, etc.) are super welcome 💙💚💛💜🧡❤️
|
| 325 |
+
|
| 326 |
+
Everyone is welcome to contribute, and we value everybody's contribution. Code is not the only way to help the community.
|
| 327 |
+
Answering questions, helping others, reaching out and improving the documentations are immensely valuable to the community.
|
| 328 |
+
We wrote a [contribution guide](https://github.com/huggingface/huggingface_hub/blob/main/CONTRIBUTING.md) to summarize
|
| 329 |
+
how to get started to contribute to this repository.
|
| 330 |
+
|
| 331 |
+
|
pythonProject/.venv/Lib/site-packages/huggingface_hub-0.34.4.dist-info/RECORD
ADDED
|
@@ -0,0 +1,327 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
../../Scripts/hf.exe,sha256=StSqlEsqGFWFge63Op8-BJoAHylmTBIDpSpvw0PHNhM,108411
|
| 2 |
+
../../Scripts/huggingface-cli.exe,sha256=j98QwPmWkWMT3e4FJeVmL4Q23uJQPEkPlL6wJhvNaqI,108429
|
| 3 |
+
../../Scripts/tiny-agents.exe,sha256=z7vLUru4du9IUGHzHWLsdzpRcm0wVAmgFHqra06vKp8,108421
|
| 4 |
+
huggingface_hub-0.34.4.dist-info/INSTALLER,sha256=zuuue4knoyJ-UwPPXg8fezS7VCrXJQrAP7zeNuwvFQg,4
|
| 5 |
+
huggingface_hub-0.34.4.dist-info/LICENSE,sha256=xx0jnfkXJvxRnG63LTGOxlggYnIysveWIZ6H3PNdCrQ,11357
|
| 6 |
+
huggingface_hub-0.34.4.dist-info/METADATA,sha256=sEy_vJu9uxm2AvxvoG49eqdJxzxDYRZPOpeBr4jKhkM,14699
|
| 7 |
+
huggingface_hub-0.34.4.dist-info/RECORD,,
|
| 8 |
+
huggingface_hub-0.34.4.dist-info/WHEEL,sha256=tZoeGjtWxWRfdplE7E3d45VPlLNQnvbKiYnx7gwAy8A,92
|
| 9 |
+
huggingface_hub-0.34.4.dist-info/entry_points.txt,sha256=HIzLhjwPTO7U_ncpW4AkmzAuaadr1ajmYagW5mdb5TM,217
|
| 10 |
+
huggingface_hub-0.34.4.dist-info/top_level.txt,sha256=8KzlQJAY4miUvjAssOAJodqKOw3harNzuiwGQ9qLSSk,16
|
| 11 |
+
huggingface_hub/__init__.py,sha256=MsUqTc5DczrwOwtEhkM9Y7tkGAlHvDtNTDbWQcEEqTA,51837
|
| 12 |
+
huggingface_hub/__pycache__/__init__.cpython-310.pyc,,
|
| 13 |
+
huggingface_hub/__pycache__/_commit_api.cpython-310.pyc,,
|
| 14 |
+
huggingface_hub/__pycache__/_commit_scheduler.cpython-310.pyc,,
|
| 15 |
+
huggingface_hub/__pycache__/_inference_endpoints.cpython-310.pyc,,
|
| 16 |
+
huggingface_hub/__pycache__/_jobs_api.cpython-310.pyc,,
|
| 17 |
+
huggingface_hub/__pycache__/_local_folder.cpython-310.pyc,,
|
| 18 |
+
huggingface_hub/__pycache__/_login.cpython-310.pyc,,
|
| 19 |
+
huggingface_hub/__pycache__/_oauth.cpython-310.pyc,,
|
| 20 |
+
huggingface_hub/__pycache__/_snapshot_download.cpython-310.pyc,,
|
| 21 |
+
huggingface_hub/__pycache__/_space_api.cpython-310.pyc,,
|
| 22 |
+
huggingface_hub/__pycache__/_tensorboard_logger.cpython-310.pyc,,
|
| 23 |
+
huggingface_hub/__pycache__/_upload_large_folder.cpython-310.pyc,,
|
| 24 |
+
huggingface_hub/__pycache__/_webhooks_payload.cpython-310.pyc,,
|
| 25 |
+
huggingface_hub/__pycache__/_webhooks_server.cpython-310.pyc,,
|
| 26 |
+
huggingface_hub/__pycache__/community.cpython-310.pyc,,
|
| 27 |
+
huggingface_hub/__pycache__/constants.cpython-310.pyc,,
|
| 28 |
+
huggingface_hub/__pycache__/dataclasses.cpython-310.pyc,,
|
| 29 |
+
huggingface_hub/__pycache__/errors.cpython-310.pyc,,
|
| 30 |
+
huggingface_hub/__pycache__/fastai_utils.cpython-310.pyc,,
|
| 31 |
+
huggingface_hub/__pycache__/file_download.cpython-310.pyc,,
|
| 32 |
+
huggingface_hub/__pycache__/hf_api.cpython-310.pyc,,
|
| 33 |
+
huggingface_hub/__pycache__/hf_file_system.cpython-310.pyc,,
|
| 34 |
+
huggingface_hub/__pycache__/hub_mixin.cpython-310.pyc,,
|
| 35 |
+
huggingface_hub/__pycache__/inference_api.cpython-310.pyc,,
|
| 36 |
+
huggingface_hub/__pycache__/keras_mixin.cpython-310.pyc,,
|
| 37 |
+
huggingface_hub/__pycache__/lfs.cpython-310.pyc,,
|
| 38 |
+
huggingface_hub/__pycache__/repocard.cpython-310.pyc,,
|
| 39 |
+
huggingface_hub/__pycache__/repocard_data.cpython-310.pyc,,
|
| 40 |
+
huggingface_hub/__pycache__/repository.cpython-310.pyc,,
|
| 41 |
+
huggingface_hub/_commit_api.py,sha256=68HxFnJE2s-QmGZRHQav5kOMTseYV_ZQi04ADaQmZUk,38979
|
| 42 |
+
huggingface_hub/_commit_scheduler.py,sha256=tfIoO1xWHjTJ6qy6VS6HIoymDycFPg0d6pBSZprrU2U,14679
|
| 43 |
+
huggingface_hub/_inference_endpoints.py,sha256=ahmbPcEXsJ_JcMb9TDgdkD8Z2z9uytkFG3_1o6dTm8g,17598
|
| 44 |
+
huggingface_hub/_jobs_api.py,sha256=Fm59WJ-P3E5mkCthQWgGugTlbXnQe4nxDKIwZ_jtpt8,5344
|
| 45 |
+
huggingface_hub/_local_folder.py,sha256=2iHXNgIT3UdSt2PvCovd0NzgVxTRypKb-rvAFLK-gZU,17305
|
| 46 |
+
huggingface_hub/_login.py,sha256=rcwx9EZdFUB3vuowC5QBiSYS4ImUnBzo04igR1Z8l40,20256
|
| 47 |
+
huggingface_hub/_oauth.py,sha256=75ya9toHxC0WRKsLOAI212CrssRjTSxs16mHWWNMb3w,18714
|
| 48 |
+
huggingface_hub/_snapshot_download.py,sha256=b-NzYQcvktsAirIfGQKgzQwu8w0S6lhBTvnJ5S6saw8,16166
|
| 49 |
+
huggingface_hub/_space_api.py,sha256=jb6rF8qLtjaNU12D-8ygAPM26xDiHCu8CHXHowhGTmg,5470
|
| 50 |
+
huggingface_hub/_tensorboard_logger.py,sha256=ZkYcAUiRC8RGL214QUYtp58O8G5tn-HF6DCWha9imcA,8358
|
| 51 |
+
huggingface_hub/_upload_large_folder.py,sha256=l2YWLZttOw69EGdihT3y_Nhr5mweLGooZG9L8smNoHY,30066
|
| 52 |
+
huggingface_hub/_webhooks_payload.py,sha256=Xm3KaK7tCOGBlXkuZvbym6zjHXrT1XCrbUFWuXiBmNY,3617
|
| 53 |
+
huggingface_hub/_webhooks_server.py,sha256=5J63wk9MUGKBNJVsOD9i60mJ-VMp0YYmlf87vQsl-L8,15767
|
| 54 |
+
huggingface_hub/cli/__init__.py,sha256=xzX1qgAvrtAX4gP59WrPlvOZFLuzuTgcjvanQvcpgHc,928
|
| 55 |
+
huggingface_hub/cli/__pycache__/__init__.cpython-310.pyc,,
|
| 56 |
+
huggingface_hub/cli/__pycache__/_cli_utils.cpython-310.pyc,,
|
| 57 |
+
huggingface_hub/cli/__pycache__/auth.cpython-310.pyc,,
|
| 58 |
+
huggingface_hub/cli/__pycache__/cache.cpython-310.pyc,,
|
| 59 |
+
huggingface_hub/cli/__pycache__/download.cpython-310.pyc,,
|
| 60 |
+
huggingface_hub/cli/__pycache__/hf.cpython-310.pyc,,
|
| 61 |
+
huggingface_hub/cli/__pycache__/jobs.cpython-310.pyc,,
|
| 62 |
+
huggingface_hub/cli/__pycache__/lfs.cpython-310.pyc,,
|
| 63 |
+
huggingface_hub/cli/__pycache__/repo.cpython-310.pyc,,
|
| 64 |
+
huggingface_hub/cli/__pycache__/repo_files.cpython-310.pyc,,
|
| 65 |
+
huggingface_hub/cli/__pycache__/system.cpython-310.pyc,,
|
| 66 |
+
huggingface_hub/cli/__pycache__/upload.cpython-310.pyc,,
|
| 67 |
+
huggingface_hub/cli/__pycache__/upload_large_folder.cpython-310.pyc,,
|
| 68 |
+
huggingface_hub/cli/_cli_utils.py,sha256=Nt6CjbkYqQQRuh70bUXVA6rZpbZt_Sa1WqBUxjQLu6g,2095
|
| 69 |
+
huggingface_hub/cli/auth.py,sha256=BO6sJJcdHhjouMEH5JpUmC0qg3vaukX4I5DtA7ohLes,7296
|
| 70 |
+
huggingface_hub/cli/cache.py,sha256=zPLEWP5MidLElo0FeLvkUjCT0EYlX1pmC-2TkCa8kD0,15871
|
| 71 |
+
huggingface_hub/cli/download.py,sha256=PUpW-nbu6ZAP6P9DpVhliAKSSlxvXWkVh0U2KZoukhQ,7115
|
| 72 |
+
huggingface_hub/cli/hf.py,sha256=SQ73_SXEQnWVJkhKT_6bwNQBHQXGOdI5qqlTTtI0XH0,2328
|
| 73 |
+
huggingface_hub/cli/jobs.py,sha256=QPF8YT31hoyjXXp_pD-LK_VHNFXF09su-nLeP4b9WGQ,21538
|
| 74 |
+
huggingface_hub/cli/lfs.py,sha256=J9MkKOGUW6GjBrKs2zZUCOaAGxpatxsEoSbBjuhDJV8,7230
|
| 75 |
+
huggingface_hub/cli/repo.py,sha256=lNDEZbOpLW8SQVBYDQ1xofw9nJ7M8AUsd2kBIV_m_do,10576
|
| 76 |
+
huggingface_hub/cli/repo_files.py,sha256=L-Ku52l2vZ04GCabp_OhVXqLzE9dsKQqaQKudGzjWg4,4831
|
| 77 |
+
huggingface_hub/cli/system.py,sha256=eLSYME7ywt5Ae3tYQnS43Tai2pR2JLtA1KGImzPt5pM,1707
|
| 78 |
+
huggingface_hub/cli/upload.py,sha256=qOGccIcBYJtodmlQlFyGfV_ZGHYhWlAARr3fxgYLDnE,14349
|
| 79 |
+
huggingface_hub/cli/upload_large_folder.py,sha256=dEb1EKbPi2nRpFS-gz4P0D49-Rlf26MAs6kgl7l2vJk,6151
|
| 80 |
+
huggingface_hub/commands/__init__.py,sha256=AkbM2a-iGh0Vq_xAWhK3mu3uZ44km8-X5uWjKcvcrUQ,928
|
| 81 |
+
huggingface_hub/commands/__pycache__/__init__.cpython-310.pyc,,
|
| 82 |
+
huggingface_hub/commands/__pycache__/_cli_utils.cpython-310.pyc,,
|
| 83 |
+
huggingface_hub/commands/__pycache__/delete_cache.cpython-310.pyc,,
|
| 84 |
+
huggingface_hub/commands/__pycache__/download.cpython-310.pyc,,
|
| 85 |
+
huggingface_hub/commands/__pycache__/env.cpython-310.pyc,,
|
| 86 |
+
huggingface_hub/commands/__pycache__/huggingface_cli.cpython-310.pyc,,
|
| 87 |
+
huggingface_hub/commands/__pycache__/lfs.cpython-310.pyc,,
|
| 88 |
+
huggingface_hub/commands/__pycache__/repo.cpython-310.pyc,,
|
| 89 |
+
huggingface_hub/commands/__pycache__/repo_files.cpython-310.pyc,,
|
| 90 |
+
huggingface_hub/commands/__pycache__/scan_cache.cpython-310.pyc,,
|
| 91 |
+
huggingface_hub/commands/__pycache__/tag.cpython-310.pyc,,
|
| 92 |
+
huggingface_hub/commands/__pycache__/upload.cpython-310.pyc,,
|
| 93 |
+
huggingface_hub/commands/__pycache__/upload_large_folder.cpython-310.pyc,,
|
| 94 |
+
huggingface_hub/commands/__pycache__/user.cpython-310.pyc,,
|
| 95 |
+
huggingface_hub/commands/__pycache__/version.cpython-310.pyc,,
|
| 96 |
+
huggingface_hub/commands/_cli_utils.py,sha256=ePYTIEWnU677nPvdNC5AdYcEB1400L6qYEUxMkVUzME,2329
|
| 97 |
+
huggingface_hub/commands/delete_cache.py,sha256=GpeDKs2I16rfxGnGX1widy9TBZSZM8PQ1qmHHRcwd-o,17734
|
| 98 |
+
huggingface_hub/commands/download.py,sha256=0QY9ho7eiAPvFndBPttGtH6vXNk3r9AioltNwc8h1Z4,8310
|
| 99 |
+
huggingface_hub/commands/env.py,sha256=qv4SmjuzUz9exo4RDMY2HqabLCKE1oRb55cBA6LN9R4,1342
|
| 100 |
+
huggingface_hub/commands/huggingface_cli.py,sha256=gDi7JueyiLD0bGclTEYfHPQWpAY_WBdPfHT7vkqa5v0,2654
|
| 101 |
+
huggingface_hub/commands/lfs.py,sha256=xdbnNRO04UuQemEhUGT809jFgQn9Rj-SnyT_0Ph-VYg,7342
|
| 102 |
+
huggingface_hub/commands/repo.py,sha256=WcRDFqUYKB0Kz0zFopegiG614ot6VOYTAf6jht0BMss,6042
|
| 103 |
+
huggingface_hub/commands/repo_files.py,sha256=ftjLCC3XCY-AMmiYiZPIdRMmIqZbqVZw-BSjBLcZup4,5054
|
| 104 |
+
huggingface_hub/commands/scan_cache.py,sha256=Er5y0paRv0feXnwxe-eAwuJA6xlZ-OdEQ1FbB80BqNU,8670
|
| 105 |
+
huggingface_hub/commands/tag.py,sha256=4fgQuXJHG59lTVyOjIUZjxdJDL4JZW4q10XDPSo-gss,6382
|
| 106 |
+
huggingface_hub/commands/upload.py,sha256=eAJIig4ljtO9FRyGjiz6HbHS-Q4MOQziRgzjQrl5Koo,14576
|
| 107 |
+
huggingface_hub/commands/upload_large_folder.py,sha256=_1id84BFtbL8HgFRKZ-el_uPrijamz1qWlzO16KbUAc,6254
|
| 108 |
+
huggingface_hub/commands/user.py,sha256=MjG1lwMq1p5QAlBolFnRX_pUxE3Kd3UiPl-nEEQSgXg,7537
|
| 109 |
+
huggingface_hub/commands/version.py,sha256=rGpCbvxImY9eQqXrshYt609Iws27R75WARmKQrIo6Ok,1390
|
| 110 |
+
huggingface_hub/community.py,sha256=4MtcoxEI9_0lmmilBEnvUEi8_O1Ivfa8p6eKxYU5-ts,12198
|
| 111 |
+
huggingface_hub/constants.py,sha256=nILseAp4rqLu_KQTZDpPGOhepVAPanD7azbomAvovj0,10313
|
| 112 |
+
huggingface_hub/dataclasses.py,sha256=sgPdEi2UDprhNPP2PPkiSlzsHdC1WcpwVTLwlHAEcr0,17224
|
| 113 |
+
huggingface_hub/errors.py,sha256=D7Lw0Jjrf8vfmD0B26LEvg-JWkU8Zq0KDPJOzFY4QLw,11201
|
| 114 |
+
huggingface_hub/fastai_utils.py,sha256=DpeH9d-6ut2k_nCAAwglM51XmRmgfbRe2SPifpVL5Yk,16745
|
| 115 |
+
huggingface_hub/file_download.py,sha256=E-NWON01pprbAsw7Kz477JX6f8HTWsdpEdQAtA37t5c,78974
|
| 116 |
+
huggingface_hub/hf_api.py,sha256=a12ywN6Lc43A12B36xxK4dJzXsgtmjQTZqYAC4gjTzY,465077
|
| 117 |
+
huggingface_hub/hf_file_system.py,sha256=nrNOoNHRwf1swebtQvZExSblRjQg9rHKxL7Cslk72uw,46899
|
| 118 |
+
huggingface_hub/hub_mixin.py,sha256=MArtbUxjXiYeOvOmNBG9I_j5t02m2xCVAcge4waip1w,38091
|
| 119 |
+
huggingface_hub/inference/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
| 120 |
+
huggingface_hub/inference/__pycache__/__init__.cpython-310.pyc,,
|
| 121 |
+
huggingface_hub/inference/__pycache__/_client.cpython-310.pyc,,
|
| 122 |
+
huggingface_hub/inference/__pycache__/_common.cpython-310.pyc,,
|
| 123 |
+
huggingface_hub/inference/_client.py,sha256=ggY3ebg_xjC9zuHUTpzG_rghFr65vDt7Btg3pATfJjE,165039
|
| 124 |
+
huggingface_hub/inference/_common.py,sha256=6qAIauugyl1eHk0FhWdjBNEXBNF33_VXC8lc1GR8t7s,15874
|
| 125 |
+
huggingface_hub/inference/_generated/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
| 126 |
+
huggingface_hub/inference/_generated/__pycache__/__init__.cpython-310.pyc,,
|
| 127 |
+
huggingface_hub/inference/_generated/__pycache__/_async_client.cpython-310.pyc,,
|
| 128 |
+
huggingface_hub/inference/_generated/_async_client.py,sha256=oklZPp7TBmCkUB5B3UpuivEXQWGgUxXJwA-Zs4YFDsM,171274
|
| 129 |
+
huggingface_hub/inference/_generated/types/__init__.py,sha256=9WvrGQ8aThtKSNzZF06j-CIE2ZuItne8FFnea1p1u38,6557
|
| 130 |
+
huggingface_hub/inference/_generated/types/__pycache__/__init__.cpython-310.pyc,,
|
| 131 |
+
huggingface_hub/inference/_generated/types/__pycache__/audio_classification.cpython-310.pyc,,
|
| 132 |
+
huggingface_hub/inference/_generated/types/__pycache__/audio_to_audio.cpython-310.pyc,,
|
| 133 |
+
huggingface_hub/inference/_generated/types/__pycache__/automatic_speech_recognition.cpython-310.pyc,,
|
| 134 |
+
huggingface_hub/inference/_generated/types/__pycache__/base.cpython-310.pyc,,
|
| 135 |
+
huggingface_hub/inference/_generated/types/__pycache__/chat_completion.cpython-310.pyc,,
|
| 136 |
+
huggingface_hub/inference/_generated/types/__pycache__/depth_estimation.cpython-310.pyc,,
|
| 137 |
+
huggingface_hub/inference/_generated/types/__pycache__/document_question_answering.cpython-310.pyc,,
|
| 138 |
+
huggingface_hub/inference/_generated/types/__pycache__/feature_extraction.cpython-310.pyc,,
|
| 139 |
+
huggingface_hub/inference/_generated/types/__pycache__/fill_mask.cpython-310.pyc,,
|
| 140 |
+
huggingface_hub/inference/_generated/types/__pycache__/image_classification.cpython-310.pyc,,
|
| 141 |
+
huggingface_hub/inference/_generated/types/__pycache__/image_segmentation.cpython-310.pyc,,
|
| 142 |
+
huggingface_hub/inference/_generated/types/__pycache__/image_to_image.cpython-310.pyc,,
|
| 143 |
+
huggingface_hub/inference/_generated/types/__pycache__/image_to_text.cpython-310.pyc,,
|
| 144 |
+
huggingface_hub/inference/_generated/types/__pycache__/image_to_video.cpython-310.pyc,,
|
| 145 |
+
huggingface_hub/inference/_generated/types/__pycache__/object_detection.cpython-310.pyc,,
|
| 146 |
+
huggingface_hub/inference/_generated/types/__pycache__/question_answering.cpython-310.pyc,,
|
| 147 |
+
huggingface_hub/inference/_generated/types/__pycache__/sentence_similarity.cpython-310.pyc,,
|
| 148 |
+
huggingface_hub/inference/_generated/types/__pycache__/summarization.cpython-310.pyc,,
|
| 149 |
+
huggingface_hub/inference/_generated/types/__pycache__/table_question_answering.cpython-310.pyc,,
|
| 150 |
+
huggingface_hub/inference/_generated/types/__pycache__/text2text_generation.cpython-310.pyc,,
|
| 151 |
+
huggingface_hub/inference/_generated/types/__pycache__/text_classification.cpython-310.pyc,,
|
| 152 |
+
huggingface_hub/inference/_generated/types/__pycache__/text_generation.cpython-310.pyc,,
|
| 153 |
+
huggingface_hub/inference/_generated/types/__pycache__/text_to_audio.cpython-310.pyc,,
|
| 154 |
+
huggingface_hub/inference/_generated/types/__pycache__/text_to_image.cpython-310.pyc,,
|
| 155 |
+
huggingface_hub/inference/_generated/types/__pycache__/text_to_speech.cpython-310.pyc,,
|
| 156 |
+
huggingface_hub/inference/_generated/types/__pycache__/text_to_video.cpython-310.pyc,,
|
| 157 |
+
huggingface_hub/inference/_generated/types/__pycache__/token_classification.cpython-310.pyc,,
|
| 158 |
+
huggingface_hub/inference/_generated/types/__pycache__/translation.cpython-310.pyc,,
|
| 159 |
+
huggingface_hub/inference/_generated/types/__pycache__/video_classification.cpython-310.pyc,,
|
| 160 |
+
huggingface_hub/inference/_generated/types/__pycache__/visual_question_answering.cpython-310.pyc,,
|
| 161 |
+
huggingface_hub/inference/_generated/types/__pycache__/zero_shot_classification.cpython-310.pyc,,
|
| 162 |
+
huggingface_hub/inference/_generated/types/__pycache__/zero_shot_image_classification.cpython-310.pyc,,
|
| 163 |
+
huggingface_hub/inference/_generated/types/__pycache__/zero_shot_object_detection.cpython-310.pyc,,
|
| 164 |
+
huggingface_hub/inference/_generated/types/audio_classification.py,sha256=Jg3mzfGhCSH6CfvVvgJSiFpkz6v4nNA0G4LJXacEgNc,1573
|
| 165 |
+
huggingface_hub/inference/_generated/types/audio_to_audio.py,sha256=2Ep4WkePL7oJwcp5nRJqApwviumGHbft9HhXE9XLHj4,891
|
| 166 |
+
huggingface_hub/inference/_generated/types/automatic_speech_recognition.py,sha256=8CEphr6rvRHgq1L5Md3tq14V0tEAmzJkemh1_7gSswo,5515
|
| 167 |
+
huggingface_hub/inference/_generated/types/base.py,sha256=4XG49q0-2SOftYQ8HXQnWLxiJktou-a7IoG3kdOv-kg,6751
|
| 168 |
+
huggingface_hub/inference/_generated/types/chat_completion.py,sha256=6EqWnpbeT0zsiLNZjoJDzrmZ34M2j01S99oMMayZg9Y,11182
|
| 169 |
+
huggingface_hub/inference/_generated/types/depth_estimation.py,sha256=rcpe9MhYMeLjflOwBs3KMZPr6WjOH3FYEThStG-FJ3M,929
|
| 170 |
+
huggingface_hub/inference/_generated/types/document_question_answering.py,sha256=6BEYGwJcqGlah4RBJDAvWFTEXkO0mosBiMy82432nAM,3202
|
| 171 |
+
huggingface_hub/inference/_generated/types/feature_extraction.py,sha256=NMWVL_TLSG5SS5bdt1-fflkZ75UMlMKeTMtmdnUTADc,1537
|
| 172 |
+
huggingface_hub/inference/_generated/types/fill_mask.py,sha256=OrTgQ7Ndn0_dWK5thQhZwTOHbQni8j0iJcx9llyhRds,1708
|
| 173 |
+
huggingface_hub/inference/_generated/types/image_classification.py,sha256=A-Y024o8723_n8mGVos4TwdAkVL62McGeL1iIo4VzNs,1585
|
| 174 |
+
huggingface_hub/inference/_generated/types/image_segmentation.py,sha256=vrkI4SuP1Iq_iLXc-2pQhYY3SHN4gzvFBoZqbUHxU7o,1950
|
| 175 |
+
huggingface_hub/inference/_generated/types/image_to_image.py,sha256=HPz1uKXk_9xvgNUi3GV6n4lw-J3G6cdGTcW3Ou_N0l8,2044
|
| 176 |
+
huggingface_hub/inference/_generated/types/image_to_text.py,sha256=OaFEBAfgT-fOVzJ7xVermGf7VODhrc9-Jg38WrM7-2o,4810
|
| 177 |
+
huggingface_hub/inference/_generated/types/image_to_video.py,sha256=bC-L_cNsDhk4s_IdSiprJ9d1NeMGePLcUp7UPpco21w,2240
|
| 178 |
+
huggingface_hub/inference/_generated/types/object_detection.py,sha256=VuFlb1281qTXoSgJDmquGz-VNfEZLo2H0Rh_F6MF6ts,2000
|
| 179 |
+
huggingface_hub/inference/_generated/types/question_answering.py,sha256=zw38a9_9l2k1ifYZefjkioqZ4asfSRM9M4nU3gSCmAQ,2898
|
| 180 |
+
huggingface_hub/inference/_generated/types/sentence_similarity.py,sha256=w5Nj1g18eBzopZwxuDLI-fEsyaCK2KrHA5yf_XfSjgo,1052
|
| 181 |
+
huggingface_hub/inference/_generated/types/summarization.py,sha256=WGGr8uDLrZg8JQgF9ZMUP9euw6uZo6zwkVZ-IfvCFI0,1487
|
| 182 |
+
huggingface_hub/inference/_generated/types/table_question_answering.py,sha256=cJnIPA2fIbQP2Ejn7X_esY48qGWoXg30fnNOqCXiOVQ,2293
|
| 183 |
+
huggingface_hub/inference/_generated/types/text2text_generation.py,sha256=v-418w1JNNSZ2tuW9DUl6a36TQQCADa438A3ufvcbOw,1609
|
| 184 |
+
huggingface_hub/inference/_generated/types/text_classification.py,sha256=FarAjygLEfPofLfKeabzJ7PKEBItlHGoUNUOzyLRpL4,1445
|
| 185 |
+
huggingface_hub/inference/_generated/types/text_generation.py,sha256=28u-1zU7elk2teP3y4u1VAtDDHzY0JZ2KEEJe5d5uvg,5922
|
| 186 |
+
huggingface_hub/inference/_generated/types/text_to_audio.py,sha256=1HR9Q6s9MXqtKGTvHPLGVMum5-eg7O-Pgv6Nd0v8_HU,4741
|
| 187 |
+
huggingface_hub/inference/_generated/types/text_to_image.py,sha256=sGGi1Fa0n5Pmd6G3I-F2SBJcJ1M7Gmqnng6sfi0AVzs,1903
|
| 188 |
+
huggingface_hub/inference/_generated/types/text_to_speech.py,sha256=ROFuR32ijROCeqbv81Jos0lmaA8SRWyIUsWrdD4yWow,4760
|
| 189 |
+
huggingface_hub/inference/_generated/types/text_to_video.py,sha256=yHXVNs3t6aYO7visrBlB5cH7kjoysxF9510aofcf_18,1790
|
| 190 |
+
huggingface_hub/inference/_generated/types/token_classification.py,sha256=iblAcgfxXeaLYJ14NdiiCMIQuBlarUknLkXUklhvcLI,1915
|
| 191 |
+
huggingface_hub/inference/_generated/types/translation.py,sha256=xww4X5cfCYv_F0oINWLwqJRPCT6SV3VBAJuPjTs_j7o,1763
|
| 192 |
+
huggingface_hub/inference/_generated/types/video_classification.py,sha256=TyydjQw2NRLK9sDGzJUVnkDeo848ebmCx588Ur8I9q0,1680
|
| 193 |
+
huggingface_hub/inference/_generated/types/visual_question_answering.py,sha256=AWrQ6qo4gZa3PGedaNpzDFqx5yOYyjhnUB6iuZEj_uo,1673
|
| 194 |
+
huggingface_hub/inference/_generated/types/zero_shot_classification.py,sha256=BAiebPjsqoNa8EU35Dx0pfIv8W2c4GSl-TJckV1MaxQ,1738
|
| 195 |
+
huggingface_hub/inference/_generated/types/zero_shot_image_classification.py,sha256=8J9n6VqFARkWvPfAZNWEG70AlrMGldU95EGQQwn06zI,1487
|
| 196 |
+
huggingface_hub/inference/_generated/types/zero_shot_object_detection.py,sha256=GUd81LIV7oEbRWayDlAVgyLmY596r1M3AW0jXDp1yTA,1630
|
| 197 |
+
huggingface_hub/inference/_mcp/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
| 198 |
+
huggingface_hub/inference/_mcp/__pycache__/__init__.cpython-310.pyc,,
|
| 199 |
+
huggingface_hub/inference/_mcp/__pycache__/_cli_hacks.cpython-310.pyc,,
|
| 200 |
+
huggingface_hub/inference/_mcp/__pycache__/agent.cpython-310.pyc,,
|
| 201 |
+
huggingface_hub/inference/_mcp/__pycache__/cli.cpython-310.pyc,,
|
| 202 |
+
huggingface_hub/inference/_mcp/__pycache__/constants.cpython-310.pyc,,
|
| 203 |
+
huggingface_hub/inference/_mcp/__pycache__/mcp_client.cpython-310.pyc,,
|
| 204 |
+
huggingface_hub/inference/_mcp/__pycache__/types.cpython-310.pyc,,
|
| 205 |
+
huggingface_hub/inference/_mcp/__pycache__/utils.cpython-310.pyc,,
|
| 206 |
+
huggingface_hub/inference/_mcp/_cli_hacks.py,sha256=cMZirVFe4N0EM9Nzzs9aEmzUBUEBYR4oYZpByTWlZCM,3182
|
| 207 |
+
huggingface_hub/inference/_mcp/agent.py,sha256=VahvSqldiC1R72CFH4T05l80uEXl5OjLwboWQFUJbsw,4281
|
| 208 |
+
huggingface_hub/inference/_mcp/cli.py,sha256=UY4K20zp5ydGx7Gzp4goVXVIDV2gB2arKFeGWK1AzwE,9625
|
| 209 |
+
huggingface_hub/inference/_mcp/constants.py,sha256=AnOp_oR5Vty0d5J3AynGGNK9i1I9KgGhCd9nPoZcD1M,2495
|
| 210 |
+
huggingface_hub/inference/_mcp/mcp_client.py,sha256=ndaTcZZPbU1ZTNUeB9-WdaOx7bHD3lsrXnKxCeiwpUg,15788
|
| 211 |
+
huggingface_hub/inference/_mcp/types.py,sha256=ic8VSR9JY1d-vPWsBVXYtXtIU669-HbGQ3m12Szs7BQ,815
|
| 212 |
+
huggingface_hub/inference/_mcp/utils.py,sha256=VsRWl0fuSZDS0zNT9n7FOMSlzA0UBbP8p8xWKWDt2Pc,4093
|
| 213 |
+
huggingface_hub/inference/_providers/__init__.py,sha256=8i8Y2X_hEhm20ndFjEQxgddHVAEQpYn38UfLeH4V0Rg,8354
|
| 214 |
+
huggingface_hub/inference/_providers/__pycache__/__init__.cpython-310.pyc,,
|
| 215 |
+
huggingface_hub/inference/_providers/__pycache__/_common.cpython-310.pyc,,
|
| 216 |
+
huggingface_hub/inference/_providers/__pycache__/black_forest_labs.cpython-310.pyc,,
|
| 217 |
+
huggingface_hub/inference/_providers/__pycache__/cerebras.cpython-310.pyc,,
|
| 218 |
+
huggingface_hub/inference/_providers/__pycache__/cohere.cpython-310.pyc,,
|
| 219 |
+
huggingface_hub/inference/_providers/__pycache__/fal_ai.cpython-310.pyc,,
|
| 220 |
+
huggingface_hub/inference/_providers/__pycache__/featherless_ai.cpython-310.pyc,,
|
| 221 |
+
huggingface_hub/inference/_providers/__pycache__/fireworks_ai.cpython-310.pyc,,
|
| 222 |
+
huggingface_hub/inference/_providers/__pycache__/groq.cpython-310.pyc,,
|
| 223 |
+
huggingface_hub/inference/_providers/__pycache__/hf_inference.cpython-310.pyc,,
|
| 224 |
+
huggingface_hub/inference/_providers/__pycache__/hyperbolic.cpython-310.pyc,,
|
| 225 |
+
huggingface_hub/inference/_providers/__pycache__/nebius.cpython-310.pyc,,
|
| 226 |
+
huggingface_hub/inference/_providers/__pycache__/novita.cpython-310.pyc,,
|
| 227 |
+
huggingface_hub/inference/_providers/__pycache__/nscale.cpython-310.pyc,,
|
| 228 |
+
huggingface_hub/inference/_providers/__pycache__/openai.cpython-310.pyc,,
|
| 229 |
+
huggingface_hub/inference/_providers/__pycache__/replicate.cpython-310.pyc,,
|
| 230 |
+
huggingface_hub/inference/_providers/__pycache__/sambanova.cpython-310.pyc,,
|
| 231 |
+
huggingface_hub/inference/_providers/__pycache__/together.cpython-310.pyc,,
|
| 232 |
+
huggingface_hub/inference/_providers/_common.py,sha256=dDtKWEUTcvO0llx96x5E3RwOSzkCIL8cKbiFk9RRh8o,11330
|
| 233 |
+
huggingface_hub/inference/_providers/black_forest_labs.py,sha256=wO7qgRyNyrIKlZtvL3vJEbS4-D19kfoXZk6PDh1dTis,2842
|
| 234 |
+
huggingface_hub/inference/_providers/cerebras.py,sha256=QOJ-1U-os7uE7p6eUnn_P_APq-yQhx28be7c3Tq2EuA,210
|
| 235 |
+
huggingface_hub/inference/_providers/cohere.py,sha256=O3tC-qIUL91mx_mE8bOHCtDWcQuKOUauhUoXSUBUCZ8,1253
|
| 236 |
+
huggingface_hub/inference/_providers/fal_ai.py,sha256=Z3-MPddtsfjWT1Zmvni0m8LPBSFjOa_wyv8xV9FNtoA,9860
|
| 237 |
+
huggingface_hub/inference/_providers/featherless_ai.py,sha256=QxBz-32O4PztxixrIjrfKuTOzvfqyUi-cVsw0Hf_zlY,1382
|
| 238 |
+
huggingface_hub/inference/_providers/fireworks_ai.py,sha256=Id226ITfPkOcFMFzly3MW9l-dZl9l4qizL4JEHWkBFk,1215
|
| 239 |
+
huggingface_hub/inference/_providers/groq.py,sha256=JTk2JV4ZOlaohho7zLAFQtk92kGVsPmLJ1hmzcwsqvQ,315
|
| 240 |
+
huggingface_hub/inference/_providers/hf_inference.py,sha256=P4FJdzk3M2T6GdSTumBvpcr8NNRixWLXymv4ptpHroI,9508
|
| 241 |
+
huggingface_hub/inference/_providers/hyperbolic.py,sha256=OQIBi2j3aNvuaSQ8BUK1K1PVeRXdrxc80G-6YmBa-ns,1985
|
| 242 |
+
huggingface_hub/inference/_providers/nebius.py,sha256=VJpTF2JZ58rznc9wxdk-57vwF8sV2vESw_WkXjXqCho,3580
|
| 243 |
+
huggingface_hub/inference/_providers/novita.py,sha256=HGVC8wPraRQUuI5uBoye1Y4Wqe4X116B71GhhbWy5yM,2514
|
| 244 |
+
huggingface_hub/inference/_providers/nscale.py,sha256=qWUsWinQmUbNUqehyKn34tVoWehu8gd-OZ2F4uj2SWM,1802
|
| 245 |
+
huggingface_hub/inference/_providers/openai.py,sha256=GCVYeNdjWIgpQQ7E_Xv8IebmdhTi0S6WfFosz3nLtps,1089
|
| 246 |
+
huggingface_hub/inference/_providers/replicate.py,sha256=BuLb1x4nUlH5SfazBwvMiFwwcs-OS99U87m3QWdx2is,3810
|
| 247 |
+
huggingface_hub/inference/_providers/sambanova.py,sha256=Unt3H3jr_kgI9vzRjmmW1DFyoEuPkKCcgIIloiOj3j8,2037
|
| 248 |
+
huggingface_hub/inference/_providers/together.py,sha256=KHF19CS3qXS7G1-CwcMiD8Z5wzPKEKi4F2DzqAthbBE,3439
|
| 249 |
+
huggingface_hub/inference_api.py,sha256=b4-NhPSn9b44nYKV8tDKXodmE4JVdEymMWL4CVGkzlE,8323
|
| 250 |
+
huggingface_hub/keras_mixin.py,sha256=WGNQZROdw6yjJ1DGTPZPwKAxf1UbkzAx1dRidkeb2fk,19553
|
| 251 |
+
huggingface_hub/lfs.py,sha256=n-TIjK7J7aXG3zi__0nkd6aNkE4djOf9CD6dYQOQ5P8,16649
|
| 252 |
+
huggingface_hub/py.typed,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
| 253 |
+
huggingface_hub/repocard.py,sha256=KxQoJTvnMVA6mKpi7660O6dKtbw4wRJMBxyXuP7vvzU,34751
|
| 254 |
+
huggingface_hub/repocard_data.py,sha256=hr4ReFpEQMNdh_9Dx-L-IJoI1ElHyk-h-8ZRqwVYYOE,34082
|
| 255 |
+
huggingface_hub/repository.py,sha256=Lerq3kr7tC-oUdZk5i1CdhAA84ZvYiqjaGR77j2iOyk,54536
|
| 256 |
+
huggingface_hub/serialization/__init__.py,sha256=kn-Fa-m4FzMnN8lNsF-SwFcfzug4CucexybGKyvZ8S0,1041
|
| 257 |
+
huggingface_hub/serialization/__pycache__/__init__.cpython-310.pyc,,
|
| 258 |
+
huggingface_hub/serialization/__pycache__/_base.cpython-310.pyc,,
|
| 259 |
+
huggingface_hub/serialization/__pycache__/_dduf.cpython-310.pyc,,
|
| 260 |
+
huggingface_hub/serialization/__pycache__/_tensorflow.cpython-310.pyc,,
|
| 261 |
+
huggingface_hub/serialization/__pycache__/_torch.cpython-310.pyc,,
|
| 262 |
+
huggingface_hub/serialization/_base.py,sha256=Df3GwGR9NzeK_SD75prXLucJAzPiNPgHbgXSw-_LTk8,8126
|
| 263 |
+
huggingface_hub/serialization/_dduf.py,sha256=s42239rLiHwaJE36QDEmS5GH7DSmQ__BffiHJO5RjIg,15424
|
| 264 |
+
huggingface_hub/serialization/_tensorflow.py,sha256=zHOvEMg-JHC55Fm4roDT3LUCDO5zB9qtXZffG065RAM,3625
|
| 265 |
+
huggingface_hub/serialization/_torch.py,sha256=jpBmuSZJymMpvLcDcMaNxDu_fE5VkY_pAVH8e8stYIo,45201
|
| 266 |
+
huggingface_hub/templates/datasetcard_template.md,sha256=W-EMqR6wndbrnZorkVv56URWPG49l7MATGeI015kTvs,5503
|
| 267 |
+
huggingface_hub/templates/modelcard_template.md,sha256=4AqArS3cqdtbit5Bo-DhjcnDFR-pza5hErLLTPM4Yuc,6870
|
| 268 |
+
huggingface_hub/utils/__init__.py,sha256=ORfVkn5D0wuLIq12jjhTzn5_c4F8fRPxB7TG-iednuQ,3722
|
| 269 |
+
huggingface_hub/utils/__pycache__/__init__.cpython-310.pyc,,
|
| 270 |
+
huggingface_hub/utils/__pycache__/_auth.cpython-310.pyc,,
|
| 271 |
+
huggingface_hub/utils/__pycache__/_cache_assets.cpython-310.pyc,,
|
| 272 |
+
huggingface_hub/utils/__pycache__/_cache_manager.cpython-310.pyc,,
|
| 273 |
+
huggingface_hub/utils/__pycache__/_chunk_utils.cpython-310.pyc,,
|
| 274 |
+
huggingface_hub/utils/__pycache__/_datetime.cpython-310.pyc,,
|
| 275 |
+
huggingface_hub/utils/__pycache__/_deprecation.cpython-310.pyc,,
|
| 276 |
+
huggingface_hub/utils/__pycache__/_dotenv.cpython-310.pyc,,
|
| 277 |
+
huggingface_hub/utils/__pycache__/_experimental.cpython-310.pyc,,
|
| 278 |
+
huggingface_hub/utils/__pycache__/_fixes.cpython-310.pyc,,
|
| 279 |
+
huggingface_hub/utils/__pycache__/_git_credential.cpython-310.pyc,,
|
| 280 |
+
huggingface_hub/utils/__pycache__/_headers.cpython-310.pyc,,
|
| 281 |
+
huggingface_hub/utils/__pycache__/_hf_folder.cpython-310.pyc,,
|
| 282 |
+
huggingface_hub/utils/__pycache__/_http.cpython-310.pyc,,
|
| 283 |
+
huggingface_hub/utils/__pycache__/_lfs.cpython-310.pyc,,
|
| 284 |
+
huggingface_hub/utils/__pycache__/_pagination.cpython-310.pyc,,
|
| 285 |
+
huggingface_hub/utils/__pycache__/_paths.cpython-310.pyc,,
|
| 286 |
+
huggingface_hub/utils/__pycache__/_runtime.cpython-310.pyc,,
|
| 287 |
+
huggingface_hub/utils/__pycache__/_safetensors.cpython-310.pyc,,
|
| 288 |
+
huggingface_hub/utils/__pycache__/_subprocess.cpython-310.pyc,,
|
| 289 |
+
huggingface_hub/utils/__pycache__/_telemetry.cpython-310.pyc,,
|
| 290 |
+
huggingface_hub/utils/__pycache__/_typing.cpython-310.pyc,,
|
| 291 |
+
huggingface_hub/utils/__pycache__/_validators.cpython-310.pyc,,
|
| 292 |
+
huggingface_hub/utils/__pycache__/_xet.cpython-310.pyc,,
|
| 293 |
+
huggingface_hub/utils/__pycache__/_xet_progress_reporting.cpython-310.pyc,,
|
| 294 |
+
huggingface_hub/utils/__pycache__/endpoint_helpers.cpython-310.pyc,,
|
| 295 |
+
huggingface_hub/utils/__pycache__/insecure_hashlib.cpython-310.pyc,,
|
| 296 |
+
huggingface_hub/utils/__pycache__/logging.cpython-310.pyc,,
|
| 297 |
+
huggingface_hub/utils/__pycache__/sha.cpython-310.pyc,,
|
| 298 |
+
huggingface_hub/utils/__pycache__/tqdm.cpython-310.pyc,,
|
| 299 |
+
huggingface_hub/utils/_auth.py,sha256=Ixve2vxdftHXXk2R2vfyLzlVoDT39Tkq-Hrou9KCUvw,8286
|
| 300 |
+
huggingface_hub/utils/_cache_assets.py,sha256=kai77HPQMfYpROouMBQCr_gdBCaeTm996Sqj0dExbNg,5728
|
| 301 |
+
huggingface_hub/utils/_cache_manager.py,sha256=osqV4gM5Mx293Kw1POuJ5uKIcoIG57gY0k6BU6vdjPA,34518
|
| 302 |
+
huggingface_hub/utils/_chunk_utils.py,sha256=kRCaj5228_vKcyLWspd8Xq01f17Jz6ds5Sr9ed5d_RU,2130
|
| 303 |
+
huggingface_hub/utils/_datetime.py,sha256=kCS5jaKV25kOncX1xujbXsz5iDLcjLcLw85semGNzxQ,2770
|
| 304 |
+
huggingface_hub/utils/_deprecation.py,sha256=HZhRGGUX_QMKBBBwHHlffLtmCSK01TOpeXHefZbPfwI,4872
|
| 305 |
+
huggingface_hub/utils/_dotenv.py,sha256=RzHqC8HgzVxE-N4DFBcnemvX0NHmXcV0My2ASK0U1OQ,2017
|
| 306 |
+
huggingface_hub/utils/_experimental.py,sha256=3-c8irbn9sJr2CwWbzhGkIrdXKg8_x7BifhHFy32ei8,2470
|
| 307 |
+
huggingface_hub/utils/_fixes.py,sha256=xQV1QkUn2WpLqLjtXNiyn9gh-454K6AF-Q3kwkYAQD8,4437
|
| 308 |
+
huggingface_hub/utils/_git_credential.py,sha256=SDdsiREr1TcAR2Ze2TB0E5cYzVJgvDZrs60od9lAsMc,4596
|
| 309 |
+
huggingface_hub/utils/_headers.py,sha256=w4ayq4hLGaZ3B7nwdEi5Zu23SmmDuOwv58It78wkakk,8868
|
| 310 |
+
huggingface_hub/utils/_hf_folder.py,sha256=WNjTnu0Q7tqcSS9EsP4ssCJrrJMcCvAt8P_-LEtmOU8,2487
|
| 311 |
+
huggingface_hub/utils/_http.py,sha256=her7UZ0KRo9WYDArpqVFyEXTusOGUECj5HNS8Eahqm8,25531
|
| 312 |
+
huggingface_hub/utils/_lfs.py,sha256=EC0Oz6Wiwl8foRNkUOzrETXzAWlbgpnpxo5a410ovFY,3957
|
| 313 |
+
huggingface_hub/utils/_pagination.py,sha256=EX5tRasSuQDaKbXuGYbInBK2odnSWNHgzw2tSgqeBRI,1906
|
| 314 |
+
huggingface_hub/utils/_paths.py,sha256=w1ZhFmmD5ykWjp_hAvhjtOoa2ZUcOXJrF4a6O3QpAWo,5042
|
| 315 |
+
huggingface_hub/utils/_runtime.py,sha256=L7SOYezdxKcwd4DovAY0UGY3qt27toXO-QjceIDwExk,11634
|
| 316 |
+
huggingface_hub/utils/_safetensors.py,sha256=GW3nyv7xQcuwObKYeYoT9VhURVzG1DZTbKBKho8Bbos,4458
|
| 317 |
+
huggingface_hub/utils/_subprocess.py,sha256=u9FFUDE7TrzQTiuEzlUnHx7S2P57GbYRV8u16GJwrFw,4625
|
| 318 |
+
huggingface_hub/utils/_telemetry.py,sha256=54LXeIJU5pEGghPAh06gqNAR-UoxOjVLvKqAQscwqZs,4890
|
| 319 |
+
huggingface_hub/utils/_typing.py,sha256=Dgp6TQUlpzStfVLoSvXHCBP4b3NzHZ8E0Gg9mYAoDS4,2903
|
| 320 |
+
huggingface_hub/utils/_validators.py,sha256=dDsVG31iooTYrIyi5Vwr1DukL0fEmJwu3ceVNduhsuE,9204
|
| 321 |
+
huggingface_hub/utils/_xet.py,sha256=f8qfk8YKePAeGUL6lQiQ1w_3bcs78oWwbeACYdUeg5k,7312
|
| 322 |
+
huggingface_hub/utils/_xet_progress_reporting.py,sha256=i1nOjoUtaUpbWV7Tkn5H9cEJZ0CduVz3Q6eIjNvyBhk,5450
|
| 323 |
+
huggingface_hub/utils/endpoint_helpers.py,sha256=9VtIAlxQ5H_4y30sjCAgbu7XCqAtNLC7aRYxaNn0hLI,2366
|
| 324 |
+
huggingface_hub/utils/insecure_hashlib.py,sha256=iAaepavFZ5Dhfa5n8KozRfQprKmvcjSnt3X58OUl9fQ,1142
|
| 325 |
+
huggingface_hub/utils/logging.py,sha256=0A8fF1yh3L9Ka_bCDX2ml4U5Ht0tY8Dr3JcbRvWFuwo,4909
|
| 326 |
+
huggingface_hub/utils/sha.py,sha256=OFnNGCba0sNcT2gUwaVCJnldxlltrHHe0DS_PCpV3C4,2134
|
| 327 |
+
huggingface_hub/utils/tqdm.py,sha256=xAKcyfnNHsZ7L09WuEM5Ew5-MDhiahLACbbN2zMmcLs,10671
|
pythonProject/.venv/Lib/site-packages/huggingface_hub-0.34.4.dist-info/WHEEL
ADDED
|
@@ -0,0 +1,5 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Wheel-Version: 1.0
|
| 2 |
+
Generator: bdist_wheel (0.45.1)
|
| 3 |
+
Root-Is-Purelib: true
|
| 4 |
+
Tag: py3-none-any
|
| 5 |
+
|
pythonProject/.venv/Lib/site-packages/huggingface_hub-0.34.4.dist-info/entry_points.txt
ADDED
|
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[console_scripts]
|
| 2 |
+
hf = huggingface_hub.cli.hf:main
|
| 3 |
+
huggingface-cli = huggingface_hub.commands.huggingface_cli:main
|
| 4 |
+
tiny-agents = huggingface_hub.inference._mcp.cli:app
|
| 5 |
+
|
| 6 |
+
[fsspec.specs]
|
| 7 |
+
hf=huggingface_hub.HfFileSystem
|
| 8 |
+
|
pythonProject/.venv/Lib/site-packages/huggingface_hub-0.34.4.dist-info/top_level.txt
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
huggingface_hub
|
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/__init__.cpython-310.pyc
ADDED
|
Binary file (20.6 kB). View file
|
|
|
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/_commit_api.cpython-310.pyc
ADDED
|
Binary file (30.1 kB). View file
|
|
|
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/_commit_scheduler.cpython-310.pyc
ADDED
|
Binary file (13.8 kB). View file
|
|
|
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/_inference_endpoints.cpython-310.pyc
ADDED
|
Binary file (15.8 kB). View file
|
|
|
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/_jobs_api.cpython-310.pyc
ADDED
|
Binary file (5.1 kB). View file
|
|
|
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/_tensorboard_logger.cpython-310.pyc
ADDED
|
Binary file (6.86 kB). View file
|
|
|
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/_upload_large_folder.cpython-310.pyc
ADDED
|
Binary file (20.7 kB). View file
|
|
|
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/_webhooks_payload.cpython-310.pyc
ADDED
|
Binary file (3.93 kB). View file
|
|
|
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/_webhooks_server.cpython-310.pyc
ADDED
|
Binary file (14 kB). View file
|
|
|
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/community.cpython-310.pyc
ADDED
|
Binary file (13 kB). View file
|
|
|
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/constants.cpython-310.pyc
ADDED
|
Binary file (6.02 kB). View file
|
|
|
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/dataclasses.cpython-310.pyc
ADDED
|
Binary file (13.6 kB). View file
|
|
|
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/errors.cpython-310.pyc
ADDED
|
Binary file (14.1 kB). View file
|
|
|
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/fastai_utils.cpython-310.pyc
ADDED
|
Binary file (14.3 kB). View file
|
|
|
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/file_download.cpython-310.pyc
ADDED
|
Binary file (50.8 kB). View file
|
|
|
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/hf_file_system.cpython-310.pyc
ADDED
|
Binary file (33.1 kB). View file
|
|
|
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/hub_mixin.cpython-310.pyc
ADDED
|
Binary file (29.1 kB). View file
|
|
|
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/inference_api.cpython-310.pyc
ADDED
|
Binary file (7.57 kB). View file
|
|
|
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/keras_mixin.cpython-310.pyc
ADDED
|
Binary file (17 kB). View file
|
|
|
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/lfs.cpython-310.pyc
ADDED
|
Binary file (13.8 kB). View file
|
|
|
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/repocard.cpython-310.pyc
ADDED
|
Binary file (29.6 kB). View file
|
|
|
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/repocard_data.cpython-310.pyc
ADDED
|
Binary file (29.4 kB). View file
|
|
|
pythonProject/.venv/Lib/site-packages/huggingface_hub/__pycache__/repository.cpython-310.pyc
ADDED
|
Binary file (45.6 kB). View file
|
|
|
pythonProject/.venv/Lib/site-packages/huggingface_hub/inference/__pycache__/__init__.cpython-310.pyc
ADDED
|
Binary file (182 Bytes). View file
|
|
|
pythonProject/.venv/Lib/site-packages/huggingface_hub/inference/__pycache__/_client.cpython-310.pyc
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0fcedb6fe5ff0f30068953849983fc2d2bc0163e249770936db741589d32fc62
|
| 3 |
+
size 137799
|
pythonProject/.venv/Lib/site-packages/huggingface_hub/inference/__pycache__/_common.cpython-310.pyc
ADDED
|
Binary file (12 kB). View file
|
|
|
pythonProject/.venv/Lib/site-packages/huggingface_hub/inference/_providers/__pycache__/__init__.cpython-310.pyc
ADDED
|
Binary file (5.64 kB). View file
|
|
|
pythonProject/.venv/Lib/site-packages/huggingface_hub/inference/_providers/__pycache__/_common.cpython-310.pyc
ADDED
|
Binary file (10.3 kB). View file
|
|
|
pythonProject/.venv/Lib/site-packages/huggingface_hub/inference/_providers/__pycache__/replicate.cpython-310.pyc
ADDED
|
Binary file (4.45 kB). View file
|
|
|
pythonProject/.venv/Lib/site-packages/huggingface_hub/inference/_providers/__pycache__/sambanova.cpython-310.pyc
ADDED
|
Binary file (2.92 kB). View file
|
|
|
pythonProject/.venv/Lib/site-packages/huggingface_hub/inference/_providers/__pycache__/together.cpython-310.pyc
ADDED
|
Binary file (4.2 kB). View file
|
|
|
pythonProject/.venv/Lib/site-packages/huggingface_hub/serialization/__init__.py
ADDED
|
@@ -0,0 +1,27 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright 2024 The HuggingFace Team. All rights reserved.
|
| 2 |
+
#
|
| 3 |
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
| 4 |
+
# you may not use this file except in compliance with the License.
|
| 5 |
+
# You may obtain a copy of the License at
|
| 6 |
+
#
|
| 7 |
+
# http://www.apache.org/licenses/LICENSE-2.0
|
| 8 |
+
#
|
| 9 |
+
# Unless required by applicable law or agreed to in writing, software
|
| 10 |
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
| 11 |
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
| 12 |
+
# See the License for the specific language governing permissions and
|
| 13 |
+
# limitations under the License.
|
| 14 |
+
# ruff: noqa: F401
|
| 15 |
+
"""Contains helpers to serialize tensors."""
|
| 16 |
+
|
| 17 |
+
from ._base import StateDictSplit, split_state_dict_into_shards_factory
|
| 18 |
+
from ._tensorflow import get_tf_storage_size, split_tf_state_dict_into_shards
|
| 19 |
+
from ._torch import (
|
| 20 |
+
get_torch_storage_id,
|
| 21 |
+
get_torch_storage_size,
|
| 22 |
+
load_state_dict_from_file,
|
| 23 |
+
load_torch_model,
|
| 24 |
+
save_torch_model,
|
| 25 |
+
save_torch_state_dict,
|
| 26 |
+
split_torch_state_dict_into_shards,
|
| 27 |
+
)
|
pythonProject/.venv/Lib/site-packages/huggingface_hub/serialization/_base.py
ADDED
|
@@ -0,0 +1,210 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright 2024 The HuggingFace Team. All rights reserved.
|
| 2 |
+
#
|
| 3 |
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
| 4 |
+
# you may not use this file except in compliance with the License.
|
| 5 |
+
# You may obtain a copy of the License at
|
| 6 |
+
#
|
| 7 |
+
# http://www.apache.org/licenses/LICENSE-2.0
|
| 8 |
+
#
|
| 9 |
+
# Unless required by applicable law or agreed to in writing, software
|
| 10 |
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
| 11 |
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
| 12 |
+
# See the License for the specific language governing permissions and
|
| 13 |
+
# limitations under the License.
|
| 14 |
+
"""Contains helpers to split tensors into shards."""
|
| 15 |
+
|
| 16 |
+
from dataclasses import dataclass, field
|
| 17 |
+
from typing import Any, Callable, Dict, List, Optional, TypeVar, Union
|
| 18 |
+
|
| 19 |
+
from .. import logging
|
| 20 |
+
|
| 21 |
+
|
| 22 |
+
TensorT = TypeVar("TensorT")
|
| 23 |
+
TensorSizeFn_T = Callable[[TensorT], int]
|
| 24 |
+
StorageIDFn_T = Callable[[TensorT], Optional[Any]]
|
| 25 |
+
|
| 26 |
+
MAX_SHARD_SIZE = "5GB"
|
| 27 |
+
SIZE_UNITS = {
|
| 28 |
+
"TB": 10**12,
|
| 29 |
+
"GB": 10**9,
|
| 30 |
+
"MB": 10**6,
|
| 31 |
+
"KB": 10**3,
|
| 32 |
+
}
|
| 33 |
+
|
| 34 |
+
|
| 35 |
+
logger = logging.get_logger(__file__)
|
| 36 |
+
|
| 37 |
+
|
| 38 |
+
@dataclass
|
| 39 |
+
class StateDictSplit:
|
| 40 |
+
is_sharded: bool = field(init=False)
|
| 41 |
+
metadata: Dict[str, Any]
|
| 42 |
+
filename_to_tensors: Dict[str, List[str]]
|
| 43 |
+
tensor_to_filename: Dict[str, str]
|
| 44 |
+
|
| 45 |
+
def __post_init__(self):
|
| 46 |
+
self.is_sharded = len(self.filename_to_tensors) > 1
|
| 47 |
+
|
| 48 |
+
|
| 49 |
+
def split_state_dict_into_shards_factory(
|
| 50 |
+
state_dict: Dict[str, TensorT],
|
| 51 |
+
*,
|
| 52 |
+
get_storage_size: TensorSizeFn_T,
|
| 53 |
+
filename_pattern: str,
|
| 54 |
+
get_storage_id: StorageIDFn_T = lambda tensor: None,
|
| 55 |
+
max_shard_size: Union[int, str] = MAX_SHARD_SIZE,
|
| 56 |
+
) -> StateDictSplit:
|
| 57 |
+
"""
|
| 58 |
+
Split a model state dictionary in shards so that each shard is smaller than a given size.
|
| 59 |
+
|
| 60 |
+
The shards are determined by iterating through the `state_dict` in the order of its keys. There is no optimization
|
| 61 |
+
made to make each shard as close as possible to the maximum size passed. For example, if the limit is 10GB and we
|
| 62 |
+
have tensors of sizes [6GB, 6GB, 2GB, 6GB, 2GB, 2GB] they will get sharded as [6GB], [6+2GB], [6+2+2GB] and not
|
| 63 |
+
[6+2+2GB], [6+2GB], [6GB].
|
| 64 |
+
|
| 65 |
+
<Tip warning={true}>
|
| 66 |
+
|
| 67 |
+
If one of the model's tensor is bigger than `max_shard_size`, it will end up in its own shard which will have a
|
| 68 |
+
size greater than `max_shard_size`.
|
| 69 |
+
|
| 70 |
+
</Tip>
|
| 71 |
+
|
| 72 |
+
Args:
|
| 73 |
+
state_dict (`Dict[str, Tensor]`):
|
| 74 |
+
The state dictionary to save.
|
| 75 |
+
get_storage_size (`Callable[[Tensor], int]`):
|
| 76 |
+
A function that returns the size of a tensor when saved on disk in bytes.
|
| 77 |
+
get_storage_id (`Callable[[Tensor], Optional[Any]]`, *optional*):
|
| 78 |
+
A function that returns a unique identifier to a tensor storage. Multiple different tensors can share the
|
| 79 |
+
same underlying storage. This identifier is guaranteed to be unique and constant for this tensor's storage
|
| 80 |
+
during its lifetime. Two tensor storages with non-overlapping lifetimes may have the same id.
|
| 81 |
+
filename_pattern (`str`, *optional*):
|
| 82 |
+
The pattern to generate the files names in which the model will be saved. Pattern must be a string that
|
| 83 |
+
can be formatted with `filename_pattern.format(suffix=...)` and must contain the keyword `suffix`
|
| 84 |
+
max_shard_size (`int` or `str`, *optional*):
|
| 85 |
+
The maximum size of each shard, in bytes. Defaults to 5GB.
|
| 86 |
+
|
| 87 |
+
Returns:
|
| 88 |
+
[`StateDictSplit`]: A `StateDictSplit` object containing the shards and the index to retrieve them.
|
| 89 |
+
"""
|
| 90 |
+
storage_id_to_tensors: Dict[Any, List[str]] = {}
|
| 91 |
+
|
| 92 |
+
shard_list: List[Dict[str, TensorT]] = []
|
| 93 |
+
current_shard: Dict[str, TensorT] = {}
|
| 94 |
+
current_shard_size = 0
|
| 95 |
+
total_size = 0
|
| 96 |
+
|
| 97 |
+
if isinstance(max_shard_size, str):
|
| 98 |
+
max_shard_size = parse_size_to_int(max_shard_size)
|
| 99 |
+
|
| 100 |
+
for key, tensor in state_dict.items():
|
| 101 |
+
# when bnb serialization is used the weights in the state dict can be strings
|
| 102 |
+
# check: https://github.com/huggingface/transformers/pull/24416 for more details
|
| 103 |
+
if isinstance(tensor, str):
|
| 104 |
+
logger.info("Skipping tensor %s as it is a string (bnb serialization)", key)
|
| 105 |
+
continue
|
| 106 |
+
|
| 107 |
+
# If a `tensor` shares the same underlying storage as another tensor, we put `tensor` in the same `block`
|
| 108 |
+
storage_id = get_storage_id(tensor)
|
| 109 |
+
if storage_id is not None:
|
| 110 |
+
if storage_id in storage_id_to_tensors:
|
| 111 |
+
# We skip this tensor for now and will reassign to correct shard later
|
| 112 |
+
storage_id_to_tensors[storage_id].append(key)
|
| 113 |
+
continue
|
| 114 |
+
else:
|
| 115 |
+
# This is the first tensor with this storage_id, we create a new entry
|
| 116 |
+
# in the storage_id_to_tensors dict => we will assign the shard id later
|
| 117 |
+
storage_id_to_tensors[storage_id] = [key]
|
| 118 |
+
|
| 119 |
+
# Compute tensor size
|
| 120 |
+
tensor_size = get_storage_size(tensor)
|
| 121 |
+
|
| 122 |
+
# If this tensor is bigger than the maximal size, we put it in its own shard
|
| 123 |
+
if tensor_size > max_shard_size:
|
| 124 |
+
total_size += tensor_size
|
| 125 |
+
shard_list.append({key: tensor})
|
| 126 |
+
continue
|
| 127 |
+
|
| 128 |
+
# If this tensor is going to tip up over the maximal size, we split.
|
| 129 |
+
# Current shard already has some tensors, we add it to the list of shards and create a new one.
|
| 130 |
+
if current_shard_size + tensor_size > max_shard_size:
|
| 131 |
+
shard_list.append(current_shard)
|
| 132 |
+
current_shard = {}
|
| 133 |
+
current_shard_size = 0
|
| 134 |
+
|
| 135 |
+
# Add the tensor to the current shard
|
| 136 |
+
current_shard[key] = tensor
|
| 137 |
+
current_shard_size += tensor_size
|
| 138 |
+
total_size += tensor_size
|
| 139 |
+
|
| 140 |
+
# Add the last shard
|
| 141 |
+
if len(current_shard) > 0:
|
| 142 |
+
shard_list.append(current_shard)
|
| 143 |
+
nb_shards = len(shard_list)
|
| 144 |
+
|
| 145 |
+
# Loop over the tensors that share the same storage and assign them together
|
| 146 |
+
for storage_id, keys in storage_id_to_tensors.items():
|
| 147 |
+
# Let's try to find the shard where the first tensor of this storage is and put all tensors in the same shard
|
| 148 |
+
for shard in shard_list:
|
| 149 |
+
if keys[0] in shard:
|
| 150 |
+
for key in keys:
|
| 151 |
+
shard[key] = state_dict[key]
|
| 152 |
+
break
|
| 153 |
+
|
| 154 |
+
# If we only have one shard, we return it => no need to build the index
|
| 155 |
+
if nb_shards == 1:
|
| 156 |
+
filename = filename_pattern.format(suffix="")
|
| 157 |
+
return StateDictSplit(
|
| 158 |
+
metadata={"total_size": total_size},
|
| 159 |
+
filename_to_tensors={filename: list(state_dict.keys())},
|
| 160 |
+
tensor_to_filename={key: filename for key in state_dict.keys()},
|
| 161 |
+
)
|
| 162 |
+
|
| 163 |
+
# Now that each tensor is assigned to a shard, let's assign a filename to each shard
|
| 164 |
+
tensor_name_to_filename = {}
|
| 165 |
+
filename_to_tensors = {}
|
| 166 |
+
for idx, shard in enumerate(shard_list):
|
| 167 |
+
filename = filename_pattern.format(suffix=f"-{idx + 1:05d}-of-{nb_shards:05d}")
|
| 168 |
+
for key in shard:
|
| 169 |
+
tensor_name_to_filename[key] = filename
|
| 170 |
+
filename_to_tensors[filename] = list(shard.keys())
|
| 171 |
+
|
| 172 |
+
# Build the index and return
|
| 173 |
+
return StateDictSplit(
|
| 174 |
+
metadata={"total_size": total_size},
|
| 175 |
+
filename_to_tensors=filename_to_tensors,
|
| 176 |
+
tensor_to_filename=tensor_name_to_filename,
|
| 177 |
+
)
|
| 178 |
+
|
| 179 |
+
|
| 180 |
+
def parse_size_to_int(size_as_str: str) -> int:
|
| 181 |
+
"""
|
| 182 |
+
Parse a size expressed as a string with digits and unit (like `"5MB"`) to an integer (in bytes).
|
| 183 |
+
|
| 184 |
+
Supported units are "TB", "GB", "MB", "KB".
|
| 185 |
+
|
| 186 |
+
Args:
|
| 187 |
+
size_as_str (`str`): The size to convert. Will be directly returned if an `int`.
|
| 188 |
+
|
| 189 |
+
Example:
|
| 190 |
+
|
| 191 |
+
```py
|
| 192 |
+
>>> parse_size_to_int("5MB")
|
| 193 |
+
5000000
|
| 194 |
+
```
|
| 195 |
+
"""
|
| 196 |
+
size_as_str = size_as_str.strip()
|
| 197 |
+
|
| 198 |
+
# Parse unit
|
| 199 |
+
unit = size_as_str[-2:].upper()
|
| 200 |
+
if unit not in SIZE_UNITS:
|
| 201 |
+
raise ValueError(f"Unit '{unit}' not supported. Supported units are TB, GB, MB, KB. Got '{size_as_str}'.")
|
| 202 |
+
multiplier = SIZE_UNITS[unit]
|
| 203 |
+
|
| 204 |
+
# Parse value
|
| 205 |
+
try:
|
| 206 |
+
value = float(size_as_str[:-2].strip())
|
| 207 |
+
except ValueError as e:
|
| 208 |
+
raise ValueError(f"Could not parse the size value from '{size_as_str}': {e}") from e
|
| 209 |
+
|
| 210 |
+
return int(value * multiplier)
|
pythonProject/.venv/Lib/site-packages/huggingface_hub/serialization/_dduf.py
ADDED
|
@@ -0,0 +1,387 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import json
|
| 2 |
+
import logging
|
| 3 |
+
import mmap
|
| 4 |
+
import os
|
| 5 |
+
import shutil
|
| 6 |
+
import zipfile
|
| 7 |
+
from contextlib import contextmanager
|
| 8 |
+
from dataclasses import dataclass, field
|
| 9 |
+
from pathlib import Path
|
| 10 |
+
from typing import Any, Dict, Generator, Iterable, Tuple, Union
|
| 11 |
+
|
| 12 |
+
from ..errors import DDUFCorruptedFileError, DDUFExportError, DDUFInvalidEntryNameError
|
| 13 |
+
|
| 14 |
+
|
| 15 |
+
logger = logging.getLogger(__name__)
|
| 16 |
+
|
| 17 |
+
DDUF_ALLOWED_ENTRIES = {
|
| 18 |
+
# Allowed file extensions in a DDUF file
|
| 19 |
+
".json",
|
| 20 |
+
".model",
|
| 21 |
+
".safetensors",
|
| 22 |
+
".txt",
|
| 23 |
+
}
|
| 24 |
+
|
| 25 |
+
DDUF_FOLDER_REQUIRED_ENTRIES = {
|
| 26 |
+
# Each folder must contain at least one of these entries
|
| 27 |
+
"config.json",
|
| 28 |
+
"tokenizer_config.json",
|
| 29 |
+
"preprocessor_config.json",
|
| 30 |
+
"scheduler_config.json",
|
| 31 |
+
}
|
| 32 |
+
|
| 33 |
+
|
| 34 |
+
@dataclass
|
| 35 |
+
class DDUFEntry:
|
| 36 |
+
"""Object representing a file entry in a DDUF file.
|
| 37 |
+
|
| 38 |
+
See [`read_dduf_file`] for how to read a DDUF file.
|
| 39 |
+
|
| 40 |
+
Attributes:
|
| 41 |
+
filename (str):
|
| 42 |
+
The name of the file in the DDUF archive.
|
| 43 |
+
offset (int):
|
| 44 |
+
The offset of the file in the DDUF archive.
|
| 45 |
+
length (int):
|
| 46 |
+
The length of the file in the DDUF archive.
|
| 47 |
+
dduf_path (str):
|
| 48 |
+
The path to the DDUF archive (for internal use).
|
| 49 |
+
"""
|
| 50 |
+
|
| 51 |
+
filename: str
|
| 52 |
+
length: int
|
| 53 |
+
offset: int
|
| 54 |
+
|
| 55 |
+
dduf_path: Path = field(repr=False)
|
| 56 |
+
|
| 57 |
+
@contextmanager
|
| 58 |
+
def as_mmap(self) -> Generator[bytes, None, None]:
|
| 59 |
+
"""Open the file as a memory-mapped file.
|
| 60 |
+
|
| 61 |
+
Useful to load safetensors directly from the file.
|
| 62 |
+
|
| 63 |
+
Example:
|
| 64 |
+
```py
|
| 65 |
+
>>> import safetensors.torch
|
| 66 |
+
>>> with entry.as_mmap() as mm:
|
| 67 |
+
... tensors = safetensors.torch.load(mm)
|
| 68 |
+
```
|
| 69 |
+
"""
|
| 70 |
+
with self.dduf_path.open("rb") as f:
|
| 71 |
+
with mmap.mmap(f.fileno(), length=0, access=mmap.ACCESS_READ) as mm:
|
| 72 |
+
yield mm[self.offset : self.offset + self.length]
|
| 73 |
+
|
| 74 |
+
def read_text(self, encoding: str = "utf-8") -> str:
|
| 75 |
+
"""Read the file as text.
|
| 76 |
+
|
| 77 |
+
Useful for '.txt' and '.json' entries.
|
| 78 |
+
|
| 79 |
+
Example:
|
| 80 |
+
```py
|
| 81 |
+
>>> import json
|
| 82 |
+
>>> index = json.loads(entry.read_text())
|
| 83 |
+
```
|
| 84 |
+
"""
|
| 85 |
+
with self.dduf_path.open("rb") as f:
|
| 86 |
+
f.seek(self.offset)
|
| 87 |
+
return f.read(self.length).decode(encoding=encoding)
|
| 88 |
+
|
| 89 |
+
|
| 90 |
+
def read_dduf_file(dduf_path: Union[os.PathLike, str]) -> Dict[str, DDUFEntry]:
|
| 91 |
+
"""
|
| 92 |
+
Read a DDUF file and return a dictionary of entries.
|
| 93 |
+
|
| 94 |
+
Only the metadata is read, the data is not loaded in memory.
|
| 95 |
+
|
| 96 |
+
Args:
|
| 97 |
+
dduf_path (`str` or `os.PathLike`):
|
| 98 |
+
The path to the DDUF file to read.
|
| 99 |
+
|
| 100 |
+
Returns:
|
| 101 |
+
`Dict[str, DDUFEntry]`:
|
| 102 |
+
A dictionary of [`DDUFEntry`] indexed by filename.
|
| 103 |
+
|
| 104 |
+
Raises:
|
| 105 |
+
- [`DDUFCorruptedFileError`]: If the DDUF file is corrupted (i.e. doesn't follow the DDUF format).
|
| 106 |
+
|
| 107 |
+
Example:
|
| 108 |
+
```python
|
| 109 |
+
>>> import json
|
| 110 |
+
>>> import safetensors.torch
|
| 111 |
+
>>> from huggingface_hub import read_dduf_file
|
| 112 |
+
|
| 113 |
+
# Read DDUF metadata
|
| 114 |
+
>>> dduf_entries = read_dduf_file("FLUX.1-dev.dduf")
|
| 115 |
+
|
| 116 |
+
# Returns a mapping filename <> DDUFEntry
|
| 117 |
+
>>> dduf_entries["model_index.json"]
|
| 118 |
+
DDUFEntry(filename='model_index.json', offset=66, length=587)
|
| 119 |
+
|
| 120 |
+
# Load model index as JSON
|
| 121 |
+
>>> json.loads(dduf_entries["model_index.json"].read_text())
|
| 122 |
+
{'_class_name': 'FluxPipeline', '_diffusers_version': '0.32.0.dev0', '_name_or_path': 'black-forest-labs/FLUX.1-dev', ...
|
| 123 |
+
|
| 124 |
+
# Load VAE weights using safetensors
|
| 125 |
+
>>> with dduf_entries["vae/diffusion_pytorch_model.safetensors"].as_mmap() as mm:
|
| 126 |
+
... state_dict = safetensors.torch.load(mm)
|
| 127 |
+
```
|
| 128 |
+
"""
|
| 129 |
+
entries = {}
|
| 130 |
+
dduf_path = Path(dduf_path)
|
| 131 |
+
logger.info(f"Reading DDUF file {dduf_path}")
|
| 132 |
+
with zipfile.ZipFile(str(dduf_path), "r") as zf:
|
| 133 |
+
for info in zf.infolist():
|
| 134 |
+
logger.debug(f"Reading entry {info.filename}")
|
| 135 |
+
if info.compress_type != zipfile.ZIP_STORED:
|
| 136 |
+
raise DDUFCorruptedFileError("Data must not be compressed in DDUF file.")
|
| 137 |
+
|
| 138 |
+
try:
|
| 139 |
+
_validate_dduf_entry_name(info.filename)
|
| 140 |
+
except DDUFInvalidEntryNameError as e:
|
| 141 |
+
raise DDUFCorruptedFileError(f"Invalid entry name in DDUF file: {info.filename}") from e
|
| 142 |
+
|
| 143 |
+
offset = _get_data_offset(zf, info)
|
| 144 |
+
|
| 145 |
+
entries[info.filename] = DDUFEntry(
|
| 146 |
+
filename=info.filename, offset=offset, length=info.file_size, dduf_path=dduf_path
|
| 147 |
+
)
|
| 148 |
+
|
| 149 |
+
# Consistency checks on the DDUF file
|
| 150 |
+
if "model_index.json" not in entries:
|
| 151 |
+
raise DDUFCorruptedFileError("Missing required 'model_index.json' entry in DDUF file.")
|
| 152 |
+
index = json.loads(entries["model_index.json"].read_text())
|
| 153 |
+
_validate_dduf_structure(index, entries.keys())
|
| 154 |
+
|
| 155 |
+
logger.info(f"Done reading DDUF file {dduf_path}. Found {len(entries)} entries")
|
| 156 |
+
return entries
|
| 157 |
+
|
| 158 |
+
|
| 159 |
+
def export_entries_as_dduf(
|
| 160 |
+
dduf_path: Union[str, os.PathLike], entries: Iterable[Tuple[str, Union[str, Path, bytes]]]
|
| 161 |
+
) -> None:
|
| 162 |
+
"""Write a DDUF file from an iterable of entries.
|
| 163 |
+
|
| 164 |
+
This is a lower-level helper than [`export_folder_as_dduf`] that allows more flexibility when serializing data.
|
| 165 |
+
In particular, you don't need to save the data on disk before exporting it in the DDUF file.
|
| 166 |
+
|
| 167 |
+
Args:
|
| 168 |
+
dduf_path (`str` or `os.PathLike`):
|
| 169 |
+
The path to the DDUF file to write.
|
| 170 |
+
entries (`Iterable[Tuple[str, Union[str, Path, bytes]]]`):
|
| 171 |
+
An iterable of entries to write in the DDUF file. Each entry is a tuple with the filename and the content.
|
| 172 |
+
The filename should be the path to the file in the DDUF archive.
|
| 173 |
+
The content can be a string or a pathlib.Path representing a path to a file on the local disk or directly the content as bytes.
|
| 174 |
+
|
| 175 |
+
Raises:
|
| 176 |
+
- [`DDUFExportError`]: If anything goes wrong during the export (e.g. invalid entry name, missing 'model_index.json', etc.).
|
| 177 |
+
|
| 178 |
+
Example:
|
| 179 |
+
```python
|
| 180 |
+
# Export specific files from the local disk.
|
| 181 |
+
>>> from huggingface_hub import export_entries_as_dduf
|
| 182 |
+
>>> export_entries_as_dduf(
|
| 183 |
+
... dduf_path="stable-diffusion-v1-4-FP16.dduf",
|
| 184 |
+
... entries=[ # List entries to add to the DDUF file (here, only FP16 weights)
|
| 185 |
+
... ("model_index.json", "path/to/model_index.json"),
|
| 186 |
+
... ("vae/config.json", "path/to/vae/config.json"),
|
| 187 |
+
... ("vae/diffusion_pytorch_model.fp16.safetensors", "path/to/vae/diffusion_pytorch_model.fp16.safetensors"),
|
| 188 |
+
... ("text_encoder/config.json", "path/to/text_encoder/config.json"),
|
| 189 |
+
... ("text_encoder/model.fp16.safetensors", "path/to/text_encoder/model.fp16.safetensors"),
|
| 190 |
+
... # ... add more entries here
|
| 191 |
+
... ]
|
| 192 |
+
... )
|
| 193 |
+
```
|
| 194 |
+
|
| 195 |
+
```python
|
| 196 |
+
# Export state_dicts one by one from a loaded pipeline
|
| 197 |
+
>>> from diffusers import DiffusionPipeline
|
| 198 |
+
>>> from typing import Generator, Tuple
|
| 199 |
+
>>> import safetensors.torch
|
| 200 |
+
>>> from huggingface_hub import export_entries_as_dduf
|
| 201 |
+
>>> pipe = DiffusionPipeline.from_pretrained("CompVis/stable-diffusion-v1-4")
|
| 202 |
+
... # ... do some work with the pipeline
|
| 203 |
+
|
| 204 |
+
>>> def as_entries(pipe: DiffusionPipeline) -> Generator[Tuple[str, bytes], None, None]:
|
| 205 |
+
... # Build an generator that yields the entries to add to the DDUF file.
|
| 206 |
+
... # The first element of the tuple is the filename in the DDUF archive (must use UNIX separator!). The second element is the content of the file.
|
| 207 |
+
... # Entries will be evaluated lazily when the DDUF file is created (only 1 entry is loaded in memory at a time)
|
| 208 |
+
... yield "vae/config.json", pipe.vae.to_json_string().encode()
|
| 209 |
+
... yield "vae/diffusion_pytorch_model.safetensors", safetensors.torch.save(pipe.vae.state_dict())
|
| 210 |
+
... yield "text_encoder/config.json", pipe.text_encoder.config.to_json_string().encode()
|
| 211 |
+
... yield "text_encoder/model.safetensors", safetensors.torch.save(pipe.text_encoder.state_dict())
|
| 212 |
+
... # ... add more entries here
|
| 213 |
+
|
| 214 |
+
>>> export_entries_as_dduf(dduf_path="stable-diffusion-v1-4.dduf", entries=as_entries(pipe))
|
| 215 |
+
```
|
| 216 |
+
"""
|
| 217 |
+
logger.info(f"Exporting DDUF file '{dduf_path}'")
|
| 218 |
+
filenames = set()
|
| 219 |
+
index = None
|
| 220 |
+
with zipfile.ZipFile(str(dduf_path), "w", zipfile.ZIP_STORED) as archive:
|
| 221 |
+
for filename, content in entries:
|
| 222 |
+
if filename in filenames:
|
| 223 |
+
raise DDUFExportError(f"Can't add duplicate entry: {filename}")
|
| 224 |
+
filenames.add(filename)
|
| 225 |
+
|
| 226 |
+
if filename == "model_index.json":
|
| 227 |
+
try:
|
| 228 |
+
index = json.loads(_load_content(content).decode())
|
| 229 |
+
except json.JSONDecodeError as e:
|
| 230 |
+
raise DDUFExportError("Failed to parse 'model_index.json'.") from e
|
| 231 |
+
|
| 232 |
+
try:
|
| 233 |
+
filename = _validate_dduf_entry_name(filename)
|
| 234 |
+
except DDUFInvalidEntryNameError as e:
|
| 235 |
+
raise DDUFExportError(f"Invalid entry name: {filename}") from e
|
| 236 |
+
logger.debug(f"Adding entry '{filename}' to DDUF file")
|
| 237 |
+
_dump_content_in_archive(archive, filename, content)
|
| 238 |
+
|
| 239 |
+
# Consistency checks on the DDUF file
|
| 240 |
+
if index is None:
|
| 241 |
+
raise DDUFExportError("Missing required 'model_index.json' entry in DDUF file.")
|
| 242 |
+
try:
|
| 243 |
+
_validate_dduf_structure(index, filenames)
|
| 244 |
+
except DDUFCorruptedFileError as e:
|
| 245 |
+
raise DDUFExportError("Invalid DDUF file structure.") from e
|
| 246 |
+
|
| 247 |
+
logger.info(f"Done writing DDUF file {dduf_path}")
|
| 248 |
+
|
| 249 |
+
|
| 250 |
+
def export_folder_as_dduf(dduf_path: Union[str, os.PathLike], folder_path: Union[str, os.PathLike]) -> None:
|
| 251 |
+
"""
|
| 252 |
+
Export a folder as a DDUF file.
|
| 253 |
+
|
| 254 |
+
AUses [`export_entries_as_dduf`] under the hood.
|
| 255 |
+
|
| 256 |
+
Args:
|
| 257 |
+
dduf_path (`str` or `os.PathLike`):
|
| 258 |
+
The path to the DDUF file to write.
|
| 259 |
+
folder_path (`str` or `os.PathLike`):
|
| 260 |
+
The path to the folder containing the diffusion model.
|
| 261 |
+
|
| 262 |
+
Example:
|
| 263 |
+
```python
|
| 264 |
+
>>> from huggingface_hub import export_folder_as_dduf
|
| 265 |
+
>>> export_folder_as_dduf(dduf_path="FLUX.1-dev.dduf", folder_path="path/to/FLUX.1-dev")
|
| 266 |
+
```
|
| 267 |
+
"""
|
| 268 |
+
folder_path = Path(folder_path)
|
| 269 |
+
|
| 270 |
+
def _iterate_over_folder() -> Iterable[Tuple[str, Path]]:
|
| 271 |
+
for path in Path(folder_path).glob("**/*"):
|
| 272 |
+
if not path.is_file():
|
| 273 |
+
continue
|
| 274 |
+
if path.suffix not in DDUF_ALLOWED_ENTRIES:
|
| 275 |
+
logger.debug(f"Skipping file '{path}' (file type not allowed)")
|
| 276 |
+
continue
|
| 277 |
+
path_in_archive = path.relative_to(folder_path)
|
| 278 |
+
if len(path_in_archive.parts) >= 3:
|
| 279 |
+
logger.debug(f"Skipping file '{path}' (nested directories not allowed)")
|
| 280 |
+
continue
|
| 281 |
+
yield path_in_archive.as_posix(), path
|
| 282 |
+
|
| 283 |
+
export_entries_as_dduf(dduf_path, _iterate_over_folder())
|
| 284 |
+
|
| 285 |
+
|
| 286 |
+
def _dump_content_in_archive(archive: zipfile.ZipFile, filename: str, content: Union[str, os.PathLike, bytes]) -> None:
|
| 287 |
+
with archive.open(filename, "w", force_zip64=True) as archive_fh:
|
| 288 |
+
if isinstance(content, (str, Path)):
|
| 289 |
+
content_path = Path(content)
|
| 290 |
+
with content_path.open("rb") as content_fh:
|
| 291 |
+
shutil.copyfileobj(content_fh, archive_fh, 1024 * 1024 * 8) # type: ignore[misc]
|
| 292 |
+
elif isinstance(content, bytes):
|
| 293 |
+
archive_fh.write(content)
|
| 294 |
+
else:
|
| 295 |
+
raise DDUFExportError(f"Invalid content type for {filename}. Must be str, Path or bytes.")
|
| 296 |
+
|
| 297 |
+
|
| 298 |
+
def _load_content(content: Union[str, Path, bytes]) -> bytes:
|
| 299 |
+
"""Load the content of an entry as bytes.
|
| 300 |
+
|
| 301 |
+
Used only for small checks (not to dump content into archive).
|
| 302 |
+
"""
|
| 303 |
+
if isinstance(content, (str, Path)):
|
| 304 |
+
return Path(content).read_bytes()
|
| 305 |
+
elif isinstance(content, bytes):
|
| 306 |
+
return content
|
| 307 |
+
else:
|
| 308 |
+
raise DDUFExportError(f"Invalid content type. Must be str, Path or bytes. Got {type(content)}.")
|
| 309 |
+
|
| 310 |
+
|
| 311 |
+
def _validate_dduf_entry_name(entry_name: str) -> str:
|
| 312 |
+
if "." + entry_name.split(".")[-1] not in DDUF_ALLOWED_ENTRIES:
|
| 313 |
+
raise DDUFInvalidEntryNameError(f"File type not allowed: {entry_name}")
|
| 314 |
+
if "\\" in entry_name:
|
| 315 |
+
raise DDUFInvalidEntryNameError(f"Entry names must use UNIX separators ('/'). Got {entry_name}.")
|
| 316 |
+
entry_name = entry_name.strip("/")
|
| 317 |
+
if entry_name.count("/") > 1:
|
| 318 |
+
raise DDUFInvalidEntryNameError(f"DDUF only supports 1 level of directory. Got {entry_name}.")
|
| 319 |
+
return entry_name
|
| 320 |
+
|
| 321 |
+
|
| 322 |
+
def _validate_dduf_structure(index: Any, entry_names: Iterable[str]) -> None:
|
| 323 |
+
"""
|
| 324 |
+
Consistency checks on the DDUF file structure.
|
| 325 |
+
|
| 326 |
+
Rules:
|
| 327 |
+
- The 'model_index.json' entry is required and must contain a dictionary.
|
| 328 |
+
- Each folder name must correspond to an entry in 'model_index.json'.
|
| 329 |
+
- Each folder must contain at least a config file ('config.json', 'tokenizer_config.json', 'preprocessor_config.json', 'scheduler_config.json').
|
| 330 |
+
|
| 331 |
+
Args:
|
| 332 |
+
index (Any):
|
| 333 |
+
The content of the 'model_index.json' entry.
|
| 334 |
+
entry_names (Iterable[str]):
|
| 335 |
+
The list of entry names in the DDUF file.
|
| 336 |
+
|
| 337 |
+
Raises:
|
| 338 |
+
- [`DDUFCorruptedFileError`]: If the DDUF file is corrupted (i.e. doesn't follow the DDUF format).
|
| 339 |
+
"""
|
| 340 |
+
if not isinstance(index, dict):
|
| 341 |
+
raise DDUFCorruptedFileError(f"Invalid 'model_index.json' content. Must be a dictionary. Got {type(index)}.")
|
| 342 |
+
|
| 343 |
+
dduf_folders = {entry.split("/")[0] for entry in entry_names if "/" in entry}
|
| 344 |
+
for folder in dduf_folders:
|
| 345 |
+
if folder not in index:
|
| 346 |
+
raise DDUFCorruptedFileError(f"Missing required entry '{folder}' in 'model_index.json'.")
|
| 347 |
+
if not any(f"{folder}/{required_entry}" in entry_names for required_entry in DDUF_FOLDER_REQUIRED_ENTRIES):
|
| 348 |
+
raise DDUFCorruptedFileError(
|
| 349 |
+
f"Missing required file in folder '{folder}'. Must contains at least one of {DDUF_FOLDER_REQUIRED_ENTRIES}."
|
| 350 |
+
)
|
| 351 |
+
|
| 352 |
+
|
| 353 |
+
def _get_data_offset(zf: zipfile.ZipFile, info: zipfile.ZipInfo) -> int:
|
| 354 |
+
"""
|
| 355 |
+
Calculate the data offset for a file in a ZIP archive.
|
| 356 |
+
|
| 357 |
+
Args:
|
| 358 |
+
zf (`zipfile.ZipFile`):
|
| 359 |
+
The opened ZIP file. Must be opened in read mode.
|
| 360 |
+
info (`zipfile.ZipInfo`):
|
| 361 |
+
The file info.
|
| 362 |
+
|
| 363 |
+
Returns:
|
| 364 |
+
int: The offset of the file data in the ZIP archive.
|
| 365 |
+
"""
|
| 366 |
+
if zf.fp is None:
|
| 367 |
+
raise DDUFCorruptedFileError("ZipFile object must be opened in read mode.")
|
| 368 |
+
|
| 369 |
+
# Step 1: Get the local file header offset
|
| 370 |
+
header_offset = info.header_offset
|
| 371 |
+
|
| 372 |
+
# Step 2: Read the local file header
|
| 373 |
+
zf.fp.seek(header_offset)
|
| 374 |
+
local_file_header = zf.fp.read(30) # Fixed-size part of the local header
|
| 375 |
+
|
| 376 |
+
if len(local_file_header) < 30:
|
| 377 |
+
raise DDUFCorruptedFileError("Incomplete local file header.")
|
| 378 |
+
|
| 379 |
+
# Step 3: Parse the header fields to calculate the start of file data
|
| 380 |
+
# Local file header: https://en.wikipedia.org/wiki/ZIP_(file_format)#File_headers
|
| 381 |
+
filename_len = int.from_bytes(local_file_header[26:28], "little")
|
| 382 |
+
extra_field_len = int.from_bytes(local_file_header[28:30], "little")
|
| 383 |
+
|
| 384 |
+
# Data offset is after the fixed header, filename, and extra fields
|
| 385 |
+
data_offset = header_offset + 30 + filename_len + extra_field_len
|
| 386 |
+
|
| 387 |
+
return data_offset
|
pythonProject/.venv/Lib/site-packages/huggingface_hub/serialization/_tensorflow.py
ADDED
|
@@ -0,0 +1,95 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright 2024 The HuggingFace Team. All rights reserved.
|
| 2 |
+
#
|
| 3 |
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
| 4 |
+
# you may not use this file except in compliance with the License.
|
| 5 |
+
# You may obtain a copy of the License at
|
| 6 |
+
#
|
| 7 |
+
# http://www.apache.org/licenses/LICENSE-2.0
|
| 8 |
+
#
|
| 9 |
+
# Unless required by applicable law or agreed to in writing, software
|
| 10 |
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
| 11 |
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
| 12 |
+
# See the License for the specific language governing permissions and
|
| 13 |
+
# limitations under the License.
|
| 14 |
+
"""Contains tensorflow-specific helpers."""
|
| 15 |
+
|
| 16 |
+
import math
|
| 17 |
+
import re
|
| 18 |
+
from typing import TYPE_CHECKING, Dict, Union
|
| 19 |
+
|
| 20 |
+
from .. import constants
|
| 21 |
+
from ._base import MAX_SHARD_SIZE, StateDictSplit, split_state_dict_into_shards_factory
|
| 22 |
+
|
| 23 |
+
|
| 24 |
+
if TYPE_CHECKING:
|
| 25 |
+
import tensorflow as tf
|
| 26 |
+
|
| 27 |
+
|
| 28 |
+
def split_tf_state_dict_into_shards(
|
| 29 |
+
state_dict: Dict[str, "tf.Tensor"],
|
| 30 |
+
*,
|
| 31 |
+
filename_pattern: str = constants.TF2_WEIGHTS_FILE_PATTERN,
|
| 32 |
+
max_shard_size: Union[int, str] = MAX_SHARD_SIZE,
|
| 33 |
+
) -> StateDictSplit:
|
| 34 |
+
"""
|
| 35 |
+
Split a model state dictionary in shards so that each shard is smaller than a given size.
|
| 36 |
+
|
| 37 |
+
The shards are determined by iterating through the `state_dict` in the order of its keys. There is no optimization
|
| 38 |
+
made to make each shard as close as possible to the maximum size passed. For example, if the limit is 10GB and we
|
| 39 |
+
have tensors of sizes [6GB, 6GB, 2GB, 6GB, 2GB, 2GB] they will get sharded as [6GB], [6+2GB], [6+2+2GB] and not
|
| 40 |
+
[6+2+2GB], [6+2GB], [6GB].
|
| 41 |
+
|
| 42 |
+
<Tip warning={true}>
|
| 43 |
+
|
| 44 |
+
If one of the model's tensor is bigger than `max_shard_size`, it will end up in its own shard which will have a
|
| 45 |
+
size greater than `max_shard_size`.
|
| 46 |
+
|
| 47 |
+
</Tip>
|
| 48 |
+
|
| 49 |
+
Args:
|
| 50 |
+
state_dict (`Dict[str, Tensor]`):
|
| 51 |
+
The state dictionary to save.
|
| 52 |
+
filename_pattern (`str`, *optional*):
|
| 53 |
+
The pattern to generate the files names in which the model will be saved. Pattern must be a string that
|
| 54 |
+
can be formatted with `filename_pattern.format(suffix=...)` and must contain the keyword `suffix`
|
| 55 |
+
Defaults to `"tf_model{suffix}.h5"`.
|
| 56 |
+
max_shard_size (`int` or `str`, *optional*):
|
| 57 |
+
The maximum size of each shard, in bytes. Defaults to 5GB.
|
| 58 |
+
|
| 59 |
+
Returns:
|
| 60 |
+
[`StateDictSplit`]: A `StateDictSplit` object containing the shards and the index to retrieve them.
|
| 61 |
+
"""
|
| 62 |
+
return split_state_dict_into_shards_factory(
|
| 63 |
+
state_dict,
|
| 64 |
+
max_shard_size=max_shard_size,
|
| 65 |
+
filename_pattern=filename_pattern,
|
| 66 |
+
get_storage_size=get_tf_storage_size,
|
| 67 |
+
)
|
| 68 |
+
|
| 69 |
+
|
| 70 |
+
def get_tf_storage_size(tensor: "tf.Tensor") -> int:
|
| 71 |
+
# Return `math.ceil` since dtype byte size can be a float (e.g., 0.125 for tf.bool).
|
| 72 |
+
# Better to overestimate than underestimate.
|
| 73 |
+
return math.ceil(tensor.numpy().size * _dtype_byte_size_tf(tensor.dtype))
|
| 74 |
+
|
| 75 |
+
|
| 76 |
+
def _dtype_byte_size_tf(dtype) -> float:
|
| 77 |
+
"""
|
| 78 |
+
Returns the size (in bytes) occupied by one parameter of type `dtype`.
|
| 79 |
+
Taken from https://github.com/huggingface/transformers/blob/74d9d0cebb0263a3f8ab9c280569170cc74651d0/src/transformers/modeling_tf_utils.py#L608.
|
| 80 |
+
NOTE: why not `tensor.numpy().nbytes`?
|
| 81 |
+
Example:
|
| 82 |
+
```py
|
| 83 |
+
>>> _dtype_byte_size(tf.float32)
|
| 84 |
+
4
|
| 85 |
+
```
|
| 86 |
+
"""
|
| 87 |
+
import tensorflow as tf
|
| 88 |
+
|
| 89 |
+
if dtype == tf.bool:
|
| 90 |
+
return 1 / 8
|
| 91 |
+
bit_search = re.search(r"[^\d](\d+)$", dtype.name)
|
| 92 |
+
if bit_search is None:
|
| 93 |
+
raise ValueError(f"`dtype` is not a valid dtype: {dtype}.")
|
| 94 |
+
bit_size = int(bit_search.groups()[0])
|
| 95 |
+
return bit_size // 8
|