serverdaun commited on
Commit
fe610d1
·
1 Parent(s): 1b4d5e8

add reqs file for hf

Browse files
Files changed (1) hide show
  1. requirements.txt +544 -0
requirements.txt ADDED
@@ -0,0 +1,544 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # This file was autogenerated by uv via the following command:
2
+ # uv export --frozen --no-hashes -o requirements.txt
3
+ aiofiles==24.1.0
4
+ # via gradio
5
+ aiohappyeyeballs==2.6.1
6
+ # via aiohttp
7
+ aiohttp==3.12.15
8
+ # via
9
+ # huggingface-hub
10
+ # langchain-community
11
+ # llama-index-core
12
+ aiosignal==1.4.0
13
+ # via aiohttp
14
+ aiosqlite==0.21.0
15
+ # via llama-index-core
16
+ annotated-types==0.7.0
17
+ # via pydantic
18
+ anyio==4.10.0
19
+ # via
20
+ # gradio
21
+ # httpx
22
+ # openai
23
+ # starlette
24
+ attrs==25.3.0
25
+ # via aiohttp
26
+ audioop-lts==0.2.2 ; python_full_version >= '3.13'
27
+ # via gradio
28
+ banks==2.2.0
29
+ # via llama-index-core
30
+ beautifulsoup4==4.13.4
31
+ # via llama-index-readers-file
32
+ black==25.1.0
33
+ # via rag-w-binary-quant
34
+ brotli==1.1.0
35
+ # via gradio
36
+ certifi==2025.8.3
37
+ # via
38
+ # httpcore
39
+ # httpx
40
+ # llama-cloud
41
+ # requests
42
+ cffi==1.17.1 ; platform_python_implementation == 'PyPy'
43
+ # via zstandard
44
+ charset-normalizer==3.4.2
45
+ # via requests
46
+ click==8.2.1
47
+ # via
48
+ # black
49
+ # llama-cloud-services
50
+ # nltk
51
+ # typer
52
+ # uvicorn
53
+ colorama==0.4.6
54
+ # via
55
+ # click
56
+ # griffe
57
+ # tqdm
58
+ dataclasses-json==0.6.7
59
+ # via
60
+ # langchain-community
61
+ # llama-index-core
62
+ defusedxml==0.7.1
63
+ # via llama-index-readers-file
64
+ deprecated==1.2.18
65
+ # via
66
+ # banks
67
+ # llama-index-core
68
+ # llama-index-instrumentation
69
+ dirtyjson==1.0.8
70
+ # via llama-index-core
71
+ distro==1.9.0
72
+ # via openai
73
+ docx2txt==0.9
74
+ # via rag-w-binary-quant
75
+ dotenv==0.9.9
76
+ # via rag-w-binary-quant
77
+ fastapi==0.116.1
78
+ # via gradio
79
+ ffmpy==0.6.1
80
+ # via gradio
81
+ filelock==3.18.0
82
+ # via
83
+ # huggingface-hub
84
+ # torch
85
+ # transformers
86
+ filetype==1.2.0
87
+ # via llama-index-core
88
+ frozenlist==1.7.0
89
+ # via
90
+ # aiohttp
91
+ # aiosignal
92
+ fsspec==2025.7.0
93
+ # via
94
+ # gradio-client
95
+ # huggingface-hub
96
+ # llama-index-core
97
+ # torch
98
+ gradio==5.41.0
99
+ # via rag-w-binary-quant
100
+ gradio-client==1.11.0
101
+ # via gradio
102
+ greenlet==3.2.3
103
+ # via sqlalchemy
104
+ griffe==1.9.0
105
+ # via banks
106
+ groovy==0.1.2
107
+ # via gradio
108
+ grpcio==1.67.1
109
+ # via pymilvus
110
+ h11==0.16.0
111
+ # via
112
+ # httpcore
113
+ # uvicorn
114
+ hf-xet==1.1.5 ; platform_machine == 'aarch64' or platform_machine == 'amd64' or platform_machine == 'arm64' or platform_machine == 'x86_64'
115
+ # via huggingface-hub
116
+ httpcore==1.0.9
117
+ # via httpx
118
+ httpx==0.28.1
119
+ # via
120
+ # gradio
121
+ # gradio-client
122
+ # langsmith
123
+ # llama-cloud
124
+ # llama-index-core
125
+ # openai
126
+ # safehttpx
127
+ httpx-sse==0.4.1
128
+ # via langchain-community
129
+ huggingface-hub==0.34.3
130
+ # via
131
+ # gradio
132
+ # gradio-client
133
+ # llama-index-embeddings-huggingface
134
+ # sentence-transformers
135
+ # tokenizers
136
+ # transformers
137
+ idna==3.10
138
+ # via
139
+ # anyio
140
+ # httpx
141
+ # requests
142
+ # yarl
143
+ isort==6.0.1
144
+ # via rag-w-binary-quant
145
+ jinja2==3.1.6
146
+ # via
147
+ # banks
148
+ # gradio
149
+ # torch
150
+ jiter==0.10.0
151
+ # via openai
152
+ joblib==1.5.1
153
+ # via
154
+ # nltk
155
+ # scikit-learn
156
+ jsonpatch==1.33
157
+ # via langchain-core
158
+ jsonpointer==3.0.0
159
+ # via jsonpatch
160
+ langchain==0.3.27
161
+ # via
162
+ # langchain-community
163
+ # rag-w-binary-quant
164
+ langchain-community==0.3.27
165
+ # via rag-w-binary-quant
166
+ langchain-core==0.3.72
167
+ # via
168
+ # langchain
169
+ # langchain-community
170
+ # langchain-openai
171
+ # langchain-text-splitters
172
+ langchain-openai==0.3.28
173
+ # via rag-w-binary-quant
174
+ langchain-text-splitters==0.3.9
175
+ # via langchain
176
+ langsmith==0.4.10
177
+ # via
178
+ # langchain
179
+ # langchain-community
180
+ # langchain-core
181
+ llama-cloud==0.1.35
182
+ # via
183
+ # llama-cloud-services
184
+ # llama-index-indices-managed-llama-cloud
185
+ llama-cloud-services==0.6.54
186
+ # via llama-parse
187
+ llama-index==0.13.0
188
+ # via rag-w-binary-quant
189
+ llama-index-cli==0.5.0
190
+ # via llama-index
191
+ llama-index-core==0.13.0
192
+ # via
193
+ # llama-cloud-services
194
+ # llama-index
195
+ # llama-index-cli
196
+ # llama-index-embeddings-huggingface
197
+ # llama-index-embeddings-openai
198
+ # llama-index-indices-managed-llama-cloud
199
+ # llama-index-llms-openai
200
+ # llama-index-readers-file
201
+ # llama-index-readers-llama-parse
202
+ llama-index-embeddings-huggingface==0.6.0
203
+ # via rag-w-binary-quant
204
+ llama-index-embeddings-openai==0.5.0
205
+ # via
206
+ # llama-index
207
+ # llama-index-cli
208
+ llama-index-indices-managed-llama-cloud==0.9.0
209
+ # via llama-index
210
+ llama-index-instrumentation==0.4.0
211
+ # via llama-index-workflows
212
+ llama-index-llms-openai==0.5.0
213
+ # via
214
+ # llama-index
215
+ # llama-index-cli
216
+ llama-index-readers-file==0.5.0
217
+ # via llama-index
218
+ llama-index-readers-llama-parse==0.5.0
219
+ # via llama-index
220
+ llama-index-workflows==1.2.0
221
+ # via llama-index-core
222
+ llama-parse==0.6.54
223
+ # via llama-index-readers-llama-parse
224
+ logging==0.4.9.6
225
+ # via rag-w-binary-quant
226
+ markdown-it-py==3.0.0 ; sys_platform != 'emscripten'
227
+ # via rich
228
+ markupsafe==3.0.2
229
+ # via
230
+ # gradio
231
+ # jinja2
232
+ marshmallow==3.26.1
233
+ # via dataclasses-json
234
+ mdurl==0.1.2 ; sys_platform != 'emscripten'
235
+ # via markdown-it-py
236
+ milvus-lite==2.5.1 ; sys_platform != 'win32'
237
+ # via pymilvus
238
+ mpmath==1.3.0
239
+ # via sympy
240
+ multidict==6.6.3
241
+ # via
242
+ # aiohttp
243
+ # yarl
244
+ mypy-extensions==1.1.0
245
+ # via
246
+ # black
247
+ # typing-inspect
248
+ nest-asyncio==1.6.0
249
+ # via llama-index-core
250
+ networkx==3.5
251
+ # via
252
+ # llama-index-core
253
+ # torch
254
+ nltk==3.9.1
255
+ # via
256
+ # llama-index
257
+ # llama-index-core
258
+ numpy==2.3.2
259
+ # via
260
+ # gradio
261
+ # langchain-community
262
+ # llama-index-core
263
+ # pandas
264
+ # rag-w-binary-quant
265
+ # scikit-learn
266
+ # scipy
267
+ # transformers
268
+ nvidia-cublas-cu12==12.6.4.1 ; platform_machine == 'x86_64' and sys_platform == 'linux'
269
+ # via
270
+ # nvidia-cudnn-cu12
271
+ # nvidia-cusolver-cu12
272
+ # torch
273
+ nvidia-cuda-cupti-cu12==12.6.80 ; platform_machine == 'x86_64' and sys_platform == 'linux'
274
+ # via torch
275
+ nvidia-cuda-nvrtc-cu12==12.6.77 ; platform_machine == 'x86_64' and sys_platform == 'linux'
276
+ # via torch
277
+ nvidia-cuda-runtime-cu12==12.6.77 ; platform_machine == 'x86_64' and sys_platform == 'linux'
278
+ # via torch
279
+ nvidia-cudnn-cu12==9.5.1.17 ; platform_machine == 'x86_64' and sys_platform == 'linux'
280
+ # via torch
281
+ nvidia-cufft-cu12==11.3.0.4 ; platform_machine == 'x86_64' and sys_platform == 'linux'
282
+ # via torch
283
+ nvidia-cufile-cu12==1.11.1.6 ; platform_machine == 'x86_64' and sys_platform == 'linux'
284
+ # via torch
285
+ nvidia-curand-cu12==10.3.7.77 ; platform_machine == 'x86_64' and sys_platform == 'linux'
286
+ # via torch
287
+ nvidia-cusolver-cu12==11.7.1.2 ; platform_machine == 'x86_64' and sys_platform == 'linux'
288
+ # via torch
289
+ nvidia-cusparse-cu12==12.5.4.2 ; platform_machine == 'x86_64' and sys_platform == 'linux'
290
+ # via
291
+ # nvidia-cusolver-cu12
292
+ # torch
293
+ nvidia-cusparselt-cu12==0.6.3 ; platform_machine == 'x86_64' and sys_platform == 'linux'
294
+ # via torch
295
+ nvidia-nccl-cu12==2.26.2 ; platform_machine == 'x86_64' and sys_platform == 'linux'
296
+ # via torch
297
+ nvidia-nvjitlink-cu12==12.6.85 ; platform_machine == 'x86_64' and sys_platform == 'linux'
298
+ # via
299
+ # nvidia-cufft-cu12
300
+ # nvidia-cusolver-cu12
301
+ # nvidia-cusparse-cu12
302
+ # torch
303
+ nvidia-nvtx-cu12==12.6.77 ; platform_machine == 'x86_64' and sys_platform == 'linux'
304
+ # via torch
305
+ openai==1.98.0
306
+ # via
307
+ # langchain-openai
308
+ # llama-index-embeddings-openai
309
+ # llama-index-llms-openai
310
+ orjson==3.11.1
311
+ # via
312
+ # gradio
313
+ # langsmith
314
+ packaging==25.0
315
+ # via
316
+ # black
317
+ # gradio
318
+ # gradio-client
319
+ # huggingface-hub
320
+ # langchain-core
321
+ # langsmith
322
+ # marshmallow
323
+ # transformers
324
+ pandas==2.2.3
325
+ # via
326
+ # gradio
327
+ # llama-index-readers-file
328
+ # pymilvus
329
+ pathspec==0.12.1
330
+ # via black
331
+ pillow==11.3.0
332
+ # via
333
+ # gradio
334
+ # llama-index-core
335
+ # sentence-transformers
336
+ platformdirs==4.3.8
337
+ # via
338
+ # banks
339
+ # black
340
+ # llama-cloud-services
341
+ # llama-index-core
342
+ propcache==0.3.2
343
+ # via
344
+ # aiohttp
345
+ # yarl
346
+ protobuf==6.31.1
347
+ # via pymilvus
348
+ pycparser==2.22 ; platform_python_implementation == 'PyPy'
349
+ # via cffi
350
+ pydantic==2.11.7
351
+ # via
352
+ # banks
353
+ # fastapi
354
+ # gradio
355
+ # langchain
356
+ # langchain-core
357
+ # langsmith
358
+ # llama-cloud
359
+ # llama-cloud-services
360
+ # llama-index-core
361
+ # llama-index-instrumentation
362
+ # llama-index-workflows
363
+ # openai
364
+ # pydantic-settings
365
+ pydantic-core==2.33.2
366
+ # via pydantic
367
+ pydantic-settings==2.10.1
368
+ # via langchain-community
369
+ pydub==0.25.1
370
+ # via gradio
371
+ pygments==2.19.2 ; sys_platform != 'emscripten'
372
+ # via rich
373
+ pymilvus==2.5.14
374
+ # via rag-w-binary-quant
375
+ pypdf==5.9.0
376
+ # via llama-index-readers-file
377
+ python-dateutil==2.9.0.post0
378
+ # via pandas
379
+ python-dotenv==1.1.1
380
+ # via
381
+ # dotenv
382
+ # llama-cloud-services
383
+ # pydantic-settings
384
+ # pymilvus
385
+ python-multipart==0.0.20
386
+ # via gradio
387
+ pytz==2025.2
388
+ # via pandas
389
+ pyyaml==6.0.2
390
+ # via
391
+ # gradio
392
+ # huggingface-hub
393
+ # langchain
394
+ # langchain-community
395
+ # langchain-core
396
+ # llama-index-core
397
+ # transformers
398
+ regex==2025.7.34
399
+ # via
400
+ # nltk
401
+ # tiktoken
402
+ # transformers
403
+ requests==2.32.4
404
+ # via
405
+ # huggingface-hub
406
+ # langchain
407
+ # langchain-community
408
+ # langsmith
409
+ # llama-index-core
410
+ # requests-toolbelt
411
+ # tiktoken
412
+ # transformers
413
+ requests-toolbelt==1.0.0
414
+ # via langsmith
415
+ rich==14.1.0 ; sys_platform != 'emscripten'
416
+ # via typer
417
+ ruff==0.12.7 ; sys_platform != 'emscripten'
418
+ # via gradio
419
+ safehttpx==0.1.6
420
+ # via gradio
421
+ safetensors==0.5.3
422
+ # via transformers
423
+ scikit-learn==1.7.1
424
+ # via sentence-transformers
425
+ scipy==1.16.1
426
+ # via
427
+ # scikit-learn
428
+ # sentence-transformers
429
+ semantic-version==2.10.0
430
+ # via gradio
431
+ sentence-transformers==5.0.0
432
+ # via llama-index-embeddings-huggingface
433
+ setuptools==80.9.0
434
+ # via
435
+ # llama-index-core
436
+ # pymilvus
437
+ # torch
438
+ # triton
439
+ shellingham==1.5.4 ; sys_platform != 'emscripten'
440
+ # via typer
441
+ six==1.17.0
442
+ # via python-dateutil
443
+ sniffio==1.3.1
444
+ # via
445
+ # anyio
446
+ # openai
447
+ soupsieve==2.7
448
+ # via beautifulsoup4
449
+ sqlalchemy==2.0.42
450
+ # via
451
+ # langchain
452
+ # langchain-community
453
+ # llama-index-core
454
+ starlette==0.47.2
455
+ # via
456
+ # fastapi
457
+ # gradio
458
+ striprtf==0.0.26
459
+ # via llama-index-readers-file
460
+ sympy==1.14.0
461
+ # via torch
462
+ tenacity==9.1.2
463
+ # via
464
+ # langchain-community
465
+ # langchain-core
466
+ # llama-cloud-services
467
+ # llama-index-core
468
+ threadpoolctl==3.6.0
469
+ # via scikit-learn
470
+ tiktoken==0.9.0
471
+ # via
472
+ # langchain-openai
473
+ # llama-index-core
474
+ tokenizers==0.21.4
475
+ # via transformers
476
+ tomlkit==0.13.3
477
+ # via gradio
478
+ torch==2.7.1
479
+ # via sentence-transformers
480
+ tqdm==4.67.1
481
+ # via
482
+ # huggingface-hub
483
+ # llama-index-core
484
+ # milvus-lite
485
+ # nltk
486
+ # openai
487
+ # sentence-transformers
488
+ # transformers
489
+ transformers==4.54.1
490
+ # via sentence-transformers
491
+ triton==3.3.1 ; platform_machine == 'x86_64' and sys_platform == 'linux'
492
+ # via torch
493
+ typer==0.16.0 ; sys_platform != 'emscripten'
494
+ # via gradio
495
+ typing-extensions==4.14.1
496
+ # via
497
+ # aiosignal
498
+ # aiosqlite
499
+ # anyio
500
+ # beautifulsoup4
501
+ # fastapi
502
+ # gradio
503
+ # gradio-client
504
+ # huggingface-hub
505
+ # langchain-core
506
+ # llama-index-core
507
+ # openai
508
+ # pydantic
509
+ # pydantic-core
510
+ # sentence-transformers
511
+ # sqlalchemy
512
+ # starlette
513
+ # torch
514
+ # typer
515
+ # typing-inspect
516
+ # typing-inspection
517
+ typing-inspect==0.9.0
518
+ # via
519
+ # dataclasses-json
520
+ # llama-index-core
521
+ typing-inspection==0.4.1
522
+ # via
523
+ # pydantic
524
+ # pydantic-settings
525
+ tzdata==2025.2
526
+ # via pandas
527
+ ujson==5.10.0
528
+ # via pymilvus
529
+ urllib3==2.5.0
530
+ # via
531
+ # gradio
532
+ # requests
533
+ uvicorn==0.35.0 ; sys_platform != 'emscripten'
534
+ # via gradio
535
+ websockets==15.0.1
536
+ # via gradio-client
537
+ wrapt==1.17.2
538
+ # via
539
+ # deprecated
540
+ # llama-index-core
541
+ yarl==1.20.1
542
+ # via aiohttp
543
+ zstandard==0.23.0
544
+ # via langsmith