Spaces:
Sleeping
Sleeping
Upload folder using huggingface_hub
Browse files- app/app.py +1 -1
- src/extractor.py +2 -2
app/app.py
CHANGED
|
@@ -1,6 +1,6 @@
|
|
| 1 |
"""
|
| 2 |
Financial Document Extractor — FastAPI Web Server
|
| 3 |
-
Hybrid Pipeline: Docling (PDF→MD) + Groq (llama-3.
|
| 4 |
"""
|
| 5 |
|
| 6 |
import os
|
|
|
|
| 1 |
"""
|
| 2 |
Financial Document Extractor — FastAPI Web Server
|
| 3 |
+
Hybrid Pipeline: Docling (PDF→MD) + Groq (llama-3.3-70b-versatile)
|
| 4 |
"""
|
| 5 |
|
| 6 |
import os
|
src/extractor.py
CHANGED
|
@@ -2,7 +2,7 @@
|
|
| 2 |
Financial Document Extraction Pipeline — Groq Edition.
|
| 3 |
|
| 4 |
Replaces the local llama-cpp-python / GGUF inference with a
|
| 5 |
-
server-side Groq API call to llama-3.
|
| 6 |
|
| 7 |
Features:
|
| 8 |
- Zero GPU / VRAM requirements — pure API call
|
|
@@ -127,7 +127,7 @@ class FinancialDocExtractor:
|
|
| 127 |
|
| 128 |
def __init__(
|
| 129 |
self,
|
| 130 |
-
model: str = "llama-3.
|
| 131 |
max_tokens: int = 4096,
|
| 132 |
temperature: float = 0.1,
|
| 133 |
max_retries: int = 3,
|
|
|
|
| 2 |
Financial Document Extraction Pipeline — Groq Edition.
|
| 3 |
|
| 4 |
Replaces the local llama-cpp-python / GGUF inference with a
|
| 5 |
+
server-side Groq API call to llama-3.3-70b-versatile.
|
| 6 |
|
| 7 |
Features:
|
| 8 |
- Zero GPU / VRAM requirements — pure API call
|
|
|
|
| 127 |
|
| 128 |
def __init__(
|
| 129 |
self,
|
| 130 |
+
model: str = "llama-3.3-70b-versatile",
|
| 131 |
max_tokens: int = 4096,
|
| 132 |
temperature: float = 0.1,
|
| 133 |
max_retries: int = 3,
|