smartytrios commited on
Commit
7e14c40
·
verified ·
1 Parent(s): 2099d25

updated readme.md

Browse files
Files changed (1) hide show
  1. README.md +49 -1
README.md CHANGED
@@ -3,7 +3,14 @@ tags:
3
  - gguf
4
  - llama.cpp
5
  - unsloth
6
-
 
 
 
 
 
 
 
7
  ---
8
 
9
  # docintel_ocr_llama_3_2_gguf : GGUF
@@ -21,3 +28,44 @@ This model was finetuned and converted to GGUF format using [Unsloth](https://gi
21
  An Ollama Modelfile is included for easy deployment.
22
  This was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth)
23
  [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  - gguf
4
  - llama.cpp
5
  - unsloth
6
+ license: mit
7
+ datasets:
8
+ - smartytrios/document_data_extractor
9
+ language:
10
+ - en
11
+ base_model:
12
+ - unsloth/Llama-3.2-1B-Instruct-bnb-4bit
13
+ pipeline_tag: zero-shot-classification
14
  ---
15
 
16
  # docintel_ocr_llama_3_2_gguf : GGUF
 
28
  An Ollama Modelfile is included for easy deployment.
29
  This was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth)
30
  [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
31
+
32
+ ---
33
+ tags:
34
+ - gguf
35
+ - llama.cpp
36
+ - unsloth
37
+ - ocr
38
+ - document-intelligence
39
+ - json-extraction
40
+ license: mit
41
+ datasets:
42
+ - smartytrios/document_data_extractor
43
+ language:
44
+ - en
45
+ base_model:
46
+ - unsloth/Llama-3.2-1B-Instruct-bnb-4bit
47
+ pipeline_tag: text-generation
48
+ library_name: transformers
49
+ ---
50
+
51
+ # docintel_ocr_llama_3_2_gguf : GGUF Optimized
52
+
53
+ This model is a fine-tuned version of **Llama-3.2-1B-Instruct**, specialized for **Document Intelligence** and **OCR-to-JSON** extraction. It was trained using the [Unsloth](https://github.com/unslothai/unsloth) library to optimize memory efficiency and training speed, then exported to GGUF format for local deployment.
54
+
55
+ ## Model Description
56
+ The primary objective of this model is to transform unstructured text generated by Optical Character Recognition (OCR) engines into structured, machine-readable JSON formats. It is specifically tuned to handle noise, line breaks (`\n`), and misalignments common in raw OCR data.
57
+
58
+ - **Architecture:** Llama 3.2 (1B Parameters)
59
+ - **Quantization:** Q4_K_M (4-bit Medium)
60
+ - **Specialization:** Invoice/Receipt data extraction, medical bill parsing, and form field mapping.
61
+ - **Fine-tuning Method:** QLoRA (Rank: 16)
62
+
63
+ ---
64
+
65
+ ## 🚀 Usage Guide
66
+
67
+ ### 1. Local Inference with llama.cpp
68
+ For the best performance on Windows, Mac, or Linux using `llama.cpp`, use the following command:
69
+
70
+ ```bash
71
+ ./llama-cli -hf smartytrios/docintel_ocr_llama_3_2_gguf --jinja -p "### OCR:\n[PASTE YOUR OCR TEXT HERE]\n### JSON:"