donribbs commited on
Commit
19d924c
·
verified ·
1 Parent(s): 817ba60

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -5
README.md CHANGED
@@ -12,7 +12,7 @@ pipeline_tag: text-generation
12
 
13
  ## 🧠 Training Configuration
14
 
15
- This model was trained using the configuration file [`configs/train_small.yaml`](https://huggingface.co/donribbs/scraps-llm-model/blob/main/configs/train_small.yaml).
16
 
17
  ### ⚙️ Model Architecture
18
  | Parameter | Value | Description |
@@ -68,7 +68,7 @@ This model was trained using the configuration file [`configs/train_small.yaml`]
68
  | **Platform** | Google Colab Pro+ |
69
  | **GPU** | NVIDIA A100 (80 GB) |
70
  | **Runtime** | PyTorch 2.x |
71
- | **Training time** | ~4–5 hours |
72
  | **Mixed precision** | Enabled (AMP) |
73
 
74
  ---
@@ -83,7 +83,7 @@ This model was trained using the configuration file [`configs/train_small.yaml`]
83
  ---
84
 
85
  ### 🧩 Summary
86
- Scraps-LLM is a **136 M-parameter decoder-only Transformer** trained to generate complete cooking recipes from a list of input ingredients.
87
  The model learns via **causal language modeling (next-token prediction)** on RecipeNLG data, producing structured, human-readable recipes that include titles and numbered steps.
88
  It was later exported to **ONNX** for lightweight CPU inference and integrated into a Hugging Face Space demo.
89
 
@@ -102,5 +102,4 @@ It was later exported to **ONNX** for lightweight CPU inference and integrated i
102
  |------|-------------|
103
  | `export/scraps.onnx` | ONNX-optimized inference graph |
104
  | `tokenizer/bpe.json` | BPE vocabulary used for encoding/decoding |
105
- | `configs/train_small.yaml` | Original training configuration |
106
- | `best_model.pt` | PyTorch checkpoint (~136 M parameters) |
 
12
 
13
  ## 🧠 Training Configuration
14
 
15
+ This model was trained using the model architecture:
16
 
17
  ### ⚙️ Model Architecture
18
  | Parameter | Value | Description |
 
68
  | **Platform** | Google Colab Pro+ |
69
  | **GPU** | NVIDIA A100 (80 GB) |
70
  | **Runtime** | PyTorch 2.x |
71
+ | **Training time** | ~18-22 hours |
72
  | **Mixed precision** | Enabled (AMP) |
73
 
74
  ---
 
83
  ---
84
 
85
  ### 🧩 Summary
86
+ Scraps-LLM is a **138 M-parameter decoder-only Transformer** trained to generate complete cooking recipes from a list of input ingredients.
87
  The model learns via **causal language modeling (next-token prediction)** on RecipeNLG data, producing structured, human-readable recipes that include titles and numbered steps.
88
  It was later exported to **ONNX** for lightweight CPU inference and integrated into a Hugging Face Space demo.
89
 
 
102
  |------|-------------|
103
  | `export/scraps.onnx` | ONNX-optimized inference graph |
104
  | `tokenizer/bpe.json` | BPE vocabulary used for encoding/decoding |
105
+ | `best_model.pt` | PyTorch checkpoint (~138 M parameters) |