Joshua Sun commited on
Update README.md
Browse files
README.md
CHANGED
|
@@ -1,8 +1,50 @@
|
|
| 1 |
---
|
| 2 |
license: mit
|
| 3 |
language:
|
| 4 |
-
- en
|
| 5 |
base_model:
|
| 6 |
-
- mistralai/Mistral-7B-v0.1
|
| 7 |
pipeline_tag: text-generation
|
| 8 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
---
|
| 2 |
license: mit
|
| 3 |
language:
|
| 4 |
+
- en
|
| 5 |
base_model:
|
| 6 |
+
- mistralai/Mistral-7B-v0.1
|
| 7 |
pipeline_tag: text-generation
|
| 8 |
+
tags:
|
| 9 |
+
- academic-writing
|
| 10 |
+
- mistral
|
| 11 |
+
- qlora
|
| 12 |
+
- fine-tuning
|
| 13 |
+
- arxiv
|
| 14 |
+
- llm
|
| 15 |
+
- uc-davis
|
| 16 |
+
---
|
| 17 |
+
|
| 18 |
+
# Mistral-7B Fine-Tuned for Academic Style (QLoRA)
|
| 19 |
+
|
| 20 |
+
This is a parameter-efficient fine-tuning of `mistralai/Mistral-7B-v0.1` using QLoRA on 500K academic abstracts. It was built for the ECS 271 final project at UC Davis.
|
| 21 |
+
|
| 22 |
+
## Intended Use
|
| 23 |
+
|
| 24 |
+
The model is designed to generate formal academic paragraphs given a paper title, useful for research drafts, educational AI tools, and academic-style assistants.
|
| 25 |
+
|
| 26 |
+
## Training Details
|
| 27 |
+
|
| 28 |
+
- Base model: `mistralai/Mistral-7B-v0.1`
|
| 29 |
+
- Method: QLoRA (low-rank adapter)
|
| 30 |
+
- Prompt format: "Write an academic paragraph given the title: ..."
|
| 31 |
+
- Dataset: 500K arXiv abstracts
|
| 32 |
+
- Epochs: 1
|
| 33 |
+
- GPU: RTX 5070 Ti (~60 hours)
|
| 34 |
+
|
| 35 |
+
## Limitations
|
| 36 |
+
|
| 37 |
+
- Generic or templated outputs
|
| 38 |
+
- No citation support
|
| 39 |
+
- Frequent hallucinations
|
| 40 |
+
|
| 41 |
+
## Example
|
| 42 |
+
|
| 43 |
+
**Prompt:**
|
| 44 |
+
"Write an academic paragraph given the title: LoRA for In-Context Learning"
|
| 45 |
+
|
| 46 |
+
**Output:**
|
| 47 |
+
"We present LoRA (Loosely Regularized Adapters), a novel approach to fine-tune large language models for in-context learning tasks. Unlike traditional methods, LoRA updates only a small number of trainable parameters, achieving comparable performance while reducing training costs."
|
| 48 |
+
|
| 49 |
+
## More Info
|
| 50 |
+
Main Project: [Github](https://github.com/Joshua-Sun-CompSci/academic-style-llms)
|