Update README.md
Browse files
README.md
CHANGED
|
@@ -1,5 +1,92 @@
|
|
| 1 |
-
---
|
| 2 |
-
license: apache-2.0
|
| 3 |
-
tags:
|
| 4 |
-
- unsloth
|
| 5 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: apache-2.0
|
| 3 |
+
tags:
|
| 4 |
+
- unsloth
|
| 5 |
+
datasets:
|
| 6 |
+
- 0xZee/UniversalScienceKownledge-finetome-top-20k
|
| 7 |
+
base_model:
|
| 8 |
+
- unsloth/Llama-3.2-3B-Instruct
|
| 9 |
+
---
|
| 10 |
+
|
| 11 |
+
|
| 12 |
+
---
|
| 13 |
+
|
| 14 |
+
## model-name: `llama3-2-3b-finetuned-ScienceKnowledge`
|
| 15 |
+
## base_model: `unsloth/Llama-3.2-3B-Instruct`
|
| 16 |
+
|
| 17 |
+
license: apache-2.0
|
| 18 |
+
tags:
|
| 19 |
+
- unsloth
|
| 20 |
+
- science
|
| 21 |
+
- knowledge
|
| 22 |
+
- fine-tuned
|
| 23 |
+
- causal-lm
|
| 24 |
+
|
| 25 |
+
## datasets:
|
| 26 |
+
- `0xZee/UniversalScienceKownledge-finetome-top-20k`
|
| 27 |
+
|
| 28 |
+
language:
|
| 29 |
+
- en
|
| 30 |
+
library_name: transformers
|
| 31 |
+
inference: true
|
| 32 |
+
---
|
| 33 |
+
|
| 34 |
+
# llama3-2-3b-finetuned-ScienceKnowledge
|
| 35 |
+
|
| 36 |
+
This model is a fine-tuned version of unsloth/Llama-3.2-3B-Instruct, specifically adapted for enhanced performance in scientific knowledge tasks.
|
| 37 |
+
|
| 38 |
+
## Model Details
|
| 39 |
+
|
| 40 |
+
- **Base Model:** [unsloth/Llama-3.2-3B-Instruct](https://huggingface.co/unsloth/Llama-3.2-3B-Instruct)
|
| 41 |
+
- **Fine-tuning Dataset:** [0xZee/UniversalScienceKownledge-finetome-top-20k](https://huggingface.co/datasets/0xZee/UniversalScienceKownledge-finetome-top-20k)
|
| 42 |
+
- **Model Type:** Causal Language Model
|
| 43 |
+
- **License:** Apache 2.0
|
| 44 |
+
- **Finetuning Technique:** The model was likely fine-tuned using techniques such as LoRA or QLoRA, which are efficient for smaller datasets.
|
| 45 |
+
|
| 46 |
+
## Intended Use
|
| 47 |
+
|
| 48 |
+
This model is designed for tasks requiring scientific understanding, question answering, and knowledge retrieval in scientific contexts. It can be used for:
|
| 49 |
+
|
| 50 |
+
- Assisting researchers with scientific queries
|
| 51 |
+
- Educational tools in science domains
|
| 52 |
+
- Automated scientific knowledge discovery and summarization
|
| 53 |
+
|
| 54 |
+
## Training Procedure
|
| 55 |
+
|
| 56 |
+
The model was fine-tuned on the UniversalScienceKownledge-finetome-top-20k dataset. Specific hyperparameters used in training are not provided, but typical fine-tuning approaches for LLMs include:
|
| 57 |
+
|
| 58 |
+
- Learning rate: Often around 2e-5 for Mistral-based models
|
| 59 |
+
- LoRA rank: Likely between 512 and 768
|
| 60 |
+
- Batch size: Small, possibly between 1 and 4
|
| 61 |
+
|
| 62 |
+
## Limitations
|
| 63 |
+
|
| 64 |
+
- Performance is dependent on the quality and scope of the training data.
|
| 65 |
+
- May not generalize well to scientific domains outside of its training data.
|
| 66 |
+
- Potential for generating plausible-sounding but incorrect information (hallucinations).
|
| 67 |
+
|
| 68 |
+
## Bias, Risks, and Limitations
|
| 69 |
+
|
| 70 |
+
As with all language models, this model may reflect biases present in its training data. Users should be aware of potential biases in scientific literature and exercise caution when using the model for sensitive applications.
|
| 71 |
+
|
| 72 |
+
## Evaluation Results
|
| 73 |
+
|
| 74 |
+
Quantitative evaluation results are not provided. Users are encouraged to evaluate the model's performance on their specific tasks and domains of interest.
|
| 75 |
+
|
| 76 |
+
## How to Use
|
| 77 |
+
|
| 78 |
+
You can use this model with the Transformers library:
|
| 79 |
+
|
| 80 |
+
```py
|
| 81 |
+
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 82 |
+
|
| 83 |
+
model_name = "0xZee/llama3-2-3b-finetuned-ScienceKnowledge"
|
| 84 |
+
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
| 85 |
+
model = AutoModelForCausalLM.from_pretrained(model_name)
|
| 86 |
+
|
| 87 |
+
prompt = "Explain the concept of quantum entanglement."
|
| 88 |
+
input_ids = tokenizer(prompt, return_tensors="pt").input_ids
|
| 89 |
+
output = model.generate(input_ids, max_new_tokens=100)
|
| 90 |
+
print(tokenizer.decode(output, skip_special_tokens=True))
|
| 91 |
+
```
|
| 92 |
+
|