0xZee commited on
Commit
fc53ff2
·
verified ·
1 Parent(s): 6996a2b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +92 -5
README.md CHANGED
@@ -1,5 +1,92 @@
1
- ---
2
- license: apache-2.0
3
- tags:
4
- - unsloth
5
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - unsloth
5
+ datasets:
6
+ - 0xZee/UniversalScienceKownledge-finetome-top-20k
7
+ base_model:
8
+ - unsloth/Llama-3.2-3B-Instruct
9
+ ---
10
+
11
+
12
+ ---
13
+
14
+ ## model-name: `llama3-2-3b-finetuned-ScienceKnowledge`
15
+ ## base_model: `unsloth/Llama-3.2-3B-Instruct`
16
+
17
+ license: apache-2.0
18
+ tags:
19
+ - unsloth
20
+ - science
21
+ - knowledge
22
+ - fine-tuned
23
+ - causal-lm
24
+
25
+ ## datasets:
26
+ - `0xZee/UniversalScienceKownledge-finetome-top-20k`
27
+
28
+ language:
29
+ - en
30
+ library_name: transformers
31
+ inference: true
32
+ ---
33
+
34
+ # llama3-2-3b-finetuned-ScienceKnowledge
35
+
36
+ This model is a fine-tuned version of unsloth/Llama-3.2-3B-Instruct, specifically adapted for enhanced performance in scientific knowledge tasks.
37
+
38
+ ## Model Details
39
+
40
+ - **Base Model:** [unsloth/Llama-3.2-3B-Instruct](https://huggingface.co/unsloth/Llama-3.2-3B-Instruct)
41
+ - **Fine-tuning Dataset:** [0xZee/UniversalScienceKownledge-finetome-top-20k](https://huggingface.co/datasets/0xZee/UniversalScienceKownledge-finetome-top-20k)
42
+ - **Model Type:** Causal Language Model
43
+ - **License:** Apache 2.0
44
+ - **Finetuning Technique:** The model was likely fine-tuned using techniques such as LoRA or QLoRA, which are efficient for smaller datasets.
45
+
46
+ ## Intended Use
47
+
48
+ This model is designed for tasks requiring scientific understanding, question answering, and knowledge retrieval in scientific contexts. It can be used for:
49
+
50
+ - Assisting researchers with scientific queries
51
+ - Educational tools in science domains
52
+ - Automated scientific knowledge discovery and summarization
53
+
54
+ ## Training Procedure
55
+
56
+ The model was fine-tuned on the UniversalScienceKownledge-finetome-top-20k dataset. Specific hyperparameters used in training are not provided, but typical fine-tuning approaches for LLMs include:
57
+
58
+ - Learning rate: Often around 2e-5 for Mistral-based models
59
+ - LoRA rank: Likely between 512 and 768
60
+ - Batch size: Small, possibly between 1 and 4
61
+
62
+ ## Limitations
63
+
64
+ - Performance is dependent on the quality and scope of the training data.
65
+ - May not generalize well to scientific domains outside of its training data.
66
+ - Potential for generating plausible-sounding but incorrect information (hallucinations).
67
+
68
+ ## Bias, Risks, and Limitations
69
+
70
+ As with all language models, this model may reflect biases present in its training data. Users should be aware of potential biases in scientific literature and exercise caution when using the model for sensitive applications.
71
+
72
+ ## Evaluation Results
73
+
74
+ Quantitative evaluation results are not provided. Users are encouraged to evaluate the model's performance on their specific tasks and domains of interest.
75
+
76
+ ## How to Use
77
+
78
+ You can use this model with the Transformers library:
79
+
80
+ ```py
81
+ from transformers import AutoModelForCausalLM, AutoTokenizer
82
+
83
+ model_name = "0xZee/llama3-2-3b-finetuned-ScienceKnowledge"
84
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
85
+ model = AutoModelForCausalLM.from_pretrained(model_name)
86
+
87
+ prompt = "Explain the concept of quantum entanglement."
88
+ input_ids = tokenizer(prompt, return_tensors="pt").input_ids
89
+ output = model.generate(input_ids, max_new_tokens=100)
90
+ print(tokenizer.decode(output, skip_special_tokens=True))
91
+ ```
92
+