Kylan12 commited on
Commit
b5fc478
·
verified ·
1 Parent(s): 2ef0496

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +28 -33
README.md CHANGED
@@ -3,35 +3,48 @@ language:
3
  - en
4
  license: apache-2.0
5
  tags:
 
6
  - qwen2.5
7
  - fine-tuned
8
  - lora
9
  - quantum-physics
 
10
  ---
11
 
12
  # qwen-25-14b-instruct-quantum-physics
13
 
14
- This model is a fine-tuned version of Qwen/Qwen2.5-14B-Instruct using LoRA (Low-Rank Adaptation) on a quantum physics dataset. This fine-tuned version scores 41.39% on a quantum physics test set, up from 24% on the base Qwen 2.5 14B Instruct model using standard Supervised Fine-Tuning (SFT)
15
- ## Model Description
16
 
17
- Fine-tuned Qwen2.5-14B model for quantum physics domain tasks.
 
 
 
 
 
 
 
 
 
18
 
19
  ## Available Formats
20
 
21
- * GGUF: `_temp_merged_qwen-25-14b-instruct-14b-quantum-physics-20260125-007.fp16.gguf` - FP16 format for inference with llama.cpp
 
22
 
23
  ## Usage
24
 
25
  ### Using GGUF (with llama.cpp, Ollama, LM Studio, etc.)
 
26
  ```bash
27
- # Download the GGUF file
28
- huggingface-cli download Kylan12/qwen-25-14b-instruct-quantum-physics _temp_merged_qwen-25-14b-instruct-14b-quantum-physics-20260125-007.fp16.gguf
29
 
30
  # Use with llama.cpp
31
- ./llama.cpp/build/bin/llama-cli -m _temp_merged_qwen-25-14b-instruct-14b-quantum-physics-20260125-007.fp16.gguf -p "Your prompt here"
32
  ```
33
 
34
  ### Using HuggingFace Transformers
 
35
  ```python
36
  from transformers import AutoModelForCausalLM, AutoTokenizer
37
 
@@ -44,37 +57,19 @@ outputs = model.generate(**inputs, max_length=200)
44
  print(tokenizer.decode(outputs[0]))
45
  ```
46
 
47
- ## Training Details standard Supervised Fine-Tuning (SFT)
48
 
49
- * **Base Model:** Qwen/Qwen2.5-14B-Instruct
50
- * **Training Method:** LoRA (Low-Rank Adaptation)
51
- * **LoRA Rank:** 16
52
- * **LoRA Alpha:** 16
53
- * **Target Modules:** q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj
54
-
55
- ## Evaluation
56
-
57
- | Metric | Base Model | Fine-Tuned |
58
- |--------|------------|------------|
59
- | Overall | 24.0% | 41.39% |
60
 
61
  ## Limitations
62
 
63
  This model inherits the limitations of the base Qwen2.5-14B-Instruct model and may have additional domain-specific limitations due to the fine-tuning dataset.
64
 
65
- ## Citation
66
-
67
- If you use this model, please cite:
68
- ```bibtex
69
- @misc{qwen_25_14b_instruct_quantum_physics,
70
- author = {Kylan12},
71
- title = {qwen-25-14b-instruct-quantum-physics},
72
- year = {2025},
73
- publisher = {HuggingFace},
74
- url = {https://huggingface.co/Kylan12/qwen-25-14b-instruct-quantum-physics}
75
- }
76
- ```
77
-
78
  ## License
79
 
80
- This model is released under the Apache 2.0 license, consistent with the base Qwen model.
 
3
  - en
4
  license: apache-2.0
5
  tags:
6
+ - gguf
7
  - qwen2.5
8
  - fine-tuned
9
  - lora
10
  - quantum-physics
11
+ base_model: Qwen/Qwen2.5-14B-Instruct
12
  ---
13
 
14
  # qwen-25-14b-instruct-quantum-physics
15
 
16
+ This model is a fine-tuned version of [Qwen/Qwen2.5-14B-Instruct](https://huggingface.co/Qwen/Qwen2.5-14B-Instruct) using LoRA (Low-Rank Adaptation) on a quantum physics dataset.
 
17
 
18
+ ## Evaluation
19
+
20
+ | Metric | Base Model | Fine-Tuned (SFT) | Fine-Tuned (latest) |
21
+ |--------|------------|-------------------|---------------------|
22
+ | Overall Accuracy | 24.0% | 41.4% | **53.7%** |
23
+ | Factual Accuracy | — | — | 55.0 |
24
+ | Completeness | — | — | 51.0 |
25
+ | Technical Precision | — | — | 54.3 |
26
+
27
+ Evaluated on [BoltzmannEntropy/QuantumLLMInstruct](https://huggingface.co/datasets/BoltzmannEntropy/QuantumLLMInstruct) with RAG-augmented judging (Semantic Scholar, 5 papers per question).
28
 
29
  ## Available Formats
30
 
31
+ - **GGUF (Q4_K_M)**: `qwen-25-14b-quantum-physics-q4_k_m.gguf` 8.4 GB, quantized for efficient inference
32
+ - **GGUF (FP16)**: `_temp_merged_qwen-25-14b-instruct-14b-quantum-physics-20260125-007.fp16.gguf` — full precision
33
 
34
  ## Usage
35
 
36
  ### Using GGUF (with llama.cpp, Ollama, LM Studio, etc.)
37
+
38
  ```bash
39
+ # Download the quantized GGUF
40
+ huggingface-cli download Kylan12/qwen-25-14b-instruct-quantum-physics qwen-25-14b-quantum-physics-q4_k_m.gguf
41
 
42
  # Use with llama.cpp
43
+ ./llama.cpp/build/bin/llama-cli -m qwen-25-14b-quantum-physics-q4_k_m.gguf -p "Your prompt here"
44
  ```
45
 
46
  ### Using HuggingFace Transformers
47
+
48
  ```python
49
  from transformers import AutoModelForCausalLM, AutoTokenizer
50
 
 
57
  print(tokenizer.decode(outputs[0]))
58
  ```
59
 
60
+ ## Training Details
61
 
62
+ - **Base Model:** Qwen/Qwen2.5-14B-Instruct
63
+ - **Training Method:** LoRA (Low-Rank Adaptation)
64
+ - **Quantization:** 4-bit NF4 via bitsandbytes
65
+ - **LoRA Rank:** 16
66
+ - **LoRA Alpha:** 16
67
+ - **Target Modules:** q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj
 
 
 
 
 
68
 
69
  ## Limitations
70
 
71
  This model inherits the limitations of the base Qwen2.5-14B-Instruct model and may have additional domain-specific limitations due to the fine-tuning dataset.
72
 
 
 
 
 
 
 
 
 
 
 
 
 
 
73
  ## License
74
 
75
+ This model is released under the Apache 2.0 license, consistent with the base Qwen model.