Irfanuruchi commited on
Commit
1bba02f
·
verified ·
1 Parent(s): 2ad0579

Adding README file

Browse files
Files changed (1) hide show
  1. README.md +122 -3
README.md CHANGED
@@ -1,3 +1,122 @@
1
- ---
2
- license: llama3.2
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - computer-engineering
4
+ - llama-3
5
+ - 3b
6
+ - lora
7
+ - 4bit
8
+ license: llama3.2
9
+ license_link: https://llama.meta.com/llama3/license
10
+ base_model:
11
+ - meta-llama/Llama-3.2-3B-Instruct
12
+ datasets:
13
+ - Wikitext-2-raw-v1
14
+ - STEM-AI-mtl
15
+ - custom-computer-engineering-corpus
16
+ - technical-documentation
17
+ - hardware-specs
18
+ ---
19
+
20
+ # 🖥️ Llama-3.2-3B-Computer-Engineering-LLM
21
+
22
+ **Specialized AI Assistant for Computer Engineering**
23
+ *Fine-tuned Meta-Llama-3-8B with 4-bit quantization + LoRA adapters*
24
+
25
+ <div align="center">
26
+ <a href="https://github.com/IrfanUruchi/Llama-3.2-3B-Computer-Engineering-LLM">
27
+ <img src="https://img.shields.io/badge/🔗_GitHub-Repo-181717?style=for-the-badge&logo=github" alt="GitHub">
28
+ </a>
29
+ <a href="https://huggingface.co/Irfanuruchi/Llama-3.2-3B-Computer-Engineering-LLM">
30
+ <img src="https://img.shields.io/badge/🤗_HuggingFace-Model_Repo-FFD21F?style=for-the-badge" alt="HuggingFace">
31
+ </a>
32
+ <br>
33
+ <img src="https://img.shields.io/badge/Model_Size-3.2B_parameters-blue" alt="Model Size">
34
+ <img src="https://img.shields.io/badge/Quantization-4bit-green" alt="Quantization">
35
+ <img src="https://img.shields.io/badge/Adapter-LoRA-orange" alt="Adapter">
36
+ <img src="https://img.shields.io/badge/Context-8k-lightgrey" alt="Context">
37
+ <img src="https://img.shields.io/badge/License-Llama_3.2-yellow" alt="License">
38
+ </div>
39
+
40
+
41
+ ## 📜 License Compliance Notice
42
+ This model is derived from Meta's Llama 3.2 and is governed by the [Llama 3.2 Community License](https://llama.meta.com/llama3/license). By using this model, you agree to:
43
+
44
+ - Not use the model or its outputs to improve other LLMs
45
+ - Not use the model for commercial purposes without separate agreement
46
+ - Include attribution to Meta and this project
47
+ - Accept the license's acceptable use policy
48
+
49
+ ---
50
+
51
+ ## 🛠️ Technical Specifications
52
+
53
+ ### Architecture
54
+ | Component | Implementation Details |
55
+ |------------------------|---------------------------------|
56
+ | Base Model | Meta-Llama-3-8B-Instruct |
57
+ | Quantization | 4-bit via BitsAndBytes |
58
+ | Adapter | LoRA (r=16, alpha=32) |
59
+ | Training Framework | PyTorch + HuggingFace Ecosystem|
60
+ | Context Window | 8,192 tokens |
61
+
62
+ ### Training Data
63
+ - Curated computer engineering corpus
64
+ - Key domains covered:
65
+ - Computer architecture
66
+ - Embedded systems
67
+ - VLSI design
68
+ - Hardware description languages
69
+ - Low-level programming
70
+
71
+ ---
72
+
73
+ ```python
74
+
75
+ from transformers import AutoTokenizer, AutoModelForCausalLM
76
+
77
+ model_id = "Irfanuruchi/Llama-3.2-3B-Computer-Engineering-LLM"
78
+ tokenizer = AutoTokenizer.from_pretrained(model_id)
79
+ model = AutoModelForCausalLM.from_pretrained(
80
+ model_id,
81
+ device_map="auto",
82
+ torch_dtype="auto"
83
+ )
84
+
85
+ prompt = """You are a computer engineering expert. Explain concisely:
86
+ Q: What's the difference between RISC and CISC architectures?
87
+ A:"""
88
+
89
+ inputs = tokenizer(prompt, return_tensors="pt").to("cuda")
90
+ outputs = model.generate(
91
+ **inputs,
92
+ max_new_tokens=150,
93
+ temperature=0.7,
94
+ do_sample=True
95
+ )
96
+
97
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
98
+
99
+ ```
100
+
101
+ ## Responsible use
102
+
103
+ This model inherits all use restrictions from the Llama 3.2 license. Special considerations:
104
+
105
+ Not for production deployment without compliance review
106
+ Outputs should be verified by domain experts
107
+ Knowledge cutoff: July 2024
108
+
109
+ ## Citation
110
+
111
+ If using this model in research, please cite:
112
+
113
+ ```bibtex
114
+ @misc{llama3.2-computer-eng,
115
+ author = {Irfanuruchi},
116
+ title = {Llama-3.2-3B-Computer-Engineering-LLM},
117
+ year = {2025},
118
+ publisher = {HuggingFace},
119
+ howpublished = {\url{https://huggingface.co/Irfanuruchi/Llama-3.2-3B-Computer-Engineering-LLM}}
120
+ }
121
+ ```
122
+