hydffgg commited on
Commit
540ccb6
·
verified ·
1 Parent(s): 4566087

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +53 -1
README.md CHANGED
@@ -7,11 +7,63 @@ pipeline_tag: text-generation
7
  tags:
8
  - llama
9
  - causal-lm
 
10
  - 270m
 
11
  base_model:
12
  - Qwen/Qwen2.5-Coder-1.5B-Instruct
13
  ---
14
 
15
  # HOS-OSS-270M
16
 
17
- Custom 270M LLaMA-style model.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7
  tags:
8
  - llama
9
  - causal-lm
10
+ - code-generation
11
  - 270m
12
+ - lightweight
13
  base_model:
14
  - Qwen/Qwen2.5-Coder-1.5B-Instruct
15
  ---
16
 
17
  # HOS-OSS-270M
18
 
19
+ HOS-OSS-270M is a lightweight 270M parameter causal language model optimized for text and code generation tasks.
20
+ It is designed for fast inference, low resource usage, and local deployment.
21
+
22
+ ---
23
+
24
+ ## 🚀 Overview
25
+
26
+ - **Model size:** ~270M parameters
27
+ - **Architecture:** LLaMA-style decoder-only transformer
28
+ - **Base model:** Qwen2.5-Coder-1.5B-Instruct (distilled / adapted)
29
+ - **Framework:** 🤗 Transformers
30
+ - **Use cases:**
31
+ - Code generation
32
+ - Instruction following
33
+ - Chat-style completion
34
+ - Lightweight local AI assistant
35
+
36
+ ---
37
+
38
+ ## ⚡ Features
39
+
40
+ - Fast inference on low-end GPUs
41
+ - Runs on Kaggle / Colab without large VRAM
42
+ - Suitable for edge deployment
43
+ - Clean instruction-response formatting
44
+
45
+ ---
46
+
47
+ ## 🧠 Example Usage
48
+
49
+ ```python
50
+ from transformers import AutoTokenizer, AutoModelForCausalLM
51
+ import torch
52
+
53
+ model_name = "hydffgg/HOS-OSS-270M"
54
+
55
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
56
+ model = AutoModelForCausalLM.from_pretrained(model_name)
57
+
58
+ prompt = "User: Write a Python Hello World\nAssistant:"
59
+
60
+ inputs = tokenizer(prompt, return_tensors="pt")
61
+
62
+ with torch.no_grad():
63
+ outputs = model.generate(
64
+ **inputs,
65
+ max_new_tokens=100,
66
+ temperature=0.7
67
+ )
68
+
69
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))