BidhanAcharya commited on
Commit
5eaa15b
·
verified ·
1 Parent(s): 21fd444

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -4
README.md CHANGED
@@ -1,29 +1,34 @@
1
  ---
2
  base_model: unsloth/qwen2.5-coder-1.5b-bnb-4bit
3
  library_name: peft
 
4
  ---
5
 
6
  # Model Card for Model ID
7
 
8
- <!-- Provide a quick summary of what the model is/does. -->
9
 
10
 
11
 
12
  ## Model Details
13
 
 
14
  ### Model Description
15
 
16
- <!-- Provide a longer summary of what this model is. -->
 
 
 
17
 
18
 
19
 
20
- - **Developed by:** [More Information Needed]
21
  - **Funded by [optional]:** [More Information Needed]
22
  - **Shared by [optional]:** [More Information Needed]
23
  - **Model type:** [More Information Needed]
24
  - **Language(s) (NLP):** [More Information Needed]
25
  - **License:** [More Information Needed]
26
- - **Finetuned from model [optional]:** [More Information Needed]
27
 
28
  ### Model Sources [optional]
29
 
 
1
  ---
2
  base_model: unsloth/qwen2.5-coder-1.5b-bnb-4bit
3
  library_name: peft
4
+ license: mit
5
  ---
6
 
7
  # Model Card for Model ID
8
 
9
+ This model is a fine-tuned version of unsloth/qwen2.5-coder-1.5b-bnb-4bit, specifically adapted to solve coding problems using the CodeAlpaca-20k dataset. The model has been optimized for generating high-quality solutions to programming questions across various languages. It leverages the benefits of low-bit quantization for efficient inference while maintaining competitive performance.
10
 
11
 
12
 
13
  ## Model Details
14
 
15
+
16
  ### Model Description
17
 
18
+ Architecture: The model is based on QWen-2.5, a 1.5-billion parameter model optimized using 4-bit quantization via Bits and Bytes. This allows for reduced memory usage and faster inference while maintaining the model’s effectiveness.
19
+ Fine-tuning Process: The model was fine-tuned on the CodeAlpaca-20k dataset, a large corpus of coding-related prompts and solutions that span multiple programming languages. The goal of the fine-tuning was to improve the model’s ability to solve real-world coding problems and generate accurate, executable code.
20
+ Max Sequence Length: 2048 tokens to accommodate larger input sizes.
21
+ Quantization: The use of 4-bit quantization significantly reduces the memory footprint without sacrificing much on model performance, making it ideal for deployment in environments with limited resources
22
 
23
 
24
 
25
+ - **Developed by:** Bidhan Acharya
26
  - **Funded by [optional]:** [More Information Needed]
27
  - **Shared by [optional]:** [More Information Needed]
28
  - **Model type:** [More Information Needed]
29
  - **Language(s) (NLP):** [More Information Needed]
30
  - **License:** [More Information Needed]
31
+ - **Finetuned from model [optional]:** Qwen/Qwen2.5-Coder-1.5B
32
 
33
  ### Model Sources [optional]
34