Haaaaarsh commited on
Commit
3bb9a71
·
verified ·
1 Parent(s): 5aca7df

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -74
README.md CHANGED
@@ -1,74 +0,0 @@
1
- ---
2
- library_name: transformers
3
- pipeline_tag: text-generation
4
- tags:
5
- - BGPT
6
- - meta
7
- - pytorch
8
- - llama
9
- - llama-3
10
- ---
11
-
12
-
13
- ### Model Description
14
-
15
- This model is a finetuned version of Llama3.2-3B-Instruct specifically designed for generating multilingual outputs across multiple Indic languages. The model has been trained on a diverse and curated dataset comprising Hindi, Punjabi, Marathi, Malayalam, Oriya, Kannada, Gujarati, Bengali, Urdu, Tamil, and Telugu. It is optimized to handle natural language tasks such as translation, summarization, and conversational generation across these languages effectively.
16
-
17
- - **Developed by:** [More Information Needed]
18
- - **Model type:** Finetuned LLaMA (Language Model for Multilingual Text Generation)
19
- - **Language(s) (NLP):** Hindi, Punjabi, Marathi, Malayalam, Oriya, Kannada, Gujarati, Bengali, Urdu, Tamil, Telugu
20
- - **Finetuned from model:** Llama3.2-3B-Instruct
21
-
22
- ## How to Get Started with the Model
23
-
24
- Make sure to update your transformers installation via `pip install --upgrade transformers`.
25
-
26
- Use the code below to get started with the model.
27
-
28
- ```python
29
- import torch
30
- from transformers import pipeline
31
-
32
- model_id = "Onkarn/ML-Test-v01"
33
- pipe = pipeline(
34
- "text-generation",
35
- model=model_id,
36
- torch_dtype=torch.bfloat16,
37
- device_map="auto",
38
- )
39
- messages = [
40
- {"role": "system", "content": "You are a helpful assistant who responds in hindi"},
41
- {"role": "user", "content": "कर्नाटक की राजधानी क्या है?"},
42
- ]
43
- outputs = pipe(
44
- messages,
45
- max_new_tokens=256,
46
- )
47
- print(outputs[0]["generated_text"][-1])
48
- ```
49
-
50
-
51
- ## Training Details
52
-
53
- ### Training Data
54
-
55
- The training dataset included a diverse collection of text sources in:
56
-
57
- - Hindi, Punjabi, Marathi, Malayalam, Oriya, Kannada, Gujarati, Bengali, Urdu, Tamil, and Telugu.
58
-
59
- ### Training Parameters
60
-
61
- - **Optimization Technique**: LoRA (Low-Rank Adaptation)
62
- - **Epochs**: 3.0
63
- - **Batch Size**: 2.0 (per device train batch size)
64
- - **Learning Rate**: 5e-05
65
-
66
-
67
- ## Environmental Impact
68
-
69
- - **Hardware Type:** T4
70
- - **Hours used:** 29 hours
71
- - **Cloud Provider:** Google Cloud Platform
72
- - **Compute Region:** asia-southeast1
73
- - **Carbon Emitted:** Total emissions are estimated to be 0.85 kgCO$_2$eq of which 100 percents were directly offset by the cloud provider.
74
-