AliMuhammad73 commited on
Commit
dbee900
·
verified ·
1 Parent(s): 4912248

Upload model and tokenizer

Browse files
Files changed (5) hide show
  1. README.md +9 -32
  2. config.json +3 -0
  3. model.safetensors +3 -0
  4. tokenizer.model +3 -0
  5. tokenizer_config.json +4 -0
README.md CHANGED
@@ -1,32 +1,9 @@
1
- # Custom Urdu LLM
2
-
3
- This is a custom transformer-based Large Language Model for Urdu.
4
-
5
- ## Model Details
6
- - **Architecture:** Transformer (GPT-based)
7
- - **Framework:** PyTorch
8
- - **Tokenizer:** SentencePiece
9
- - **Hyperparameters:**
10
- - Vocabulary Size: 20,000
11
- - Embedding Size: 768
12
- - Attention Heads: 12
13
- - Layers: 12
14
- - Dropout: 0.2
15
-
16
- ## Usage
17
-
18
- ```python
19
- from transformers import AutoModel, AutoTokenizer
20
-
21
- model = AutoModel.from_pretrained("AliMuhammad73/testing-model")
22
- tokenizer = AutoTokenizer.from_pretrained("AliMuhammad73/testing-model")
23
-
24
- prompt = <prompt in urdu>
25
- inputs = tokenizer(prompt, return_tensors="pt")
26
- output = model.generate(inputs.input_ids, max_new_tokens=tokens_to_generate)
27
- print(tokenizer.decode(output[0]))
28
- ```
29
-
30
- ---
31
- license: apache-2.0
32
- ---
 
1
+ ---
2
+ tags:
3
+ - model_hub_mixin
4
+ - pytorch_model_hub_mixin
5
+ ---
6
+
7
+ This model has been pushed to the Hub using the [PytorchModelHubMixin](https://huggingface.co/docs/huggingface_hub/package_reference/mixins#huggingface_hub.PyTorchModelHubMixin) integration:
8
+ - Library: [More Information Needed]
9
+ - Docs: [More Information Needed]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
config.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "vocab_size": 20000
3
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:66ee862a2d28074541cad0ee5b4ec9d9aa98b8ab1efcd4782e1e5bef64c7adae
3
+ size 404903720
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:45858257f60a18fca1e9c91f723a2ded05ebc8e7de3e8ba64af819b095728c50
3
+ size 395464
tokenizer_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "type": "llama",
3
+ "vocab_size": 0
4
+ }