dineth554 commited on
Commit
2ddf9d9
·
verified ·
1 Parent(s): 40578d9

Upload config.json with huggingface_hub

Browse files
Files changed (1) hide show
  1. config.json +64 -0
config.json ADDED
@@ -0,0 +1,64 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model_type": "gpt2",
3
+ "architectures": [
4
+ "GPT2LMHeadModel"
5
+ ],
6
+ "vocab_size": 16000,
7
+ "d_model": 576,
8
+ "num_layers": 13,
9
+ "num_heads": 16,
10
+ "d_ff": 1280,
11
+ "max_seq_len": 1024,
12
+ "dropout": 0.1,
13
+ "pad_token_id": 0,
14
+ "eos_token_id": 1,
15
+ "unk_token_id": 2,
16
+ "torch_dtype": "float32",
17
+ "transformers_version": "4.36.0",
18
+ "task": "text-generation",
19
+ "pipeline_tag": "text-generation",
20
+ "library_name": "transformers",
21
+ "license": "apache-2.0",
22
+ "language": ["en", "code"],
23
+ "tags": [
24
+ "transformers",
25
+ "pytorch",
26
+ "safetensors",
27
+ "text-generation",
28
+ "code-generation",
29
+ "python",
30
+ "javascript",
31
+ "coding",
32
+ "programming",
33
+ "sagemaker",
34
+ "amazon-sagemaker",
35
+ "cpu",
36
+ "compact",
37
+ "efficient",
38
+ "nvdya-kit",
39
+ "death-legion",
40
+ "vllm",
41
+ "sglang",
42
+ "llama-cpp",
43
+ "ollama",
44
+ "lm-studio",
45
+ "year-2026",
46
+ "next-gen"
47
+ ],
48
+ "datasets": ["the-stack-v2"],
49
+ "metrics": ["perplexity", "accuracy"],
50
+ "inference": {
51
+ "parameters": {
52
+ "temperature": 0.8,
53
+ "top_p": 0.95,
54
+ "top_k": 50,
55
+ "max_new_tokens": 200
56
+ }
57
+ },
58
+ "sagemaker": {
59
+ "sdk_version": "2.200.0",
60
+ "instance_type": "ml.m5.large",
61
+ "instance_count": 1,
62
+ "container_image": "huggingface-pytorch-inference:2.0.0-transformers4.28.1-cpu-py310-ubuntu20.04-v1.0"
63
+ }
64
+ }