Lin-K76 commited on
Commit
f7639e0
·
verified ·
1 Parent(s): 473e39f

Upload folder using huggingface_hub

Browse files
config.json ADDED
@@ -0,0 +1,67 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/home/linghao/.cache/huggingface/hub/models--mistralai--Mistral-Large-Instruct-2407/snapshots/5c9ce5b5f7a7ad62d03e8c66c719b66d586de26b",
3
+ "architectures": [
4
+ "MistralForCausalLM"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 1,
8
+ "compression_config": {
9
+ "config_groups": {
10
+ "group_0": {
11
+ "input_activations": {
12
+ "block_structure": null,
13
+ "dynamic": false,
14
+ "group_size": null,
15
+ "num_bits": 8,
16
+ "observer": "minmax",
17
+ "observer_kwargs": {},
18
+ "strategy": "tensor",
19
+ "symmetric": true,
20
+ "type": "float"
21
+ },
22
+ "output_activations": null,
23
+ "targets": [
24
+ "Linear"
25
+ ],
26
+ "weights": {
27
+ "block_structure": null,
28
+ "dynamic": false,
29
+ "group_size": null,
30
+ "num_bits": 8,
31
+ "observer": "minmax",
32
+ "observer_kwargs": {},
33
+ "strategy": "tensor",
34
+ "symmetric": true,
35
+ "type": "float"
36
+ }
37
+ }
38
+ },
39
+ "format": "naive-quantized",
40
+ "global_compression_ratio": 1.464555049372147,
41
+ "ignore": [
42
+ "lm_head"
43
+ ],
44
+ "kv_cache_scheme": null,
45
+ "quant_method": "compressed-tensors",
46
+ "quantization_status": "frozen"
47
+ },
48
+ "eos_token_id": 2,
49
+ "head_dim": 128,
50
+ "hidden_act": "silu",
51
+ "hidden_size": 12288,
52
+ "initializer_range": 0.02,
53
+ "intermediate_size": 28672,
54
+ "max_position_embeddings": 131072,
55
+ "model_type": "mistral",
56
+ "num_attention_heads": 96,
57
+ "num_hidden_layers": 88,
58
+ "num_key_value_heads": 8,
59
+ "rms_norm_eps": 1e-05,
60
+ "rope_theta": 1000000.0,
61
+ "sliding_window": null,
62
+ "tie_word_embeddings": false,
63
+ "torch_dtype": "float16",
64
+ "transformers_version": "4.44.0.dev0",
65
+ "use_cache": true,
66
+ "vocab_size": 32768
67
+ }
generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 2,
5
+ "transformers_version": "4.44.0.dev0"
6
+ }
model-00001-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e303a2be698ab106f40e07f98e11c8753f2e543e912858cd9dce737e072ba58
3
+ size 4957822828
model-00002-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ca172335578310dd6e303174eec8b33db844eff48eb267509521362a7c007bb
3
+ size 4831994920
model-00003-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c0572b3c57d8ea384352b42562b038e6b6224bd9731e39ffe1eb6ff098bcbe4e
3
+ size 4857209108
model-00004-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a22145c894c982a900f1282f48650ff8577a5fe9ed3aff02bf0fe6b78db10f7
3
+ size 4831995008
model-00005-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8371c5b0847cfa8f892967b7c51307563ed6c1a42c24ad5281dd59ba73a01b56
3
+ size 4857209188
model-00006-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd9c84907549aa46721b7c36597c0c2b11f53cdb91b1a2189479c58b0a721653
3
+ size 4831995008
model-00007-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4df398ef4c5467f2df0687ff5f51e745a1a01bda13e82258d238fb527d724c56
3
+ size 4857209188
model-00008-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac18afa459b35b4d1799661af61e52940d9043cdf979b08b29dca2528dcb3bf7
3
+ size 4831995008
model-00009-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2dc1a3b35a4322f9d3fe75be589cb7b894bf5ea8ca887ec50c8a91bb86d578ec
3
+ size 4857209188
model-00010-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e2e7ad1788eb47780e2adc9eaaaa957fcaa2108a2001c915c7278416ee8ff7f
3
+ size 4831995008
model-00011-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:267a55f5f585ef0e3a69cd9852760f6a57b8028a727c9238f6981955f20b75f5
3
+ size 4857209188
model-00012-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b372ebbc6d6cf10cb3617bd646ce6faad0cf0e7dc25b7886bd761277349a0440
3
+ size 4831995008
model-00013-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c666c80487a3527a4f44049fcb5a9f726b2bc517ad10fad725c0009f5f73b414
3
+ size 4857209188
model-00014-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c395cd4fd5aa7ad298045856b7a3e0babc399bd846970f9923d3908084f12145
3
+ size 4831995008
model-00015-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6658ddab9358d459df21e5e996e69152f89a1062ea5550021f63f709ad360888
3
+ size 4857209188
model-00016-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:206b58cdb0ba2dc059e6c5a0a26366bbceb71abfc99a9c8821e01240a49b0cf6
3
+ size 4831995008
model-00017-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d8076d50a0e807ede502fb4a63c9fd502cff5ff53045bb27beff2cf12af6c070
3
+ size 4857209188
model-00018-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca6e33972a6e540d275cf87f50a8997ceb0b856c06f857feb7fb41006313bf86
3
+ size 4831995008
model-00019-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:762ddca452f8ac42410588b0cac0e23b8ac60e65545cbcb325350b16e21913de
3
+ size 4857209188
model-00020-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6435d44d08e4261b009757e0ce6853ea01062766ab742d250f9e72dbad906f79
3
+ size 4831995008
model-00021-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f180828684a3e48e3aa4850bfeb4c9045f981f6c0bf1f39ae4b8525ac895034
3
+ size 4857209188
model-00022-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13cdae4888e1b3c1f7da031f7c71156556e360f702e922c49b5a10d8b3b40f69
3
+ size 4831995008
model-00023-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5968fc71ab265802dcb56cc1b93e91aa8816364d249f2a1e0d9543bc85044d5f
3
+ size 4857209188
model-00024-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:45c728ab9a1a67d16f0e2f1fbd3e263a4ec879b2c3d4dcc06ba29f507d5de67b
3
+ size 4831995008
model-00025-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be3a465c79f5e75e3d6a8c49e5798d858ac4a73b2f7f62b303ec9f1f7f2b24cf
3
+ size 4857209188
model-00026-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a698337aa393632e015938a35c6ef8187f071df2e791006f39ac6b681bdaae8
3
+ size 2189503292
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
recipe.yaml ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ quant_stage:
2
+ quant_modifiers:
3
+ QuantizationModifier:
4
+ ignore: [lm_head]
5
+ config_groups:
6
+ group_0:
7
+ weights: {num_bits: 8, type: float, strategy: tensor, dynamic: false, symmetric: true}
8
+ input_activations: {num_bits: 8, type: float, strategy: tensor, dynamic: false,
9
+ symmetric: true}
10
+ targets: [Linear]
special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "</s>",
17
+ "unk_token": {
18
+ "content": "<unk>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:59f95e28944c062244741268596badc900df86c7f5ded05088d2da22a7379e06
3
+ size 587583
tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff