mtr0930 commited on
Commit
2e8eb39
ยท
1 Parent(s): 8ec7988

first commit

Browse files
added_tokens.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"์‚ฌ๊ต๊ฐ€/์€๊ทผ์—ฐ๊ตฌ์ž(4/1)": 35016, "์—๋„ˆ์ž์ด์ €": 35000, "๊ด€์กฐ์ž/์€๊ทผ๋„์ „๊ฐ€(6/3)": 35026, "๋‹ฌ์ฃผ๊ธฐ๊ฒฐ์ •๋ฐฉ์‹": 35035, "์€๋‘”์ž/์€๊ทผํ•ด๊ฒฐ์‚ฌ(2/5)": 35010, "๋„์ „๊ฐ€/์€๊ทผ๊ด€์กฐ์ž(3/6)": 35014, "๋ฐฉํ–ฅ์„ฑ๊ฒฐ์ •๋ฐฉ์‹": 35032, "์—๊ณ ํˆฌ์‚ฌ๊ฒฐ์ •๋ฐฉ์‹": 35033, "์ง๊ด€์„ผํ„ฐ": 35042, "ํ•ด๊ฒฐ์‚ฌ/์€๊ทผ์—ฐ๊ตฌ์ž(5/1)": 35020, "์—ฐ๊ตฌ์ž/์€๊ทผ์‚ฌ๊ต๊ฐ€": 35007, "(UNDEFINED)": 35051, "์„ธ๋ฌถ์Œํ๋ฆ„": 35038, "์‚ฌ๊ต๊ฐ€/์€๊ทผ๊ด€์กฐ์ž": 35019, "์ƒ๊ฐ์„ผํ„ฐ": 35045, "๊ฑฐ์šธ์ข…์กฑ": 35003, "๊ฐ์ •์„ผํ„ฐ": 35043, "์—ฐ๊ตฌ์ž/์€๊ทผ๋„์ „๊ฐ€": 35005, "์€๋‘”์ž/์€๊ทผ์‚ฌ๊ต๊ฐ€(2/4)": 35008, "ํ™œ๋ ฅ๊ฒฐ์ •๋ฐฉ์‹": 35028, "๋ฐฉํ–ฅ์„ผํ„ฐ": 35046, "์™ธ๋ถ€ํ™˜๊ฒฝ๊ฒฐ์ •๋ฐฉ์‹": 35034, "์—๊ณ ์„ผํ„ฐ": 35044, "์—๊ณ ํ˜„์‹œ๊ฒฐ์ •๋ฐฉ์‹": 35031, "์ ˆ์ „๋ชจ๋“œ": 35036, "๊ด€์กฐ์ž/์€๊ทผ์€๋‘”์ž(6/2)": 35024, "ํ™œ๋ ฅ์„ผํ„ฐ": 35041, "๊ด€์กฐ์ž/์€๊ทผ์€๋‘”์ž": 35025, "์‚ฌ๊ต๊ฐ€/์€๊ทผ๊ด€์กฐ์ž(4/6)": 35018, "ํ˜์‹ ์ฃผ๋„๊ฐ€": 35002, "๋„์ „๊ฐ€/์€๊ทผํ•ด๊ฒฐ์‚ฌ": 35013, "์—ฐ๊ตฌ์ž/์€๊ทผ๋„์ „๊ฐ€(1/3)": 35004, "์Šคํ”ผ๋“œ์—๋„ˆ์ž์ด์ €": 35001, "์ง๊ด€๊ฒฐ์ •๋ฐฉ์‹": 35030, "๋‘๋ฌถ์Œํ๋ฆ„": 35037, "์—ฐ๋ฃŒ์„ผํ„ฐ": 35040, "๋„์ „๊ฐ€/์€๊ทผํ•ด๊ฒฐ์‚ฌ(3/5)": 35012, "์€๋‘”์ž/์€๊ทผ์‚ฌ๊ต๊ฐ€": 35009, "ํ•ด๊ฒฐ์‚ฌ/์€๊ทผ์€๋‘”์ž": 35023, "๋„ค๋ฌถ์Œํ๋ฆ„": 35039, "ํ•ด๊ฒฐ์‚ฌ/์€๊ทผ์€๋‘”์ž(5/2)": 35022, "(DEFINED)": 35050, "์—ฐ๊ตฌ์ž/์€๊ทผ์‚ฌ๊ต๊ฐ€(1/4)": 35006, "ํ•ด๊ฒฐ์‚ฌ/์€๊ทผ์—ฐ๊ตฌ์ž": 35021, "ํ•œ๋ฌถ์Œํ๋ฆ„": 35049, "์‚ฌ๊ต๊ฐ€/์€๊ทผ์—ฐ๊ตฌ์ž": 35017, "ํ‘œํ˜„์„ผํ„ฐ": 35048, "๊ฐ์ •๊ฒฐ์ •๋ฐฉ์‹": 35029, "๊ด€์กฐ์ž/์€๊ทผ๋„์ „๊ฐ€": 35027, "์˜๊ฐ์„ผํ„ฐ": 35047, "๋„์ „๊ฐ€/์€๊ทผ๊ด€์กฐ์ž": 35015, "์€๋‘”์ž/์€๊ทผํ•ด๊ฒฐ์‚ฌ": 35011}
config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "mtr0930/koelectra-base-v3_epoch-100",
3
+ "architectures": [
4
+ "ElectraForQuestionAnswering"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "embedding_size": 768,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.1,
10
+ "hidden_size": 768,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 3072,
13
+ "layer_norm_eps": 1e-12,
14
+ "max_position_embeddings": 512,
15
+ "model_type": "electra",
16
+ "num_attention_heads": 12,
17
+ "num_hidden_layers": 12,
18
+ "pad_token_id": 0,
19
+ "position_embedding_type": "absolute",
20
+ "summary_activation": "gelu",
21
+ "summary_last_dropout": 0.1,
22
+ "summary_type": "first",
23
+ "summary_use_proj": true,
24
+ "torch_dtype": "float32",
25
+ "transformers_version": "4.9.2",
26
+ "type_vocab_size": 2,
27
+ "vocab_size": 35052
28
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3879d9f9e59c07d41d12b3da43661f6ca341cbb2616772b1e01f9aca75ce6430
3
+ size 449570737
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"do_lower_case": false, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "do_basic_tokenize": true, "never_split": null, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "mtr0930/koelectra-base-v3_epoch-100", "tokenizer_class": "ElectraTokenizer"}
vocab.txt ADDED
The diff for this file is too large to render. See raw diff