V4.22.0 processor update

#2
by lewtun HF Staff - opened
merges.txt CHANGED
The diff for this file is too large to render. See raw diff
 
preprocessor_config.json CHANGED
@@ -1,6 +1,7 @@
1
  {
2
- "crop_size": 224,
3
  "do_center_crop": true,
 
4
  "do_normalize": true,
5
  "do_resize": true,
6
  "feature_extractor_type": "CLIPFeatureExtractor",
@@ -14,6 +15,7 @@
14
  0.26130258,
15
  0.27577711
16
  ],
 
17
  "resample": 3,
18
- "size": 224
19
  }
 
1
  {
2
+ "crop_size": 30,
3
  "do_center_crop": true,
4
+ "do_convert_rgb": true,
5
  "do_normalize": true,
6
  "do_resize": true,
7
  "feature_extractor_type": "CLIPFeatureExtractor",
 
15
  0.26130258,
16
  0.27577711
17
  ],
18
+ "processor_class": "CLIPProcessor",
19
  "resample": 3,
20
+ "size": 30
21
  }
special_tokens_map.json CHANGED
@@ -1 +1,24 @@
1
- {"bos_token": {"content": "<|startoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "pad_token": "<|endoftext|>"}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|startoftext|>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|endoftext|>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "<|endoftext|>",
17
+ "unk_token": {
18
+ "content": "<|endoftext|>",
19
+ "lstrip": false,
20
+ "normalized": true,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -1 +1,37 @@
1
- {"unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "<|startoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "pad_token": "<|endoftext|>", "add_prefix_space": false, "errors": "replace", "do_lower_case": true, "name_or_path": "openai/clip-vit-base-patch32", "model_max_length": 77, "special_tokens_map_file": "/home/lysandre/.cache/huggingface/transformers/18a566598f286c9139f88160c99f84eec492a26bd22738fa9cb44d5b7e0a5c76.cce1206abbad28826f000510f22f354e53e66a97f7c23745a7dfe27609cc07f5", "tokenizer_class": "CLIPTokenizer"}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "bos_token": {
4
+ "__type": "AddedToken",
5
+ "content": "<|startoftext|>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false
10
+ },
11
+ "crop_size": 30,
12
+ "do_lower_case": true,
13
+ "eos_token": {
14
+ "__type": "AddedToken",
15
+ "content": "<|endoftext|>",
16
+ "lstrip": false,
17
+ "normalized": true,
18
+ "rstrip": false,
19
+ "single_word": false
20
+ },
21
+ "errors": "replace",
22
+ "model_max_length": 77,
23
+ "name_or_path": "openai/clip-vit-base-patch32",
24
+ "pad_token": "<|endoftext|>",
25
+ "processor_class": "CLIPProcessor",
26
+ "size": 30,
27
+ "special_tokens_map_file": "/Users/lewtun/.cache/huggingface/hub/models--openai--clip-vit-base-patch32/snapshots/f4881ba48ee4d21b7ed5602603b9e3e92eb1b346/special_tokens_map.json",
28
+ "tokenizer_class": "CLIPTokenizer",
29
+ "unk_token": {
30
+ "__type": "AddedToken",
31
+ "content": "<|endoftext|>",
32
+ "lstrip": false,
33
+ "normalized": true,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ }
37
+ }
vocab.json CHANGED
The diff for this file is too large to render. See raw diff