t0m-R commited on
Commit
f921783
·
1 Parent(s): 18cc612

Switch model file to safetensors format

Browse files
.gitignore ADDED
@@ -0,0 +1 @@
 
 
1
+ backup/
README.md CHANGED
@@ -9,13 +9,13 @@ tags:
9
  - materials-science
10
  - nffa-di
11
  base_model:
12
- - google/vit-base-patch16-224-in21k
13
  pipeline_tag: image-classification
14
  ---
15
 
16
  # Vision Transformer for STM Multi-Tip Artifact Detection
17
 
18
- This is a fine-tuned **Vision Transformer (ViT-B/16)** model for classifying Scanning Tunneling Microscopy (STM) images. It is designed to detect the presence of **multi-tip artifacts**, a common distortion that results in duplicated signals and complicates data interpretation.
19
 
20
  This model was developed as part of the **NFFA-DI (Nano Foundries and Fine Analysis Digital Infrastructure)** project, funded by the European Union's NextGenerationEU program.
21
 
@@ -23,7 +23,7 @@ This model was developed as part of the **NFFA-DI (Nano Foundries and Fine Analy
23
 
24
  ## Model Description
25
 
26
- The model is a `ViT-B/16` pre-trained on ImageNet-21k. It was fine-tuned to classify an STM image as either `Artifact-Free` or `Multi-Tip Artifact`.
27
 
28
  A key feature of this model is its use of a **Fast Fourier Transform (FFT)** based preprocessing method. The model's input is not a standard image but a 3-channel tensor composed of:
29
  1. The grayscale STM image.
 
9
  - materials-science
10
  - nffa-di
11
  base_model:
12
+ - google/vit-base-patch32-224-in21k
13
  pipeline_tag: image-classification
14
  ---
15
 
16
  # Vision Transformer for STM Multi-Tip Artifact Detection
17
 
18
+ This is a fine-tuned **Vision Transformer (ViT-B/32)** model for classifying Scanning Tunneling Microscopy (STM) images. It is designed to detect the presence of **multi-tip artifacts**, a common distortion that results in duplicated signals and complicates data interpretation.
19
 
20
  This model was developed as part of the **NFFA-DI (Nano Foundries and Fine Analysis Digital Infrastructure)** project, funded by the European Union's NextGenerationEU program.
21
 
 
23
 
24
  ## Model Description
25
 
26
+ The model is a `ViT-B/32` pre-trained on ImageNet-21k. It was fine-tuned to classify an STM image as either `Artifact-Free` or `Multi-Tip Artifact`.
27
 
28
  A key feature of this model is its use of a **Fast Fourier Transform (FFT)** based preprocessing method. The model's input is not a standard image but a 3-channel tensor composed of:
29
  1. The grayscale STM image.
config.json CHANGED
@@ -1,16 +1,31 @@
1
  {
2
- "_name_or_path": "google/vit-base-patch16-224-in21k",
3
  "architectures": [
4
  "ViTForImageClassification"
5
  ],
6
- "model_type": "vit",
7
- "num_labels": 2,
 
 
 
8
  "id2label": {
9
  "0": "Artifact-Free",
10
  "1": "Multi-Tip Artifact"
11
  },
 
 
 
12
  "label2id": {
13
  "Artifact-Free": 0,
14
  "Multi-Tip Artifact": 1
15
- }
16
- }
 
 
 
 
 
 
 
 
 
 
1
  {
2
+ "_name_or_path": "google/vit-base-patch32-224-in21k",
3
  "architectures": [
4
  "ViTForImageClassification"
5
  ],
6
+ "attention_probs_dropout_prob": 0.0,
7
+ "encoder_stride": 16,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.0,
10
+ "hidden_size": 768,
11
  "id2label": {
12
  "0": "Artifact-Free",
13
  "1": "Multi-Tip Artifact"
14
  },
15
+ "image_size": 224,
16
+ "initializer_range": 0.02,
17
+ "intermediate_size": 3072,
18
  "label2id": {
19
  "Artifact-Free": 0,
20
  "Multi-Tip Artifact": 1
21
+ },
22
+ "layer_norm_eps": 1e-12,
23
+ "model_type": "vit",
24
+ "num_attention_heads": 12,
25
+ "num_channels": 3,
26
+ "num_hidden_layers": 12,
27
+ "patch_size": 32,
28
+ "qkv_bias": true,
29
+ "torch_dtype": "float32",
30
+ "transformers_version": "4.41.2"
31
+ }
pytorch_model.bin → model.safetensors RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4d3aaaf677542934b42ab898915c555d07337b4a904bd533eb6f50720a92f8d3
3
- size 343264618
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b314f59f218e478a5f78cca820c95f0848da4f74ae7f55ffd1dbeabdbd3e5a5
3
+ size 349850288