imirandam commited on
Commit
321a3c1
·
verified ·
1 Parent(s): 20e4bc4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +35 -3
README.md CHANGED
@@ -1,3 +1,35 @@
1
- ---
2
- license: mit
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ datasets:
4
+ - imirandam/TROHN-Img
5
+ ---
6
+
7
+ # Model Card for CLIP_Detectos
8
+ ## Model Description
9
+ - **Homepage:** https://imirandam.github.io/BiVLC_project_page/
10
+ - **Repository:** https://github.com/IMirandaM/BiVLC
11
+ - **Paper:**
12
+ - **Point of Contact:** [Imanol Miranda](mailto:imanol.miranda@ehu.eus)
13
+ ### Model Summary
14
+ CLIP_Detector is a model presented in the [BiVLC](https://github.com/IMirandaM/BiVLC) paper for experimentation. It has been trained with the OpenCLIP framework using the CLIP ViT-B-32 model pre-trained by 'openai' as a basis. The encoders are kept frozen, and a sigmoid neuron is added on top of each encoder (more details in the paper). The objective of the model is to classify text and images as natural or synthetic. Hyperparameters:
15
+
16
+ * Learning rate: 1e-6.
17
+ * Optimizer: Adam optimizer with beta1 = 0.9, beta2 = 0.999, eps = 1e-08 and without weight decay.
18
+ * Loss function: Binary cross-entropy loss (BCELoss).
19
+ * Batch size: We define a batch size of 400.
20
+ * Epochs: We trained the text detector over 10 epochs and the image detectors over 1 epoch. We used validation accuracy as the model selection criterion, i.e. we selected the model with highest accuracy in the corresponding validation set.
21
+ * Data: Then sigmoid neuron is trained with [TROHN-Img](https://huggingface.co/datasets/imirandam/TROHN-Img) dataset.
22
+
23
+ ### Licensing Information
24
+ This work is licensed under a MIT License.
25
+
26
+ ## Citation Information
27
+ If you find this dataset useful, please consider citing our paper:
28
+ ```
29
+ @inproceedings{,
30
+ title={},
31
+ author={},
32
+ booktitle={},
33
+ year={}
34
+ }
35
+ ```