bitpoint commited on
Commit
c076c03
·
verified ·
1 Parent(s): 644b7ec

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +29 -0
README.md ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ # Model Card for vit-gpt2-image-captioning
3
+
4
+ ## Model Details
5
+ This model is a VisionEncoderDecoderModel using a ViT encoder and GPT-2 decoder to generate captions for images. It was fine-tuned by adding context information to assist in generating meaningful captions.
6
+
7
+ - **Base Model**: nlpconnect/vit-gpt2-image-captioning
8
+ - **Processor**: ViTImageProcessor
9
+ - **Tokenizer**: GPT-2 Tokenizer
10
+ - **Generated Caption Example**: "{generated_text}"
11
+
12
+ ## Intended Use
13
+ This model is intended for generating captions for stock-related images, with an initial context provided for more accurate descriptions.
14
+
15
+ ## Limitations
16
+ - The model might generate incorrect or biased descriptions depending on the input image or context.
17
+ - It requires specific context inputs for the best performance.
18
+
19
+ ## How to Use
20
+ ```python
21
+ from transformers import VisionEncoderDecoderModel, ViTImageProcessor, AutoTokenizer
22
+ model = VisionEncoderDecoderModel.from_pretrained("your_username/your_model_name")
23
+ processor = ViTImageProcessor.from_pretrained("your_username/your_model_name")
24
+ tokenizer = AutoTokenizer.from_pretrained("your_username/your_model_name")
25
+ ```
26
+
27
+ ## License
28
+ This model is licensed under the same terms as the original nlpconnect/vit-gpt2-image-captioning.
29
+