File size: 1,544 Bytes
c7ebb32
0959f52
818bff5
0959f52
10bef9c
ae8437f
 
 
818bff5
 
 
 
 
ae8437f
0959f52
 
 
 
 
10bef9c
0959f52
 
 
 
ae8437f
 
 
 
818bff5
10bef9c
818bff5
 
 
 
 
0959f52
 
 
 
 
10bef9c
0959f52
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
from PIL import Image
import torch
import numpy as np
from transformers import VisionEncoderDecoderModel, ViTImageProcessor, AutoTokenizer

def preprocess_image(image, target_size=(224, 224)):
    # Resize the image to a fixed size
    image = image.resize(target_size, Image.LANCZOS)
    # Convert to numpy array and normalize
    image_array = np.array(image) / 255.0
    # Transpose to (channels, height, width) format
    image_array = np.transpose(image_array, (2, 0, 1))
    return image_array

def generate_caption(image):
    # Load pre-trained model and tokenizer
    model = VisionEncoderDecoderModel.from_pretrained("nlpconnect/vit-gpt2-image-captioning")
    feature_extractor = ViTImageProcessor.from_pretrained("nlpconnect/vit-gpt2-image-captioning")
    tokenizer = AutoTokenizer.from_pretrained("nlpconnect/vit-gpt2-image-captioning")

    device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
    model.to(device)

    # Prepare image
    if not isinstance(image, Image.Image):
        image = Image.open(image)
    
    image = image.convert('RGB')
    image_array = preprocess_image(image)
    
    # Create a batch with a single image
    batch = np.expand_dims(image_array, axis=0)
    
    # Convert to tensor
    pixel_values = torch.tensor(batch).float()
    pixel_values = pixel_values.to(device)

    # Generate caption
    output_ids = model.generate(pixel_values, max_length=16, num_return_sequences=1)
    preds = tokenizer.batch_decode(output_ids, skip_special_tokens=True)
    
    return preds[0]