ZTWHHH's picture
Add files using upload-large-folder tool
56e4d17 verified
import argparse
import json
from math import ceil
import os
import random
import uuid
from collections import defaultdict
from typing import Callable
import more_itertools
import numpy as np
import torch
from coco_metric import compute_cider, postprocess_captioning_generation
from eval_datasets import COCOFlickrDataset, VQADataset, ImageNetDataset
from tqdm import tqdm
from open_flamingo.eval.ok_vqa_utils import postprocess_ok_vqa_generation
from vqa_metric import compute_vqa_accuracy, postprocess_vqa_generation
from open_flamingo.eval.classification import (
compute_per_sample_probs,
compute_per_sample_loss,
)
from open_flamingo.eval.imagenet_utils import (
openai_imagenet_classnames,
IMAGENET_1K_CLASS_ID_TO_LABEL,
)
from open_flamingo.src.factory import create_model_and_transforms
parser = argparse.ArgumentParser()
parser.add_argument("--lm_path", type=str, default="facebook/opt-1.3b")
parser.add_argument("--lm_tokenizer_path", type=str, default="facebook/opt-30b")
parser.add_argument("--vision_encoder_path", default="ViT-L-14", type=str)
parser.add_argument("--vision_encoder_pretrained", default="openai", type=str)
parser.add_argument("--checkpoint_path", type=str, required=True)
parser.add_argument(
"--cross_attn_every_n_layers",
type=int,
default=1,
help="how often to add a cross-attention layer after each transformer layer",
)
parser.add_argument(
"--results_file", type=str, default=None, help="JSON file to save results"
)
# Trial arguments
parser.add_argument("--shots", nargs="+", default=[0, 4, 8, 16, 32], type=int)
parser.add_argument(
"--num_trials",
type=int,
default=1,
help="Number of trials to run for each shot using different demonstrations",
)
parser.add_argument(
"--trial_seeds",
nargs="+",
default=[0],
help="Seeds to use for each trial for picking demonstrations and eval sets",
)
parser.add_argument(
"--num_samples", type=int, default=5000, help="Number of samples to evaluate on"
)
parser.add_argument("--batch_size", type=int, default=8)
parser.add_argument("--device", type=int, default=0)
# Per-dataset evaluation flags
parser.add_argument(
"--eval_coco",
action="store_true",
default=False,
help="Whether to evaluate on COCO.",
)
parser.add_argument(
"--eval_vqav2",
action="store_true",
default=False,
help="Whether to evaluate on VQAV2.",
)
parser.add_argument(
"--eval_ok_vqa",
action="store_true",
default=False,
help="Whether to evaluate on OK-VQA.",
)
parser.add_argument(
"--eval_imagenet",
action="store_true",
default=False,
help="Whether to evaluate on ImageNet.",
)
parser.add_argument(
"--eval_flickr30",
action="store_true",
default=False,
help="Whether to evaluate on Flickr30.",
)
# Dataset arguments
## Flickr30 Dataset
parser.add_argument(
"--flickr_image_dir_path",
type=str,
help="Path to the flickr30/flickr30k_images directory.",
default=None,
)
parser.add_argument(
"--flickr_annotations_json_path",
type=str,
help="Path to the dataset_flickr30k_coco_style.json file.",
default=None,
)
## COCO Dataset
parser.add_argument(
"--coco_image_dir_path",
type=str,
help="Path to the flickr30/flickr30k_images directory.",
default=None,
)
parser.add_argument(
"--coco_annotations_json_path",
type=str,
default=None,
)
## VQAV2 Dataset
parser.add_argument(
"--vqav2_image_dir_path",
type=str,
default=None,
)
parser.add_argument(
"--vqav2_questions_json_path",
type=str,
default=None,
)
parser.add_argument(
"--vqav2_annotations_json_path",
type=str,
default=None,
)
## OK-VQA Dataset
parser.add_argument(
"--ok_vqa_image_dir_path",
type=str,
help="Path to the vqav2/train2014 directory.",
default=None,
)
parser.add_argument(
"--ok_vqa_questions_json_path",
type=str,
help="Path to the v2_OpenEnded_mscoco_train2014_questions.json file.",
default=None,
)
parser.add_argument(
"--ok_vqa_annotations_json_path",
type=str,
help="Path to the v2_mscoco_train2014_annotations.json file.",
default=None,
)
## Imagenet dataset
parser.add_argument("--imagenet_root", type=str, default="/tmp")
def main():
args = parser.parse_args()
# load model
flamingo, image_processor, tokenizer = create_model_and_transforms(
args.vision_encoder_path,
args.vision_encoder_pretrained,
args.lm_path,
args.lm_tokenizer_path,
cross_attn_every_n_layers=args.cross_attn_every_n_layers,
)
checkpoint = torch.load(args.checkpoint_path, map_location="cpu")
flamingo.load_state_dict(checkpoint, strict=False)
flamingo.to(args.device if args.device >= 0 else "cpu")
results = defaultdict(list)
if args.eval_flickr30:
print("Evaluating on Flickr30...")
for shot in args.shots:
scores = []
for seed, trial in zip(args.trial_seeds, range(args.num_trials)):
cider_score = evaluate_coco_flickr(
model=flamingo,
tokenizer=tokenizer,
image_processor=image_processor,
batch_size=args.batch_size,
image_dir_path=args.flickr_image_dir_path,
annotations_json_path=args.flickr_annotations_json_path,
num_samples=args.num_samples,
num_shots=shot,
device=args.device,
seed=seed,
is_flickr=True,
)
print(f"Shots {shot} Trial {trial} CIDEr score: {cider_score}")
scores.append(cider_score)
print(f"Shots {shot} Mean CIDEr score: {np.mean(scores)}")
results["flickr30"].append(
{"shots": shot, "trials": scores, "mean": np.mean(scores)}
)
results = defaultdict(list)
if args.eval_coco:
print("Evaluating on COCO...")
for shot in args.shots:
scores = []
for seed, trial in zip(args.trial_seeds, range(args.num_trials)):
cider_score = evaluate_coco_flickr(
model=flamingo,
tokenizer=tokenizer,
image_processor=image_processor,
batch_size=args.batch_size,
image_dir_path=args.coco_image_dir_path,
annotations_json_path=args.coco_annotations_json_path,
num_samples=args.num_samples,
num_shots=shot,
device=args.device,
seed=seed,
)
print(f"Shots {shot} Trial {trial} CIDEr score: {cider_score}")
scores.append(cider_score)
print(f"Shots {shot} Mean CIDEr score: {np.mean(scores)}")
results["coco"].append(
{"shots": shot, "trials": scores, "mean": np.mean(scores)}
)
if args.eval_ok_vqa:
print("Evaluating on OK-VQA...")
for shot in args.shots:
scores = []
for seed, trial in zip(args.trial_seeds, range(args.num_trials)):
ok_vqa_score = evaluate_vqa(
model=flamingo,
tokenizer=tokenizer,
image_processor=image_processor,
batch_size=args.batch_size,
num_samples=args.num_samples,
num_shots=shot,
device=args.device,
seed=seed,
image_dir_path=args.ok_vqa_image_dir_path,
questions_json_path=args.ok_vqa_questions_json_path,
annotations_json_path=args.ok_vqa_annotations_json_path,
vqa_dataset="ok_vqa",
)
print(f"Shots {shot} Trial {trial} OK-VQA score: {ok_vqa_score}")
scores.append(ok_vqa_score)
print(f"Shots {shot} Mean OK-VQA score: {np.mean(scores)}")
results["ok_vqa"].append(
{"shots": shot, "trials": scores, "mean": np.mean(scores)}
)
if args.eval_vqav2:
print("Evaluating on VQAv2...")
for shot in args.shots:
scores = []
for seed, trial in zip(args.trial_seeds, range(args.num_trials)):
vqa_score = evaluate_vqa(
model=flamingo,
tokenizer=tokenizer,
image_processor=image_processor,
batch_size=args.batch_size,
num_samples=args.num_samples,
num_shots=shot,
device=args.device,
seed=seed,
image_dir_path=args.vqav2_image_dir_path,
questions_json_path=args.vqav2_questions_json_path,
annotations_json_path=args.vqav2_annotations_json_path,
vqa_dataset="vqa",
)
print(f"Shots {shot} Trial {trial} VQA score: {vqa_score}")
scores.append(vqa_score)
print(f"Shots {shot} Mean VQA score: {np.mean(scores)}")
results["vqav2"].append(
{"shots": shot, "trials": scores, "mean": np.mean(scores)}
)
if args.eval_imagenet:
print("Evaluating on ImageNet...")
for shot in args.shots:
scores = []
for seed, trial in zip(args.trial_seeds, range(args.num_trials)):
imagenet_score = evaluate_imagenet(
model=flamingo,
tokenizer=tokenizer,
image_processor=image_processor,
batch_size=args.batch_size,
num_samples=args.num_samples,
num_shots=shot,
device=args.device,
seed=seed,
imagenet_root=args.imagenet_root,
)
print(
f"Shots {shot} Trial {trial} " f"ImageNet score: {imagenet_score}"
)
scores.append(imagenet_score)
print(f"Shots {shot} Mean ImageNet score: {np.mean(scores)}")
results["imagenet"].append(
{"shots": shot, "trials": scores, "mean": np.mean(scores)}
)
if args.results_file is not None:
with open(args.results_file, "w") as f:
json.dump(results, f)
def get_random_indices(num_samples, query_set_size, full_dataset, seed):
if num_samples + query_set_size > len(full_dataset):
raise ValueError(
f"num_samples + num_shots must be less than {len(full_dataset)}"
)
# get a random subset of the dataset
np.random.seed(seed)
random_indices = np.random.choice(
len(full_dataset), num_samples + query_set_size, replace=False
)
return random_indices
def prepare_eval_samples_and_dataset(full_dataset, random_indices, query_set_size):
# get in context samples
in_context_samples = [full_dataset[i] for i in random_indices[:query_set_size]]
eval_dataset = torch.utils.data.Subset(
full_dataset, random_indices[query_set_size:]
)
return in_context_samples, eval_dataset
def get_context_images(image_processor, in_context_samples, num_shots):
if num_shots > 0:
context_images = [
image_processor(s["image"]).unsqueeze(0) for s in in_context_samples
]
context_images = torch.cat(context_images, dim=0)
context_images = context_images.unsqueeze(1).unsqueeze(0)
else:
context_images = None
return context_images
def get_context_text(
get_prompt: Callable[[dict], str],
in_context_samples,
effective_num_shots,
num_shots,
) -> str:
context_text = (
"".join([get_prompt(s) for s in in_context_samples])
if effective_num_shots > 0
else ""
)
if num_shots == 0:
context_text = context_text.replace("<image>", "")
return context_text
def prepare_batch_images(batch, image_processor, context_images, num_shots):
batch_images = None
for b, sample_imgs in zip(batch, context_images):
b_image = image_processor(b["image"]).unsqueeze(0).unsqueeze(1).unsqueeze(0)
b_image = torch.cat([sample_imgs, b_image], dim=1) if num_shots > 0 else b_image
if batch_images is None:
batch_images = b_image
else:
batch_images = torch.cat([batch_images, b_image], dim=0)
return batch_images
def sample_batch_demos_from_query_set(query_set, num_samples, batch_size):
return [random.sample(query_set, num_samples) for _ in range(batch_size)]
def get_outputs(
model,
batch_images,
device,
attention_mask,
max_generation_length,
num_beams,
length_penalty,
input_ids,
):
with torch.inference_mode():
outputs = model.generate(
batch_images.to(device if device >= 0 else "cpu"),
input_ids.to(device if device >= 0 else "cpu"),
attention_mask=attention_mask.to(device if device >= 0 else "cpu"),
max_new_tokens=max_generation_length,
num_beams=num_beams,
length_penalty=length_penalty,
)
outputs = outputs[:, len(input_ids[0]) :]
return outputs
def evaluate_coco_flickr(
model,
tokenizer,
image_processor,
batch_size,
image_dir_path,
annotations_json_path,
seed=42,
max_generation_length=20,
num_beams=3,
length_penalty=-2.0,
num_samples=5000,
query_set_size=2048,
num_shots=8,
device=-1,
is_flickr=False,
):
"""Evaluate a model on COCO dataset.
Args:
model (nn.Module): model to evaluate
tokenizer (transformers.PreTrainedTokenizer): tokenizer for the model
image_processor : image processor for the model
batch_size (int): batch size
image_dir_path (str, optional): path to the directory containing the images.
annotations_json_path (str, optional): path to the json file containing the annotations.
seed (int, optional): seed for random number generator. Defaults to 42.
max_generation_length (int, optional): maximum length of the generated caption. Defaults to 10.
num_beams (int, optional): number of beams to use for beam search. Defaults to 3.
length_penalty (float, optional): length penalty for beam search. Defaults to -2.0.
num_samples (int, optional): number of samples to evaluate on. Defaults to 5000.
query_set_size (int, optional): number of samples to use for query set. Defaults to 2048.
num_shots (int, optional): number of in-context samples to use. Defaults to 8.
device (int, optional): device to use. Defaults to -1.
num_workers (int, optional): number of workers to use for dataloader. Defaults to 4.
is_flickr (bool): defines if that data is COCO or Flickr. Defaults to False (COCO).
Returns:
float: CIDEr score
"""
full_dataset = COCOFlickrDataset(
image_dir_path=image_dir_path,
annotations_path=annotations_json_path,
is_flickr=is_flickr,
)
effective_num_shots = num_shots if num_shots > 0 else 2
random_indices = get_random_indices(num_samples, query_set_size, full_dataset, seed)
in_context_samples, eval_dataset = prepare_eval_samples_and_dataset(
full_dataset=full_dataset,
random_indices=random_indices,
query_set_size=query_set_size,
)
model.eval()
def get_prompt(sample):
return f"<image>Output:{sample['caption'].strip()}<|endofchunk|>"
predictions = defaultdict()
desc = "Running inference Flickr30" if is_flickr else "Running inference COCO"
for batch in more_itertools.chunked(tqdm(eval_dataset, desc=desc), batch_size):
batch_demo_samples = sample_batch_demos_from_query_set(
in_context_samples, effective_num_shots, len(batch)
)
context_images = [
get_context_images(
image_processor=image_processor,
in_context_samples=batch_demo_samples[i],
num_shots=num_shots,
)
for i in range(len(batch))
]
context_text = [
get_context_text(
get_prompt,
in_context_samples=batch_demo_samples[i],
effective_num_shots=effective_num_shots,
num_shots=num_shots,
)
for i in range(len(batch))
]
batch_images = prepare_batch_images(
batch=batch,
image_processor=image_processor,
context_images=context_images,
num_shots=num_shots,
)
batch_text = [f"{context_text[i]}<image>Output:" for i in range(len(batch))]
tokenizer.padding_side = "left"
encodings = tokenizer(
batch_text,
padding="longest",
truncation=True,
return_tensors="pt",
max_length=2000,
)
input_ids = encodings["input_ids"]
attention_mask = encodings["attention_mask"]
outputs = get_outputs(
model=model,
batch_images=batch_images,
device=device,
attention_mask=attention_mask,
max_generation_length=max_generation_length,
num_beams=num_beams,
length_penalty=length_penalty,
input_ids=input_ids,
)
new_predictions = [
postprocess_captioning_generation(out).replace('"', "")
for out in tokenizer.batch_decode(outputs, skip_special_tokens=True)
]
for i, sample in enumerate(batch):
predictions[sample["image_id"]] = {
"caption": new_predictions[i],
}
# save the predictions to a temporary file
random_uuid = str(uuid.uuid4())
results_path = (
f"flickrresults_{random_uuid}.json"
if is_flickr
else f"cocoresults_{random_uuid}.json"
)
with open(results_path, "w") as f:
f.write(
json.dumps(
[
{"image_id": k, "caption": predictions[k]["caption"]}
for k in predictions
],
indent=4,
)
)
metrics = compute_cider(
result_path=results_path,
annotations_path=annotations_json_path,
)
# delete the temporary file
os.remove(results_path)
return metrics["CIDEr"] * 100.0
def evaluate_vqa(
model,
tokenizer,
image_processor,
batch_size,
image_dir_path,
questions_json_path,
annotations_json_path,
seed=42,
max_generation_length=5,
num_beams=3,
length_penalty=-2.0,
num_samples=5000,
query_set_size=2048,
num_shots=8,
device=-1,
vqa_dataset="vqa",
):
"""
Evaluate a model on VQA datasets. Currently supports VQA v2.0.
Args:
model (nn.Module): model to evaluate
tokenizer (transformers.PreTrainedTokenizer): tokenizer for the model
image_processor : image processor for the model
batch_size (int): batch size
image_dir_path (str): path to image directory
questions_json_path (str): path to questions json file
annotations_json_path (str): path to annotations json file
seed (int, optional): random seed. Defaults to 42.
max_generation_length (int, optional): max generation length. Defaults to 5.
num_beams (int, optional): number of beams to use for beam search. Defaults to 3.
length_penalty (float, optional): length penalty for beam search. Defaults to -2.0.
num_samples (int, optional): number of samples to evaluate on. Defaults to 5000 samples.
query_set_size (int, optional): size of the query set. Defaults to 2048.
num_shots (int, optional): number of shots to use. Defaults to 8.
device (int, optional): device to use. Defaults to -1 (cpu).
num_workers (int, optional): number of workers to use. Defaults to 4.
vqa_dataset (string): type of vqa dataset: currently supports vqa, ok_vqa. Defaults to vqa.
Returns:
float: accuracy score
"""
full_dataset = VQADataset(
image_dir_path=image_dir_path,
question_path=questions_json_path,
annotations_path=annotations_json_path,
vqa_dataset=vqa_dataset,
)
effective_num_shots = num_shots if num_shots > 0 else 2
if num_samples + effective_num_shots > len(full_dataset):
raise ValueError(
f"num_samples + num_shots must be less than or equal to {len(full_dataset)}"
)
random_indices = get_random_indices(num_samples, query_set_size, full_dataset, seed)
def get_prompt(sample, train=True):
return f"<image>Question:{sample['question'].strip()} Short Answer:{sample['answers'][0].strip() if train else ''}{'<|endofchunk|>' if train else ''}"
in_context_samples, eval_dataset = prepare_eval_samples_and_dataset(
full_dataset=full_dataset,
random_indices=random_indices,
query_set_size=query_set_size,
)
model.eval()
predictions = []
for batch in more_itertools.chunked(
tqdm(eval_dataset, desc="Running inference"), batch_size
):
batch_demo_samples = sample_batch_demos_from_query_set(
in_context_samples, effective_num_shots, len(batch)
)
context_images = [
get_context_images(
image_processor=image_processor,
in_context_samples=batch_demo_samples[i],
num_shots=num_shots,
)
for i in range(len(batch))
]
context_text = [
get_context_text(
get_prompt,
in_context_samples=batch_demo_samples[i],
effective_num_shots=effective_num_shots,
num_shots=num_shots,
)
for i in range(len(batch))
]
batch_images = prepare_batch_images(
batch=batch,
image_processor=image_processor,
context_images=context_images,
num_shots=num_shots,
)
batch_text = [
context_text[i] + get_prompt(s, train=False) for i, s in enumerate(batch)
]
tokenizer.padding_side = "left"
encodings = tokenizer(
batch_text,
return_tensors="pt",
padding="longest",
truncation=True,
max_length=2000,
)
input_ids = encodings["input_ids"].to(device if device >= 0 else "cpu")
attention_mask = encodings["attention_mask"].to(
device if device >= 0 else "cpu"
)
outputs = get_outputs(
model=model,
batch_images=batch_images,
device=device,
attention_mask=attention_mask,
max_generation_length=max_generation_length,
num_beams=num_beams,
length_penalty=length_penalty,
input_ids=input_ids,
)
process_function = (
postprocess_vqa_generation
if vqa_dataset == "vqa"
else postprocess_ok_vqa_generation
)
new_predictions = [
process_function(out)
for out in tokenizer.batch_decode(outputs, skip_special_tokens=True)
]
predictions.extend(
[
{"answer": p, "question_id": sample["question_id"]}
for p, sample in zip(new_predictions, batch)
]
)
# save the predictions to a temporary file
random_uuid = str(uuid.uuid4())
with open(f"{vqa_dataset}results_{random_uuid}.json", "w") as f:
f.write(json.dumps(predictions, indent=4))
acc = compute_vqa_accuracy(
f"{vqa_dataset}results_{random_uuid}.json",
questions_json_path,
annotations_json_path,
)
# delete the temporary file
os.remove(f"{vqa_dataset}results_{random_uuid}.json")
return acc
def evaluate_imagenet(
model,
tokenizer,
image_processor,
batch_size: int,
imagenet_root: str,
seed: int = 42,
num_samples: int = 5000,
num_shots: int = 8,
device: int = -1,
):
"""
Evaluate a model on ImageNet dataset.
Args:
model: model to evaluate
tokenizer (transformers.PreTrainedTokenizer): tokenizer for the model
image_processor : image processor for the model
batch_size (int): batch size
imagenet_root (str): path to imagenet root for the specified split.
seed (int, optional): random seed. Defaults to 42.
num_samples (int, optional): number of samples to evaluate on. Defaults to 5000 samples.
num_shots (int, optional): number of shots to use. Defaults to 8.
device (int, optional): device to use. Defaults to -1 (cpu).
Returns:
float: accuracy score
"""
full_dataset = ImageNetDataset(root=imagenet_root)
effective_num_shots = num_shots if num_shots > 0 else 2
if num_samples + effective_num_shots > len(full_dataset):
raise ValueError(
f"num_samples + num_shots must be less than or equal to "
f"{len(full_dataset)} "
)
random_indices = get_random_indices(
num_samples, effective_num_shots, full_dataset, seed
)
eoc_token = "<|endofchunk|>"
eoc_token_id = tokenizer.additional_special_tokens_ids[
tokenizer.additional_special_tokens.index(eoc_token)
]
# Padding from right allows efficient precomputing of context activations.
tokenizer.padding_side = "right"
def _imagenet_prompt(class_name, is_context: bool = True):
"""Construct an imagenet prompt for a given label."""
prefix = "<image>A photo of a "
if is_context:
return prefix + class_name.strip()
else:
# Not a context example; insert EOS token before the class name
# so that we can compute the loss on the class name tokens only.
return prefix + tokenizer.eos_token + class_name.strip()
def get_imagenet_prompt(x: dict, is_context: bool = True) -> str:
"""Construct an ImageNet prompt for an example, using its label."""
return _imagenet_prompt(x["class_name"], is_context=is_context)
in_context_samples, eval_dataset = prepare_eval_samples_and_dataset(
full_dataset=full_dataset,
random_indices=random_indices,
query_set_size=effective_num_shots, # NOTE: here we replace query_set_size with effective_num_shots but this is not the ideal evaluation setting.
# TODO: We should add a query_set_size argument to the function and use it to randomly sample the context for each example.
# This will be more consistent with the evaluation setting in the paper but will require some reworking of the caching.
)
device = device if device >= 0 else "cpu"
model.eval()
# Predictions based on the class target sequence with the maximal
# predicted probability
predictions_max_prob = []
# Predictions based on the class target sequence with the minimal loss on
# the model logits
predictions_min_loss = []
labels = []
context_images = [
get_context_images(
image_processor=image_processor,
in_context_samples=in_context_samples,
num_shots=num_shots,
)
for _ in range(batch_size)
]
context_text = get_context_text(
get_imagenet_prompt,
in_context_samples=in_context_samples,
effective_num_shots=effective_num_shots,
num_shots=num_shots,
)
# kwargs to use when calling tokenizer
tokenizer_kwargs = {
"return_tensors": "pt",
"padding": True,
"truncation": True,
"max_length": 256,
}
for i, batch in enumerate(more_itertools.chunked(eval_dataset, batch_size)):
print(f"processing batch {i} of {ceil(len(eval_dataset) / batch_size)}")
batch_per_class_probs = []
batch_per_class_losses = []
batch_images = prepare_batch_images(
batch=batch,
image_processor=image_processor,
context_images=context_images,
num_shots=num_shots,
)
# Process the images only once.
batch_images = batch_images.to(device)
model._encode_vision_x(vision_x=batch_images)
# Process the context text only once.
context_encodings = tokenizer([context_text] * batch_size, **tokenizer_kwargs)
context_ids = context_encodings["input_ids"].to(device)
context_len = context_ids.shape[-1]
context_precomputed = model(
None,
context_ids,
use_cached_vision_x=True,
clear_conditioned_layers=False,
use_cache=True,
)
# For each ImageNet class, construct the output prompt, compute a
# forward pass, and store the results.
for imagenet_class_name in tqdm(openai_imagenet_classnames):
batch_text = [
context_text + _imagenet_prompt(imagenet_class_name, False) + eoc_token
] * batch_size
full_batch_encodings = tokenizer(batch_text, **tokenizer_kwargs)
# full_batch_input_ids has shape [batch_size, seq_len], but we
# only need to run inference on the [batch_size,
# context_len:] inputs that have not been precomputed and
# vary per class.
full_batch_input_ids = full_batch_encodings["input_ids"].to(device)
full_batch_attention_mask = full_batch_encodings["attention_mask"].to(
device
)
# Sanity check that the encoded inputs with context are the same
# as the encoded context alone, for every example in the batch
assert torch.all(
context_ids[0, :] == full_batch_input_ids[:, :context_len]
).item()
# Clone the nested structure of the past key values
past_key_values = tuple(
[
tuple([x.clone() for x in inner])
for inner in context_precomputed.past_key_values
]
)
# Compute the outputs without recomputing context representations.
outputs = model(
vision_x=None,
lang_x=full_batch_input_ids[:, context_len:],
attention_mask=full_batch_attention_mask,
use_cached_vision_x=True,
clear_conditioned_layers=False,
past_key_values=past_key_values,
use_cache=True,
)
logits = torch.concat((context_precomputed.logits, outputs.logits), 1)
per_sample_probs = compute_per_sample_probs(
encodings=full_batch_encodings,
tokenizer=tokenizer,
logits=logits,
eoc_token_id=eoc_token_id,
)
per_sample_loss = compute_per_sample_loss(
encodings=full_batch_encodings,
tokenizer=tokenizer,
logits=logits,
eoc_token_id=eoc_token_id,
)
batch_per_class_probs.append(per_sample_probs.detach())
batch_per_class_losses.append(per_sample_loss.detach())
# Tensor of shape [batch_size, 1000] where the [i,j]th element is
# the (probability or loss) for batch element i on imagenet class j.
batch_probs = torch.stack(batch_per_class_probs, 1)
batch_losses = torch.stack(batch_per_class_losses, 1)
predictions_max_prob.extend(torch.argmax(batch_probs, 1).detach().tolist())
predictions_min_loss.extend(torch.argmin(batch_losses, 1).detach().tolist())
labels.extend(x["class_id"] for x in batch)
acc_max_prob = (np.array(predictions_max_prob) == np.array(labels)).mean()
acc_min_loss = (np.array(predictions_min_loss) == np.array(labels)).mean()
print(f"[DEBUG] ImageNet accuracy with max prob method is {acc_max_prob}")
print(f"[DEBUG] ImageNet accuracy with min loss method is {acc_min_loss}")
print(f"[DEBUG] printing ImageNet predictions and labels:")
for yhat_prob, yhat_loss, y in zip(
predictions_max_prob, predictions_min_loss, labels
):
print(
" " * 30 + f"label: {IMAGENET_1K_CLASS_ID_TO_LABEL[y]}"
f"\nprediction (max prob method): "
f"{IMAGENET_1K_CLASS_ID_TO_LABEL[yhat_prob]}"
f"\nprediction (min loss method): "
f"{IMAGENET_1K_CLASS_ID_TO_LABEL[yhat_loss]}\n"
"#" * 25
)
return acc_max_prob
if __name__ == "__main__":
main()