|
|
import json |
|
|
import random |
|
|
import jsonlines |
|
|
import os |
|
|
|
|
|
def load_data_jsonl(data_path): |
|
|
data = [] |
|
|
with open(data_path, "r+", encoding="utf8") as f: |
|
|
for item in jsonlines.Reader(f): |
|
|
data.append(item) |
|
|
|
|
|
return data |
|
|
|
|
|
def load_data(data_path): |
|
|
with open(data_path, 'r') as f: |
|
|
data = json.load(f) |
|
|
|
|
|
return data |
|
|
|
|
|
def ensure_dir_exists(path): |
|
|
"""Create directory if it doesn't exist""" |
|
|
directory = os.path.dirname(path) |
|
|
if not os.path.exists(directory): |
|
|
os.makedirs(directory) |
|
|
print(f"Created directory: {directory}") |
|
|
|
|
|
def build_dataset(data_list, path): |
|
|
with open('/fs-computility/niuyazhe/lixueyan/meme/dataset-meme-rewardmodel/prompt/reward_model_prompt.txt', 'r') as f: |
|
|
PROMPT = f.read() |
|
|
|
|
|
dict_list = [] |
|
|
for id, d in enumerate(data_list): |
|
|
data_json = {'id': id, |
|
|
'image': d["image_list"], |
|
|
'conversations': [ |
|
|
{'from': 'human', 'value': f'{PROMPT}\nFirst image: <image>\nSecond image:<image>'}, |
|
|
{'from': 'gpt', 'value': d["label"]} |
|
|
]} |
|
|
dict_list.append(data_json) |
|
|
with open(path, 'w', encoding='utf-8') as file: |
|
|
for entry in dict_list: |
|
|
json.dump(entry, file) |
|
|
file.write('\n') |
|
|
return len(dict_list) |
|
|
|
|
|
def build_dataset_multihead(data_list, path, mask): |
|
|
with open('/fs-computility/niuyazhe/lixueyan/meme/dataset-meme-rewardmodel/prompt/reward_model_prompt.txt', 'r') as f: |
|
|
PROMPT = f.read() |
|
|
|
|
|
dict_list = [] |
|
|
for id, d in enumerate(data_list): |
|
|
data_json = {'id': id, |
|
|
'image': d["image_list"], |
|
|
'conversations': [ |
|
|
{'from': 'human', 'value': f'{PROMPT}\nFirst image: <image>\nSecond image:<image>'}, |
|
|
{'from': 'gpt', 'value': [[d["label"]]*2, mask]} |
|
|
]} |
|
|
dict_list.append(data_json) |
|
|
with open(path, 'w', encoding='utf-8') as file: |
|
|
for entry in dict_list: |
|
|
json.dump(entry, file) |
|
|
file.write('\n') |
|
|
return len(dict_list) |
|
|
|
|
|
def build_dataset_cross(data_list, path, TYPE): |
|
|
with open('/fs-computility/niuyazhe/lixueyan/meme/dataset-meme-rewardmodel/prompt/reward_model_prompt.txt', 'r') as f: |
|
|
PROMPT = f.read() |
|
|
|
|
|
dict_list = [] |
|
|
origin_image_list = [] |
|
|
boring_image_list = [] |
|
|
origin_text_lengths = [] |
|
|
boring_text_lengths = [] |
|
|
for id, d in enumerate(data_list): |
|
|
if d["label"] == 0: |
|
|
origin_image_list.append(d["image_list"][0]) |
|
|
boring_image_list.append(d["image_list"][1]) |
|
|
origin_text_lengths.append(d["text_lengths"][0]) |
|
|
boring_text_lengths.append(d["text_lengths"][1]) |
|
|
elif d["label"] == 1: |
|
|
origin_image_list.append(d["image_list"][1]) |
|
|
boring_image_list.append(d["image_list"][0]) |
|
|
origin_text_lengths.append(d["text_lengths"][1]) |
|
|
boring_text_lengths.append(d["text_lengths"][0]) |
|
|
else: |
|
|
raise ValueError("Wrong label") |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
print(f'sorting the boring images') |
|
|
|
|
|
boring_with_lengths = list(zip(boring_image_list, boring_text_lengths)) |
|
|
boring_with_lengths.sort(key=lambda x: x[1]) |
|
|
|
|
|
print(f'generating the pairs') |
|
|
for id, origin in enumerate(origin_image_list): |
|
|
original_length = origin_text_lengths[id] |
|
|
|
|
|
|
|
|
longer_idx = 0 |
|
|
while longer_idx < len(boring_with_lengths) and boring_with_lengths[longer_idx][1] <= original_length: |
|
|
longer_idx += 1 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
boring = random.choice(boring_with_lengths)[0] |
|
|
|
|
|
pos_neg = random.choice(["pos", "neg"]) |
|
|
if pos_neg == 'pos': |
|
|
data_json = {'id': id, |
|
|
'image': [origin, boring], |
|
|
'conversations': [ |
|
|
{'from': 'human', 'value': f'{PROMPT}\nFirst image: <image>\nSecond image:<image>'}, |
|
|
{'from': 'gpt', 'value': 0} |
|
|
]} |
|
|
dict_list.append(data_json) |
|
|
else: |
|
|
data_json = {'id': id, |
|
|
'image': [boring, origin], |
|
|
'conversations': [ |
|
|
{'from': 'human', 'value': f'{PROMPT}\nFirst image: <image>\nSecond image:<image>'}, |
|
|
{'from': 'gpt', 'value': 1} |
|
|
]} |
|
|
dict_list.append(data_json) |
|
|
with open(path, 'w', encoding='utf-8') as file: |
|
|
for entry in dict_list: |
|
|
json.dump(entry, file) |
|
|
file.write('\n') |
|
|
return len(dict_list) |
|
|
|
|
|
def build_dataset_cross_multihead(data_list, path, TYPE, mask): |
|
|
with open('/fs-computility/niuyazhe/lixueyan/meme/dataset-meme-rewardmodel/prompt/reward_model_prompt.txt', 'r') as f: |
|
|
PROMPT = f.read() |
|
|
|
|
|
dict_list = [] |
|
|
origin_image_list = [] |
|
|
boring_image_list = [] |
|
|
origin_text_lengths = [] |
|
|
boring_text_lengths = [] |
|
|
for id, d in enumerate(data_list): |
|
|
if d["label"] == 0: |
|
|
origin_image_list.append(d["image_list"][0]) |
|
|
boring_image_list.append(d["image_list"][1]) |
|
|
origin_text_lengths.append(d["text_lengths"][0]) |
|
|
boring_text_lengths.append(d["text_lengths"][1]) |
|
|
elif d["label"] == 1: |
|
|
origin_image_list.append(d["image_list"][1]) |
|
|
boring_image_list.append(d["image_list"][0]) |
|
|
origin_text_lengths.append(d["text_lengths"][1]) |
|
|
boring_text_lengths.append(d["text_lengths"][0]) |
|
|
else: |
|
|
raise ValueError("Wrong label") |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
print(f'sorting the boring images') |
|
|
|
|
|
boring_with_lengths = list(zip(boring_image_list, boring_text_lengths)) |
|
|
boring_with_lengths.sort(key=lambda x: x[1]) |
|
|
|
|
|
print(f'generating the pairs') |
|
|
for id, origin in enumerate(origin_image_list): |
|
|
original_length = origin_text_lengths[id] |
|
|
|
|
|
|
|
|
longer_idx = 0 |
|
|
while longer_idx < len(boring_with_lengths) and boring_with_lengths[longer_idx][1] <= original_length: |
|
|
longer_idx += 1 |
|
|
|
|
|
|
|
|
if longer_idx < len(boring_with_lengths) and random.random() < 0.7: |
|
|
|
|
|
boring = random.choice(boring_with_lengths[longer_idx:])[0] |
|
|
else: |
|
|
|
|
|
if longer_idx > 0: |
|
|
boring = random.choice(boring_with_lengths[:longer_idx])[0] |
|
|
else: |
|
|
boring = random.choice(boring_with_lengths)[0] |
|
|
|
|
|
pos_neg = random.choice(["pos", "neg"]) |
|
|
if pos_neg == 'pos': |
|
|
data_json = {'id': id, |
|
|
'image': [origin, boring], |
|
|
'conversations': [ |
|
|
{'from': 'human', 'value': f'{PROMPT}\nFirst image: <image>\nSecond image:<image>'}, |
|
|
{'from': 'gpt', 'value': [[0]*2, mask]} |
|
|
]} |
|
|
dict_list.append(data_json) |
|
|
else: |
|
|
data_json = {'id': id, |
|
|
'image': [boring, origin], |
|
|
'conversations': [ |
|
|
{'from': 'human', 'value': f'{PROMPT}\nFirst image: <image>\nSecond image:<image>'}, |
|
|
{'from': 'gpt', 'value': [[1]*2, mask]} |
|
|
]} |
|
|
dict_list.append(data_json) |
|
|
with open(path, 'w', encoding='utf-8') as file: |
|
|
for entry in dict_list: |
|
|
json.dump(entry, file) |
|
|
file.write('\n') |
|
|
return len(dict_list) |
|
|
|
|
|
def build_json(dataset_path_list, length_list, name_list, json_path): |
|
|
dict_list = [] |
|
|
for dataset_path, length, name in zip(dataset_path_list, length_list, name_list): |
|
|
dict = { |
|
|
f"{name}": { |
|
|
"root": "", |
|
|
"annotation": dataset_path, |
|
|
"data_augment": False, |
|
|
"repeat_time": 1, |
|
|
"length": length |
|
|
} |
|
|
} |
|
|
dict_list.append(dict) |
|
|
|
|
|
with open(json_path, 'w', encoding='utf-8') as file: |
|
|
for dict in dict_list: |
|
|
json.dump(dict, file) |
|
|
file.write('\n') |
|
|
|
|
|
def split_train_test(data, train_path, test_path): |
|
|
random.shuffle(data) |
|
|
|
|
|
selected_items = data[:int(len(data) * 0.9)] |
|
|
unselected_items = data[int(len(data) * 0.9):] |
|
|
|
|
|
with open(train_path, 'w') as f: |
|
|
json.dump(selected_items, f) |
|
|
|
|
|
with open(test_path, 'w') as f: |
|
|
json.dump(unselected_items, f) |
|
|
|
|
|
return selected_items, unselected_items |
|
|
|
|
|
def split_train_test_original(original_dataset): |
|
|
|
|
|
original_data = load_data(original_dataset) |
|
|
random.shuffle(original_data) |
|
|
|
|
|
|
|
|
train_data_original = original_data[:int(len(original_data) * 0.9)] |
|
|
test_data_original = original_data[int(len(original_data) * 0.9):] |
|
|
|
|
|
|
|
|
train_image_ids = [] |
|
|
for item in train_data_original: |
|
|
|
|
|
filename = item["original_image"].split("/")[-1] |
|
|
train_image_ids.append(filename) |
|
|
|
|
|
test_image_ids = [] |
|
|
for item in test_data_original: |
|
|
|
|
|
filename = item["original_image"].split("/")[-1] |
|
|
test_image_ids.append(filename) |
|
|
|
|
|
with open('/fs-computility/niuyazhe/lixueyan/meme/dataset-meme-rewardmodel/Eimages_train_ids.jsonl', 'w') as f: |
|
|
json.dump(train_image_ids, f) |
|
|
|
|
|
with open('/fs-computility/niuyazhe/lixueyan/meme/dataset-meme-rewardmodel/Eimages_test_ids.jsonl', 'w') as f: |
|
|
json.dump(test_image_ids, f) |
|
|
|
|
|
if __name__ == '__main__': |
|
|
NAME_list = ['object_add'] |
|
|
TYPE_list = ['cross', ''] |
|
|
|
|
|
mask_dict = { |
|
|
'text_replaced': [1, 1], |
|
|
'lowperformancememe': [1, 0], |
|
|
'irrelevantmeme': [0, 1], |
|
|
'boringmeme': [1, 0] |
|
|
} |
|
|
|
|
|
for NAME in NAME_list: |
|
|
for TYPE in TYPE_list: |
|
|
if NAME == 'lowperformancememe': |
|
|
dataset = f'/fs-computility/niuyazhe/lixueyan/meme/memetrash/{NAME}.jsonl' |
|
|
elif NAME == 'text_replaced' or NAME == 'boring_detailed': |
|
|
dataset = f'/fs-computility/niuyazhe/lixueyan/meme/memetrash/Eimages_{NAME}.json' |
|
|
else: |
|
|
|
|
|
dataset = "/fs-computility/niuyazhe/shared/meme/data/meme/Eimages/Eimages_object_2.jsonl" |
|
|
|
|
|
|
|
|
original_dataset = '/fs-computility/niuyazhe/lixueyan/jmj/DIlab/meme/memetrash/processed_dections_Eimage_UPDATED.json' |
|
|
train_image_ids = load_data_jsonl('/fs-computility/niuyazhe/lixueyan/meme/dataset-meme-rewardmodel/Eimages_train_ids.jsonl') |
|
|
test_image_ids = load_data_jsonl('/fs-computility/niuyazhe/lixueyan/meme/dataset-meme-rewardmodel/Eimages_test_ids.jsonl') |
|
|
|
|
|
|
|
|
|
|
|
if TYPE != '': |
|
|
dataset_path_train =f'/fs-computility/niuyazhe/lixueyan/meme/dataset-meme-rewardmodel/{NAME}_{TYPE}/Ejson/{NAME}_{TYPE}_train.jsonl' |
|
|
dataset_path_test = f'/fs-computility/niuyazhe/lixueyan/meme/dataset-meme-rewardmodel/{NAME}_{TYPE}/Ejson/{NAME}_{TYPE}_test.jsonl' |
|
|
json_path_train = f'/fs-computility/niuyazhe/lixueyan/meme/dataset-meme-rewardmodel/{NAME}_{TYPE}/{NAME}_{TYPE}_train.jsonl' |
|
|
json_path_test = f'/fs-computility/niuyazhe/lixueyan/meme/dataset-meme-rewardmodel/{NAME}_{TYPE}/{NAME}_{TYPE}_test.jsonl' |
|
|
|
|
|
else: |
|
|
dataset_path_train =f'/fs-computility/niuyazhe/lixueyan/meme/dataset-meme-rewardmodel/{NAME}/Ejson/{NAME}_train.jsonl' |
|
|
dataset_path_test = f'/fs-computility/niuyazhe/lixueyan/meme/dataset-meme-rewardmodel/{NAME}/Ejson/{NAME}_test.jsonl' |
|
|
json_path_train = f'/fs-computility/niuyazhe/lixueyan/meme/dataset-meme-rewardmodel/{NAME}/{NAME}_train.jsonl' |
|
|
json_path_test = f'/fs-computility/niuyazhe/lixueyan/meme/dataset-meme-rewardmodel/{NAME}/{NAME}_test.jsonl' |
|
|
|
|
|
train_path = f'/fs-computility/niuyazhe/lixueyan/meme/dataset-meme-rewardmodel/{NAME}/raw_data/train.json' |
|
|
test_path = f'/fs-computility/niuyazhe/lixueyan/meme/dataset-meme-rewardmodel/{NAME}/raw_data/test.json' |
|
|
|
|
|
|
|
|
ensure_dir_exists(dataset_path_train) |
|
|
ensure_dir_exists(dataset_path_test) |
|
|
ensure_dir_exists(json_path_train) |
|
|
ensure_dir_exists(json_path_test) |
|
|
ensure_dir_exists(train_path) |
|
|
ensure_dir_exists(test_path) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
train_data = load_data(train_path) |
|
|
test_data = load_data(test_path) |
|
|
|
|
|
if 'meme' in NAME: |
|
|
name = NAME[:-4] |
|
|
else: |
|
|
name = NAME |
|
|
|
|
|
if TYPE == '': |
|
|
length_train = build_dataset(train_data, dataset_path_train) |
|
|
build_json([dataset_path_train], [length_train], [name], json_path_train) |
|
|
length_test = build_dataset(test_data, dataset_path_test) |
|
|
build_json([dataset_path_test], [length_test], [name], json_path_test) |
|
|
|
|
|
elif TYPE == 'cross': |
|
|
length_train = build_dataset_cross(train_data, dataset_path_train, NAME) |
|
|
build_json([dataset_path_train], [length_train], [name+'_'+TYPE], json_path_train) |
|
|
length_test = build_dataset_cross(test_data, dataset_path_test, NAME) |
|
|
build_json([dataset_path_test], [length_test], [name+'_'+TYPE], json_path_test) |
|
|
|
|
|
elif TYPE == 'align_multihead': |
|
|
length_train = build_dataset_multihead(train_data, dataset_path_train, mask_dict[NAME]) |
|
|
build_json([dataset_path_train], [length_train], [name], json_path_train) |
|
|
length_test = build_dataset_multihead(test_data, dataset_path_test, mask_dict[NAME]) |
|
|
build_json([dataset_path_test], [length_test], [name], json_path_test) |
|
|
elif TYPE == 'cross_multihead': |
|
|
length_train = build_dataset_cross_multihead(train_data, dataset_path_train, NAME, mask_dict[NAME]) |
|
|
build_json([dataset_path_train], [length_train], [name+'_'+TYPE], json_path_train) |
|
|
length_test = build_dataset_cross_multihead(test_data, dataset_path_test, NAME, mask_dict[NAME]) |
|
|
build_json([dataset_path_test], [length_test], [name+'_'+TYPE], json_path_test) |
|
|
|
|
|
print(f'Done {NAME} {TYPE}') |
|
|
|