Datasets:

ArXiv:
shulin16's picture
Upload folder using huggingface_hub
9f3bc09 verified
You are a helpful AI scientist to build up the codebase for me.
This project is to train the open-sourced model to deploy CoT-like reasoning format on text-to-image and text-to-video generation quality assessment. You are using the LLaMAFactory to train the model, and write evaluation functions.
# Preparation
## Data
There are a folder called `ea-data/agent` and there are 3 subfolders:
* `vbench_results`: which stores the results for using proprietary models to evaluate different dimensions in vbench, and the results are CoT style.
* `t2i_results`: which stores the results for using proprietary models to evaluate different dimensions in T2I-CompBench, and the results are CoT style.
* `open_results`: which store the results for using proprietary models to evaluate open-ended queries.
Your first job is to write and execute the python script to clean the data in those aforementioned folders and convert them into the format align with `/home/data2/sltian/code/evaluation_agent_dev/LLaMA-Factory/data/alpaca_en_demo.json`.
如果指定, system 列对应的内容将被作为系统提示词。
history 列是由多个字符串二元组构成的列表,分别代表历史消息中每轮对话的指令和回答。注意在指令监督微调时,历史消息中的回答内容也会被用于模型学习。
指令监督微调数据集 格式要求 如下:
[
{
"instruction": "人类指令(必填)",
"input": "人类输入(选填)",
"output": "模型回答(必填)",
"system": "系统提示词(选填)",
"history": [
["第一轮指令(选填)", "第一轮回答(选填)"],
["第二轮指令(选填)", "第二轮回答(选填)"]
]
}
]
下面提供一个 alpaca 格式 多轮 对话的例子,对于单轮对话只需省略 history 列即可。
[
{
"instruction": "今天的天气怎么样?",
"input": "",
"output": "今天的天气不错,是晴天。",
"history": [
[
"今天会下雨吗?",
"今天不会下雨,是个好天气。"
],
[
"今天适合出去玩吗?",
"非常适合,空气质量很好。"
]
]
}
]
对于上述格式的数据, dataset_info.json 中的 数据集描述 应为:
"数据集名称": {
"file_name": "data.json",
"columns": {
"prompt": "instruction",
"query": "input",
"response": "output",
"system": "system",
"history": "history"
}
}
## Train
After cleaning and collecting the data, you should write a script to train the `Qwen2.5-3B-Instruct` model using this created dataset.
The training is using `LLaMA-Factory`. You should read the dir and write a script to train the model.