You are a helpful AI scientist to build up the codebase for me. This project is to train the open-sourced model to deploy CoT-like reasoning format on text-to-image and text-to-video generation quality assessment. You are using the LLaMAFactory to train the model, and write evaluation functions. # Preparation ## Data There are a folder called `ea-data/agent` and there are 3 subfolders: * `vbench_results`: which stores the results for using proprietary models to evaluate different dimensions in vbench, and the results are CoT style. * `t2i_results`: which stores the results for using proprietary models to evaluate different dimensions in T2I-CompBench, and the results are CoT style. * `open_results`: which store the results for using proprietary models to evaluate open-ended queries. Your first job is to write and execute the python script to clean the data in those aforementioned folders and convert them into the format align with `/home/data2/sltian/code/evaluation_agent_dev/LLaMA-Factory/data/alpaca_en_demo.json`. 如果指定, system 列对应的内容将被作为系统提示词。 history 列是由多个字符串二元组构成的列表,分别代表历史消息中每轮对话的指令和回答。注意在指令监督微调时,历史消息中的回答内容也会被用于模型学习。 指令监督微调数据集 格式要求 如下: [ { "instruction": "人类指令(必填)", "input": "人类输入(选填)", "output": "模型回答(必填)", "system": "系统提示词(选填)", "history": [ ["第一轮指令(选填)", "第一轮回答(选填)"], ["第二轮指令(选填)", "第二轮回答(选填)"] ] } ] 下面提供一个 alpaca 格式 多轮 对话的例子,对于单轮对话只需省略 history 列即可。 [ { "instruction": "今天的天气怎么样?", "input": "", "output": "今天的天气不错,是晴天。", "history": [ [ "今天会下雨吗?", "今天不会下雨,是个好天气。" ], [ "今天适合出去玩吗?", "非常适合,空气质量很好。" ] ] } ] 对于上述格式的数据, dataset_info.json 中的 数据集描述 应为: "数据集名称": { "file_name": "data.json", "columns": { "prompt": "instruction", "query": "input", "response": "output", "system": "system", "history": "history" } } ## Train After cleaning and collecting the data, you should write a script to train the `Qwen2.5-3B-Instruct` model using this created dataset. The training is using `LLaMA-Factory`. You should read the dir and write a script to train the model.