--- license: cc-by-4.0 dataset_info: features: - name: question_id dtype: string - name: question_type_id dtype: string - name: question_type_name dtype: string - name: figure_id list: string - name: question dtype: string - name: answer dtype: string - name: instructions dtype: string - name: url dtype: string - name: extra_input_figure_ids list: string - name: extra_input_figure_bboxes sequence: sequence: int64 - name: data_fact dtype: string - name: difficulty dtype: string - name: chart_type dtype: string splits: - name: text num_bytes: 25387749 num_examples: 50920 - name: visual_metaphor num_bytes: 407203 num_examples: 462 - name: visual_basic num_bytes: 7166829 num_examples: 7475 download_size: 10423864 dataset_size: 32961781 configs: - config_name: default data_files: - split: text path: data/text-* - split: visual_metaphor path: data/visual_metaphor-* - split: visual_basic path: data/visual_basic-* --- # InfoChartQA: Benchmark for Multimodal Question Answering on Infographic Charts 🤗[Dataset](https://huggingface.co/datasets/Jietson/InfoChartQA) # Dataset You can find our dataset on huggingface: 🤗[InfoChartQA Dataset](https://huggingface.co/datasets/Jietson/InfoChartQA) # Usage Each question entry is arranged as follows. **Note that for visual questions, there may be some extra input figures, which are cropped from the orginal figure. We present their bboxes in "extra_input_figure_bbox".** ``` { "question_id": id of the question, "question_type_name": question type name, for example: "extreme" questions, "question_type_id": question type id, this is only used for evaluation! For example: 72 means "extreme" questions, "figure_id": id of the figure, "question": question text, "answer": ground truth answer, "instructions": instructions, "url": url of the input image, "extra_input_figure_ids": ids of the extra input figures, "extra_input_figure_bboxes": bbox of the extra input figures, in [x,y,w,h] format w/o normalization, "data_fact": data fact of the question, only for text-based questions, "difficulty": difficulty level, "chart_type": chart_type, } ``` Each question is built by: ``` input_image: item["url"] (may need to download for models that don't support url input) extra_input_image: Cropped input_image using item["extra_input_figure_bboxes"], input_text: item["question"] + item["instructions"] (if any) ``` where ``item`` is an entry of the dataset. # Evaluate You should store and evaluate model's response as: ```python # Example code for evaluate def build_question(query): question = query['question'] if "instructions" in query: question += query["instructions"] return question #### Run your model and save your answer Responses = {} for query in tqdm(ds): query_idx = query["question_id"] input_text = build_question(query) input_figure = query["url"] # This should be a list of url for models that support url input """ Note that for models that do not support url input, you may need to download images first. For example, for model like Qwen2.5-VL, you may need to down load the image first and pass the local image path to the model, like: input_figure = YOUR_LOCAL_IMAGE_PATH OF query['figure_id'] Moreover, for questions with extra figure input, you may need to crop figure, for example, extra_input_figures = [crop(input_figure,bbox) for bbox in query["extra_input_figure_bboxes"]] """ # Replace with your model response = model.generate(input_text, input_figure) Responses[query_idx] = { "qtype": int(query["question_type_id"]), # Note that "question_type_id" are used for evaluation only! "answer": query["answer"], "question_id": query_idx, "response": response, } with open("./model_response.json", "w", encoding="utf-8") as f: json.dump(Responses, f, indent=2, ensure_ascii=False) ```