Generative Representational Instruction Tuning
Paper • 2402.09906 • Published • 54
This model is a fine-tuned version of GritLM/GritLM-7B on the CheckThat! 2025 Subtask 4B dataset for scientific claim verification and source retrieval.
GritLM (Generative Representational Instruction Tuned Language Model) is a unified model that handles both text generation and embedding tasks. This fine-tuned version is specifically optimized for retrieving scientific paper abstracts given claims from social media (tweets).
This model is designed for:
The model was fine-tuned on the CheckThat! 2025 Subtask 4B dataset:
Data Preparation:
Training Configuration:
learning_rate: 2e-5
per_device_train_batch_size: 2
gradient_accumulation_steps: 16
num_train_epochs: 3
max_seq_length: 512
warmup_steps: 300
fp16: false
bf16: true
gradient_checkpointing: true
deepspeed_stage: 2
seed: 42
Based on GritLM-7B which uses:
If you use this model, please cite:
@misc{gritlm-checkthat-2025,
author = {suitch},
title = {GritLM Fine-tuned on CheckThat! 2025 Dataset},
year = {2025},
publisher = {HuggingFace},
url = {https://huggingface.co/suitch/gritlm-checkthat-2025}
}
Base Model Citation:
@misc{muennighoff2024generative,
title={Generative Representational Instruction Tuning},
author={Niklas Muennighoff and Hongjin Su and Liang Wang and Nan Yang and Furu Wei and Tao Yu and Amanpreet Singh and Douwe Kiela},
year={2024},
eprint={2402.09906},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
Dataset Citation:
@inproceedings{checkthat2025,
title={CheckThat! at CLEF 2025: Check-Worthiness, Subjectivity, Persuasion, Roles, Authorities, and Veracity},
booktitle={Proceedings of CLEF 2025},
year={2025}
}
This model inherits the Apache 2.0 license from the base GritLM-7B model.