LIMO: Less is More for Reasoning
Paper
• 2502.03387 • Published
• 62
This model is a fine-tuned version of meta-llama/Llama-3.1-8B-Instruct model. The fine-tuning was performed using Low-Rank Adaptation (LoRA) on the LIMO dataset to enhance the model's reasoning capabilities, based on the work in the paper: LIMO: Less is More for Reasoning.
This repo contains the merged model weights. The LoRA adapter version can be found from here.
To utilize this model for text generation tasks, follow the steps below:
Ensure you have the necessary libraries installed:
pip install torch transformers
from transformers import AutoModelForCausalLM, AutoTokenizer
model_name = "t83714/llama-3.1-8b-instruct-limo"
model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype="auto", device_map="auto")
tokenizer = AutoTokenizer.from_pretrained(model_name)
prompt = "How much is (2+5)x5/7"
# Tokenize the input
inputs = tokenizer(prompt, return_tensors="pt").to("cuda")
# Generate the output
output = model.generate(**inputs, max_length=8000)
print(tokenizer.decode(output[0], skip_special_tokens=True))
The following hyperparameters were used during training:
This model is trained based on the work of Ye et al. (2025). If you use this model, please also consider citing their paper:
@misc{ye2025limoreasoning,
title={LIMO: Less is More for Reasoning},
author={Yixin Ye and Zhen Huang and Yang Xiao and Ethan Chern and Shijie Xia and Pengfei Liu},
year={2025},
eprint={2502.03387},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2502.03387},
}
Base model
meta-llama/Llama-3.1-8B