The Llama 3 Herd of Models
Paper
•
2407.21783
•
Published
•
117
This model is part of the Annoy project for code reasoning and execution specification.
This model has been fine-tuned for speculative execution reasoning tasks. It excels at predicting code execution outcomes and reasoning about program behavior.
The model was trained using the Annoy methodology on the PythonEdu-Rs dataset. The training process involved:
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("toolevalxm/llama3.1-8b_spec")
tokenizer = AutoTokenizer.from_pretrained("toolevalxm/llama3.1-8b_spec")
If you use this model, please cite our paper.
Base Model
This model is fine-tuned from meta-llama/Llama-3.1-8B.
@article{grattafiori2024llama,
title={The Llama 3 Herd of Models},
author={Grattafiori, Aaron and Dubey, Abhimanyu and Jauhri, Abhinav and Pandey, Abhinav and Kadian, Abhishek and Al-Dahle, Ahmad and Letman, Aiesha and Mathur, Akhil and Schelten, Alan and Vaughan, Alex and others},
journal={arXiv preprint arXiv:2407.21783},
year={2024}
}