--- license: mit inference: false datasets: - freecs/ArtificialThinkerSet base_model: microsoft/phi-2 --- # The First Open-Source Reasoning LLM **December 28, 2023** - This model was created 11 months before OpenAI's o1 release. ## Historical Context In late 2023, I was experimenting with fine-tuning open-source models. Working with limited computational resources (primarily free Colab notebooks with T4 GPUs), I focused on developing novel approaches and new paradigms to significantly enhance LLM capabilities without simply scaling the number of parameters, since that would have required substantial computational resources. **Proof of timeline:** Check the [initial commit](https://huggingface.co/freecs/ArtificialThinker-Phi2/commit/8ce7acd72fb187cd3c3e76a8c0c58b8246e85d23) - December 28, 2023. ## Technical Approach The model uses a custom chat template that includes a "reasoning" step before providing the output to the user: ``` <|system|>sys_message <|prompt|>prompt <|reasoning|>reasoning <|response|>response<|endoftext|> ``` To test this approach, I created the [ArtificialThinkerSet](https://huggingface.co/datasets/freecs/ArtificialThinkerSet) dataset to fine-tune Phi-2. I also wrote ["Reasoning Is All You Need"](https://freecs.org/paper.html) - a blog post explaining this approach. You can find me at [gr.bio](https://gr.bio/).