Update README.md
Browse files
README.md
CHANGED
|
@@ -17,6 +17,8 @@ tags:
|
|
| 17 |
😸<a href="https://github.com/prsdm/fine-tuning-llms/blob/main/Fine-tuning-phi-2-model.ipynb">GitHub</a> •📝<a href="https://medium.com/@prasadmahamulkar/fine-tuning-phi-2-a-step-by-step-guide-e672e7f1d009">Article</a> • Models & Datasets on: 🤗<a href="https://huggingface.co/prsdm">Hugging Face</a>
|
| 18 |
</p>
|
| 19 |
|
|
|
|
|
|
|
| 20 |
This repository provides a collection of Jupyter notebooks that demonstrate how to fine-tune large language models using various tools and techniques.
|
| 21 |
|
| 22 |
fine-tuning or instruction tuning is the process where the pre-trained model is further trained on the smaller dataset to adapt its knowledge for a specific task or domain. This process tweaks the model’s parameters to perform specific tasks. In fine-tuning, there are two methods:
|
|
@@ -35,7 +37,3 @@ In RLHF, the model interacts with users, generates responses, and receives feedb
|
|
| 35 |
| [llama-2](https://huggingface.co/meta-llama/Llama-2-7b-chat-hf) | This model has been fine-tuned on a dataset with human-generated prompts to answer questions related to general knowledge. (used SFT method) | [Dataset](https://huggingface.co/datasets/prsdm/finance-llama2-1k) | [](https://colab.research.google.com/github/prsdm/fine-tuning-llms/blob/main/Fine-tuning-llama-2-model.ipynb) | [llama-2-finance](https://huggingface.co/prsdm/llama-2-finance) |
|
| 36 |
|
| 37 |
|
| 38 |
-
### Diagram:
|
| 39 |
-
|
| 40 |
-

|
| 41 |
-
|
|
|
|
| 17 |
😸<a href="https://github.com/prsdm/fine-tuning-llms/blob/main/Fine-tuning-phi-2-model.ipynb">GitHub</a> •📝<a href="https://medium.com/@prasadmahamulkar/fine-tuning-phi-2-a-step-by-step-guide-e672e7f1d009">Article</a> • Models & Datasets on: 🤗<a href="https://huggingface.co/prsdm">Hugging Face</a>
|
| 18 |
</p>
|
| 19 |
|
| 20 |
+

|
| 21 |
+
|
| 22 |
This repository provides a collection of Jupyter notebooks that demonstrate how to fine-tune large language models using various tools and techniques.
|
| 23 |
|
| 24 |
fine-tuning or instruction tuning is the process where the pre-trained model is further trained on the smaller dataset to adapt its knowledge for a specific task or domain. This process tweaks the model’s parameters to perform specific tasks. In fine-tuning, there are two methods:
|
|
|
|
| 37 |
| [llama-2](https://huggingface.co/meta-llama/Llama-2-7b-chat-hf) | This model has been fine-tuned on a dataset with human-generated prompts to answer questions related to general knowledge. (used SFT method) | [Dataset](https://huggingface.co/datasets/prsdm/finance-llama2-1k) | [](https://colab.research.google.com/github/prsdm/fine-tuning-llms/blob/main/Fine-tuning-llama-2-model.ipynb) | [llama-2-finance](https://huggingface.co/prsdm/llama-2-finance) |
|
| 38 |
|
| 39 |
|
|
|
|
|
|
|
|
|
|
|
|