File size: 2,043 Bytes
76d9c4f
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
---
title: Llama2-7B Fine-tuning
description: Fine-tune Llama2-7B with a code instructions dataset
version: EN
---

## Try out this model on [VESSL Hub](https://vessl.ai/hub).

This example fine-tunes [Llama 2](https://ai.meta.com/llama/) on a code instruction dataset. The code instruction dataset is consisted of 1.6K samples and follows the format of Stanford's [Alpaca dataset](https://github.com/gururise/AlpacaDataCleaned). 
To optimize the training process into a single GPU with moderate memory, the model uses [8 bit quantization](https://huggingface.co/blog/hf-bitsandbytes-integration) and LoRA (Low-Rank Adaptation).

In the code we are referencing under `/code/`, we added our Python SDK for logging key metrics like loss and learning rate. You can check these values in real-time under Plots. The run completes by uploading the model checkpoint to VESSL AI model registry, as defined under `export`. 

<img
  className="rounded-md"
  src="/images/llama2-metrics.png"
/>
<img
  className="rounded-md"
  src="/images/llama2-uploaded-model.png"
/>

## Running the model

You can run the model with our quick command.
```sh
vessl run create -f llama2_fine-tuning.yaml
```

Here's a rundown of the `llama2_fine-tuning.yaml` file.
```yaml
name: llama2-finetuning
description: finetune llama2 with code instruction alpaca dataset
resources:
  cluster: vessl-gcp-oregon
  preset: v1.l4-1.mem-27
image: quay.io/vessl-ai/hub:torch2.1.0-cuda12.2-202312070053
import:
  /model/: vessl-model://vessl-ai/llama2/1
  /code/:
    git:
      url: https://github.com/vessl-ai/hub-model
      ref: main
  /dataset/: vessl-dataset://vessl-ai/code_instructions_small_alpaca
export:
  /trained_model/: vessl-model://vessl-ai/llama2-finetuned
  /artifacts/: vessl-artifact://
run:
  - command: |-
      pip install -r requirements.txt
      mkdir /model_
      cd /model
      mv llama_2_7b_hf.zip /model_
      cd /model_
      unzip llama_2_7b_hf.zip
      cd /code/llama2-finetuning
      python finetuning.py
    workdir: /code/llama2-finetuning
```