Commit ·
486cf71
1
Parent(s): f50f250
Update README.md
Browse files
README.md
CHANGED
|
@@ -14,6 +14,18 @@ Eluwa is a fine-tuned Low-Rank Adapter (LoRA) model for Facebook's OPT 2.7b. It
|
|
| 14 |
The idea was that OPT 2.7 was too curt (and frankly, a bit of an asshole) for a model of its size, and that we could finetune it like Alpaca did to Llama.
|
| 15 |
|
| 16 |
This repository contains the Eluwa 2.7b 2 epoch model, which represents a significant improvements in question-answering ability compared to the default OPT 2.7b model.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 17 |
Below are the results of Vicuna-style testing: 80 questions in various categories, with the responses rated by GPT-4.
|
| 18 |
|
| 19 |
| Model | OPT 2.7b base | Eluwa 2.7b 1000 iter | Eluwa 2.7b 2 epoch |
|
|
@@ -32,4 +44,6 @@ Below are the results of Vicuna-style testing: 80 questions in various categorie
|
|
| 32 |
(A sheet of questions, answers and GPT's reviews are also included in this repo).
|
| 33 |
|
| 34 |
Because of its small size, Eluwa can be used as research into conversational models with older and slower hardware. To load it in a UI like oobabooga,
|
| 35 |
-
download the model's .bin and .json files, put them in a folder inside the /loras folder, and load it with the OPT 2.7b model.
|
|
|
|
|
|
|
|
|
| 14 |
The idea was that OPT 2.7 was too curt (and frankly, a bit of an asshole) for a model of its size, and that we could finetune it like Alpaca did to Llama.
|
| 15 |
|
| 16 |
This repository contains the Eluwa 2.7b 2 epoch model, which represents a significant improvements in question-answering ability compared to the default OPT 2.7b model.
|
| 17 |
+
Despite Eluwa being not too different from OPT 2.7b, it often does a lot better than default OPT2.7b at accuracy and coherency.
|
| 18 |
+
|
| 19 |
+
For example, when asked "What is the capital of Sri Lanka", OPT 2.7b generates nonsense:
|
| 20 |
+
```
|
| 21 |
+
A: Sri Lanka is the country where the sun rises every morning on a beautiful beach. It has many beaches, but only one city - Colombo. (The sun sets at night.)
|
| 22 |
+
```
|
| 23 |
+
While Eluwa correctly says,
|
| 24 |
+
```
|
| 25 |
+
A: Colombo.
|
| 26 |
+
```
|
| 27 |
+
|
| 28 |
+
Likewise, when asked how to become a data scientist, Eluwa tries to be useful, whereas OPT 2.7B ends up insulting the user.
|
| 29 |
Below are the results of Vicuna-style testing: 80 questions in various categories, with the responses rated by GPT-4.
|
| 30 |
|
| 31 |
| Model | OPT 2.7b base | Eluwa 2.7b 1000 iter | Eluwa 2.7b 2 epoch |
|
|
|
|
| 44 |
(A sheet of questions, answers and GPT's reviews are also included in this repo).
|
| 45 |
|
| 46 |
Because of its small size, Eluwa can be used as research into conversational models with older and slower hardware. To load it in a UI like oobabooga,
|
| 47 |
+
download the model's .bin and .json files, put them in a folder inside the /loras folder, and load it with the OPT 2.7b model.
|
| 48 |
+
|
| 49 |
+
|