Update README.md
Browse files
README.md
CHANGED
|
@@ -36,9 +36,6 @@ For test run results (and good indicator of target use cases), please see the fi
|
|
| 36 |
- **License:** Apache 2.0
|
| 37 |
- **Finetuned from model:** TinyLlama-1.1b - 2.5T checkpoint
|
| 38 |
|
| 39 |
-
## Uses
|
| 40 |
-
|
| 41 |
-
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
|
| 42 |
|
| 43 |
### Direct Use
|
| 44 |
|
|
@@ -60,7 +57,7 @@ Any model can provide inaccurate or incomplete information, and should be used i
|
|
| 60 |
|
| 61 |
## How to Get Started with the Model
|
| 62 |
|
| 63 |
-
The fastest way to get started with
|
| 64 |
|
| 65 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
| 66 |
tokenizer = AutoTokenizer.from_pretrained("bling-tiny-llama-v0", trust_remote_code=True)
|
|
@@ -68,7 +65,7 @@ The fastest way to get started with dRAGon is through direct import in transform
|
|
| 68 |
|
| 69 |
Please refer to the generation_test .py files in the Files repository, which includes 200 samples and script to test the model. The **generation_test_llmware_script.py** includes built-in llmware capabilities for fact-checking, as well as easy integration with document parsing and actual retrieval to swap out the test set for RAG workflow consisting of business documents.
|
| 70 |
|
| 71 |
-
The
|
| 72 |
|
| 73 |
full_prompt = "<human>: " + my_prompt + "\n" + "<bot>:"
|
| 74 |
|
|
|
|
| 36 |
- **License:** Apache 2.0
|
| 37 |
- **Finetuned from model:** TinyLlama-1.1b - 2.5T checkpoint
|
| 38 |
|
|
|
|
|
|
|
|
|
|
| 39 |
|
| 40 |
### Direct Use
|
| 41 |
|
|
|
|
| 57 |
|
| 58 |
## How to Get Started with the Model
|
| 59 |
|
| 60 |
+
The fastest way to get started with BLING is through direct import in transformers:
|
| 61 |
|
| 62 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
| 63 |
tokenizer = AutoTokenizer.from_pretrained("bling-tiny-llama-v0", trust_remote_code=True)
|
|
|
|
| 65 |
|
| 66 |
Please refer to the generation_test .py files in the Files repository, which includes 200 samples and script to test the model. The **generation_test_llmware_script.py** includes built-in llmware capabilities for fact-checking, as well as easy integration with document parsing and actual retrieval to swap out the test set for RAG workflow consisting of business documents.
|
| 67 |
|
| 68 |
+
The BLING model was fine-tuned with a simple "\<human> and \<bot> wrapper", so to get the best results, wrap inference entries as:
|
| 69 |
|
| 70 |
full_prompt = "<human>: " + my_prompt + "\n" + "<bot>:"
|
| 71 |
|