Upload README.md with huggingface_hub
Browse files
README.md
CHANGED
|
@@ -36,6 +36,7 @@ She is intended for **creative writing**, **roleplay**, **period-appropriate cor
|
|
| 36 |
|
| 37 |
## Known Issues / Limitations
|
| 38 |
- The 160M model series are very brittle in chat, but are very good at text completion in general.
|
|
|
|
| 39 |
- Ages and dates are unreliable (even within 1800–1899).
|
| 40 |
- Because parts of the corpus were derived from OCR, occasional stray modern tokens may appear (e.g., “http”, “Google”, “Internet Archive”).
|
| 41 |
- Training data includes UK and US English from the era.
|
|
@@ -53,14 +54,14 @@ Violet 160M was built on a corpus spanning 1800–1899 mostly sourced from Proje
|
|
| 53 |
This project began as an attempt to build a local LLM without relying on copyrighted training sources. The author also values local models that can run on a user’s machine without sending data to the cloud.
|
| 54 |
|
| 55 |
## Demo Resources
|
| 56 |
-
- HF Space: [Transformers.js Demo](
|
| 57 |
- CloudFlare Mirror: [Transformers.js Demo](https://pub-353f427e6227415cb077f3645638c125.r2.dev/index.html)
|
| 58 |
- Both of these are intended to use WebGPU and run local on your system -- No data is sent to the cloud.
|
| 59 |
|
| 60 |
## Related repos
|
| 61 |
-
- `Zakarth/violet-160m-chat` (
|
| 62 |
-
- `Zakarth/violet-1b4` (base/completion)
|
| 63 |
-
- `Zakarth/violet-1b4-chat
|
| 64 |
|
| 65 |
## Prompt Format
|
| 66 |
Use this structure:
|
|
|
|
| 36 |
|
| 37 |
## Known Issues / Limitations
|
| 38 |
- The 160M model series are very brittle in chat, but are very good at text completion in general.
|
| 39 |
+
- Punctuation for the 160M text completion model are variable at best albeit mostly conformant; this is likely due to variable differences seen from OCR sources. If used programmatically, some post-processing will almost certainly be required.
|
| 40 |
- Ages and dates are unreliable (even within 1800–1899).
|
| 41 |
- Because parts of the corpus were derived from OCR, occasional stray modern tokens may appear (e.g., “http”, “Google”, “Internet Archive”).
|
| 42 |
- Training data includes UK and US English from the era.
|
|
|
|
| 54 |
This project began as an attempt to build a local LLM without relying on copyrighted training sources. The author also values local models that can run on a user’s machine without sending data to the cloud.
|
| 55 |
|
| 56 |
## Demo Resources
|
| 57 |
+
- HF Space: [Transformers.js Demo](/spaces/zakarth/violetdemo)
|
| 58 |
- CloudFlare Mirror: [Transformers.js Demo](https://pub-353f427e6227415cb077f3645638c125.r2.dev/index.html)
|
| 59 |
- Both of these are intended to use WebGPU and run local on your system -- No data is sent to the cloud.
|
| 60 |
|
| 61 |
## Related repos
|
| 62 |
+
- `Zakarth/violet-160m-chat` (Fine tuned 160M model for chat. Brittle, but fun)
|
| 63 |
+
- `Zakarth/violet-1b4` (base/completion for 1.4B model)
|
| 64 |
+
- `Zakarth/violet-1b4-chat` (The most powerful "Violet" chat model)
|
| 65 |
|
| 66 |
## Prompt Format
|
| 67 |
Use this structure:
|