zakarth commited on
Commit
770e99b
·
verified ·
1 Parent(s): a119b97

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +5 -4
README.md CHANGED
@@ -36,6 +36,7 @@ She is intended for **creative writing**, **roleplay**, **period-appropriate cor
36
 
37
  ## Known Issues / Limitations
38
  - The 160M model series are very brittle in chat, but are very good at text completion in general.
 
39
  - Ages and dates are unreliable (even within 1800–1899).
40
  - Because parts of the corpus were derived from OCR, occasional stray modern tokens may appear (e.g., “http”, “Google”, “Internet Archive”).
41
  - Training data includes UK and US English from the era.
@@ -53,14 +54,14 @@ Violet 160M was built on a corpus spanning 1800–1899 mostly sourced from Proje
53
  This project began as an attempt to build a local LLM without relying on copyrighted training sources. The author also values local models that can run on a user’s machine without sending data to the cloud.
54
 
55
  ## Demo Resources
56
- - HF Space: [Transformers.js Demo](Zakarth/violetdemo)
57
  - CloudFlare Mirror: [Transformers.js Demo](https://pub-353f427e6227415cb077f3645638c125.r2.dev/index.html)
58
  - Both of these are intended to use WebGPU and run local on your system -- No data is sent to the cloud.
59
 
60
  ## Related repos
61
- - `Zakarth/violet-160m-chat` (base/completion for 160M series)
62
- - `Zakarth/violet-1b4` (base/completion)
63
- - `Zakarth/violet-1b4-chat-onnx` (WebGPU INT8)
64
 
65
  ## Prompt Format
66
  Use this structure:
 
36
 
37
  ## Known Issues / Limitations
38
  - The 160M model series are very brittle in chat, but are very good at text completion in general.
39
+ - Punctuation for the 160M text completion model are variable at best albeit mostly conformant; this is likely due to variable differences seen from OCR sources. If used programmatically, some post-processing will almost certainly be required.
40
  - Ages and dates are unreliable (even within 1800–1899).
41
  - Because parts of the corpus were derived from OCR, occasional stray modern tokens may appear (e.g., “http”, “Google”, “Internet Archive”).
42
  - Training data includes UK and US English from the era.
 
54
  This project began as an attempt to build a local LLM without relying on copyrighted training sources. The author also values local models that can run on a user’s machine without sending data to the cloud.
55
 
56
  ## Demo Resources
57
+ - HF Space: [Transformers.js Demo](/spaces/zakarth/violetdemo)
58
  - CloudFlare Mirror: [Transformers.js Demo](https://pub-353f427e6227415cb077f3645638c125.r2.dev/index.html)
59
  - Both of these are intended to use WebGPU and run local on your system -- No data is sent to the cloud.
60
 
61
  ## Related repos
62
+ - `Zakarth/violet-160m-chat` (Fine tuned 160M model for chat. Brittle, but fun)
63
+ - `Zakarth/violet-1b4` (base/completion for 1.4B model)
64
+ - `Zakarth/violet-1b4-chat` (The most powerful "Violet" chat model)
65
 
66
  ## Prompt Format
67
  Use this structure: