license: other
license_name: yi-license
license_link: LICENSE
widget:
- example_title: Yi-34B-Chat
text: hi
output:
text: ' Hello! How can I assist you today?'
- example_title: Yi-34B
text: >-
There's a place where time stands still. A place of breath taking wonder,
but also
output:
text: >2-
an eerie sense that something is just not right…
Between the two worlds lies The Forgotten Kingdom - home to creatures
long since thought extinct and ancient magic so strong it defies belief!
Only here can you find what has been lost for centuries: An Elixir Of
Life which will restore youth and vitality if only those who seek its
power are brave enough to face up against all manner of dangers lurking
in this mysterious land! But beware; some say there may even exist
powerful entities beyond our comprehension whose intentions towards
humanity remain unclear at best ---- they might want nothing more than
destruction itself rather then anything else from their quest after
immortality (and maybe someone should tell them about modern medicine)?
In any event though – one thing remains true regardless : whether or not
success comes easy depends entirely upon how much effort we put into
conquering whatever challenges lie ahead along with having faith deep
down inside ourselves too ;) So let’s get started now shall We?
pipeline_tag: text-generation
quantized_by: bartowski
Exllama v2 Quantizations of Yi-9B-200K
Using turboderp's ExLlamaV2 v0.0.15 for quantization.
The "main" branch only contains the measurement.json, download one of the other branches for the model (see below)
Each branch contains an individual bits per weight, with the main one containing only the meaurement.json for further conversions.
Conversion was done using the default calibration dataset.
Default arguments used except when the bits per weight is above 6.0, at that point the lm_head layer is quantized at 8 bits per weight instead of the default 6.
Original model: https://huggingface.co/01-ai/Yi-9B-200K
Download instructions
With git:
git clone --single-branch --branch 6_5 https://huggingface.co/bartowski/Yi-9B-200K-exl2
With huggingface hub (credit to TheBloke for instructions):
pip3 install huggingface-hub
To download the main (only useful if you only care about measurement.json) branch to a folder called Yi-9B-200K-exl2:
mkdir Yi-9B-200K-exl2
huggingface-cli download bartowski/Yi-9B-200K-exl2 --local-dir Yi-9B-200K-exl2 --local-dir-use-symlinks False
To download from a different branch, add the --revision parameter:
Linux:
mkdir Yi-9B-200K-exl2-6_5
huggingface-cli download bartowski/Yi-9B-200K-exl2 --revision 6_5 --local-dir Yi-9B-200K-exl2-6_5 --local-dir-use-symlinks False
Windows (which apparently doesn't like _ in folders sometimes?):
mkdir Yi-9B-200K-exl2-6.5
huggingface-cli download bartowski/Yi-9B-200K-exl2 --revision 6_5 --local-dir Yi-9B-200K-exl2-6.5 --local-dir-use-symlinks False