Yi-9B-200K-exl2 / README.md
bartowski's picture
measurement.json
78df73f verified
metadata
license: other
license_name: yi-license
license_link: LICENSE
widget:
  - example_title: Yi-34B-Chat
    text: hi
    output:
      text: ' Hello! How can I assist you today?'
  - example_title: Yi-34B
    text: >-
      There's a place where time stands still. A place of breath taking wonder,
      but also
    output:
      text: >2-
         an eerie sense that something is just not right…
        Between the two worlds lies The Forgotten Kingdom - home to creatures
        long since thought extinct and ancient magic so strong it defies belief!
        Only here can you find what has been lost for centuries: An Elixir Of
        Life which will restore youth and vitality if only those who seek its
        power are brave enough to face up against all manner of dangers lurking
        in this mysterious land! But beware; some say there may even exist
        powerful entities beyond our comprehension whose intentions towards
        humanity remain unclear at best ---- they might want nothing more than
        destruction itself rather then anything else from their quest after
        immortality (and maybe someone should tell them about modern medicine)?
        In any event though  one thing remains true regardless : whether or not
        success comes easy depends entirely upon how much effort we put into
        conquering whatever challenges lie ahead along with having faith deep
        down inside ourselves too ;) So let’s get started now shall We?
pipeline_tag: text-generation
quantized_by: bartowski

Exllama v2 Quantizations of Yi-9B-200K

Using turboderp's ExLlamaV2 v0.0.15 for quantization.

The "main" branch only contains the measurement.json, download one of the other branches for the model (see below)

Each branch contains an individual bits per weight, with the main one containing only the meaurement.json for further conversions.

Conversion was done using the default calibration dataset.

Default arguments used except when the bits per weight is above 6.0, at that point the lm_head layer is quantized at 8 bits per weight instead of the default 6.

Original model: https://huggingface.co/01-ai/Yi-9B-200K

8.0 bits per weight

6.5 bits per weight

5.0 bits per weight

4.25 bits per weight

3.5 bits per weight

Download instructions

With git:

git clone --single-branch --branch 6_5 https://huggingface.co/bartowski/Yi-9B-200K-exl2

With huggingface hub (credit to TheBloke for instructions):

pip3 install huggingface-hub

To download the main (only useful if you only care about measurement.json) branch to a folder called Yi-9B-200K-exl2:

mkdir Yi-9B-200K-exl2
huggingface-cli download bartowski/Yi-9B-200K-exl2 --local-dir Yi-9B-200K-exl2 --local-dir-use-symlinks False

To download from a different branch, add the --revision parameter:

Linux:

mkdir Yi-9B-200K-exl2-6_5
huggingface-cli download bartowski/Yi-9B-200K-exl2 --revision 6_5 --local-dir Yi-9B-200K-exl2-6_5 --local-dir-use-symlinks False

Windows (which apparently doesn't like _ in folders sometimes?):

mkdir Yi-9B-200K-exl2-6.5
huggingface-cli download bartowski/Yi-9B-200K-exl2 --revision 6_5 --local-dir Yi-9B-200K-exl2-6.5 --local-dir-use-symlinks False