YAML Metadata Warning:empty or missing yaml metadata in repo card

Check out the documentation for more information.

Quantization made by Richard Erkhov.

Github

Discord

Request more models

Cucumber-7b-10k - bnb 4bits

Original model description:

license: cc-by-4.0 datasets: - HuggingFaceH4/no_robots - Ba2han/Reddit-instruct-curated_rated-1.2k - Ba2han/databricks-dolly_rated - Open-Orca/OpenOrca

image/png The training dataset consists of 2k (longest) examples from no_robots, reddit_instruct, dolly, OpenOrca plus two other personal datasets.

Please use with ChatML and the default system message or enter your own. It was trained with various system messages, the one in the config being the default one.

The model is:

  • Very good at generating long and coherent text.

  • Creative due to data from Reddit ELI5 and a few other sources.

  • Better at handling longer input.

  • Not great with short text both in input and generation.

The aim is to see how the "Long is More for Alignment" paper holds. This is basically a combination of LIMA + LMA. There should be no benchmark contamination as far as I am aware of. Around 70% of the data is from the mentioned datasets. I am happy with how it turned out.

image/png

Downloads last month
-
Safetensors
Model size
7B params
Tensor type
F32
F16
U8
Inference Providers NEW
This model isn't deployed by any Inference Provider. 馃檵 Ask for provider support