| | --- |
| | tags: |
| | - llama |
| | - cot |
| | - vicuna |
| | - uncensored |
| | - merge |
| | - mix |
| | - gptq |
| | --- |
| | |
| | ## 13B-Chimera |
| |
|
| | ## Composition: |
| | [] = applied as LoRA to a composite model | () = combined as composite models |
| |
|
| | ((MantiCore3E+VicunaCocktail)+[SuperCOT+[StorytellingV2+(SuperHOTProtoType-8192ctx+Metharme)]]) |
| |
|
| | This model is the result of an experimental use of LoRAs on language models and model merges that are not the base HuggingFace-format LLaMA model they were intended for. |
| | The desired outcome is to additively apply desired features without paradoxically watering down a model's effective behavior. |
| |
|
| | Potential limitations - LoRAs applied on top of each other may intercompete. |
| |
|
| | Subjective results - very promising. Further experimental tests and objective tests are required. |
| |
|
| | Instruct and Setup Suggestions: |
| |
|
| | Alpaca instruct verified working, Vicuna instruct formats should work. |
| | If using KoboldAI or Text-Generation-WebUI, recommend switching between Godlike and Storywriter presets and adjusting output length + instructions in memory. |
| | Other presets as well as custom settings can yield highly different results, especially Temperature. |
| | If poking it with a stick doesn't work try another stick. |
| |
|
| | ## Language Models and LoRAs Used Credits: |
| |
|
| | manticore-13b [Epoch3] by openaccess-ai-collective |
| |
|
| | https://huggingface.co/openaccess-ai-collective/manticore-13b |
| |
|
| | vicuna-13b-cocktail by reeducator |
| |
|
| | https://huggingface.co/reeducator/vicuna-13b-cocktail |
| |
|
| | SuperCOT-LoRA [13B] by kaiokendev |
| |
|
| | https://huggingface.co/kaiokendev/SuperCOT-LoRA |
| |
|
| | Storytelling-LLaMa-LoRA [13B, Version 2] by GamerUnTouch |
| |
|
| | https://huggingface.co/GamerUntouch/Storytelling-LLaMa-LoRAs |
| |
|
| | SuperHOT Prototype [13b 8k ctx] by kaiokendev |
| |
|
| | https://huggingface.co/kaiokendev/SuperHOT-LoRA-prototype |
| |
|
| | Metharme 13b by PygmalionAI |
| |
|
| | https://huggingface.co/PygmalionAI/metharme-13b |
| |
|
| | Also thanks to Meta for LLaMA. |
| |
|
| | Each model and LoRA was hand picked and considered for what it could contribute to this ensemble. |
| | Thanks to each and every one of you for your incredible work developing some of the best things |
| | to come out of this community. |
| | # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard) |
| | Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_digitous__13B-Chimera) |
| |
|
| | | Metric | Value | |
| | |-----------------------|---------------------------| |
| | | Avg. | 52.19 | |
| | | ARC (25-shot) | 57.59 | |
| | | HellaSwag (10-shot) | 81.5 | |
| | | MMLU (5-shot) | 49.86 | |
| | | TruthfulQA (0-shot) | 52.59 | |
| | | Winogrande (5-shot) | 77.27 | |
| | | GSM8K (5-shot) | 10.69 | |
| | | DROP (3-shot) | 35.84 | |
| |
|