| license: apache-2.0 | |
| inference: false | |
| # Info | |
| This is the model [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) which I cut all the intermediate(feed_forward_length) size with 14336 down to 3072, resulting in a ~2.81B model. | |
| It's necessary to pre-train this model, cause at the moment is generating just gibberish. |