| | --- |
| | language: |
| | - en |
| | license: mit |
| | tags: |
| | - pretrained |
| | - security |
| | - redteam |
| | - blueteam |
| | pipeline_tag: text-generation |
| | inference: |
| | parameters: |
| | temperature: 0.7 |
| | extra_gated_description: >- |
| | If you want to learn more about how we process your personal data, please read |
| | our <a href="https://mistral.ai/terms/">Privacy Policy</a>. |
| | --- |
| | |
| | # TylerG01/Indigo-v0.1 |
| | Refer to the [original model card](https://huggingface.co/mistralai/Mistral-7B-v0.1) for more details on the model. |
| | ## Project Goals |
| | This is v0.1 (alpha) release of the Indigo LLM project, which used LoRA Fine-Tuning to train Mistral 7B on more than 400 books, pamphlets, |
| | training documents, code snippets and other works in the cyber security field, openly sourced on the surface web. This version used 16 LoRA layers |
| | and had a val loss of 1.601 after the 4th training epoch. However, my goal for the LoRA version of this model is to produce a val loss of <1.51 after |
| | some modification to the dataset and training approach. |
| |
|
| | For more information on this project, check out the blog post at https://t2-security.com/indigo-llm-503cd6e22fe4. |
| |
|