| Quantization made by Richard Erkhov. | |
| [Github](https://github.com/RichardErkhov) | |
| [Discord](https://discord.gg/pvy7H8DZMG) | |
| [Request more models](https://github.com/RichardErkhov/quant_request) | |
| Magicoder-S-DS-6.7B - bnb 8bits | |
| - Model creator: https://huggingface.co/ise-uiuc/ | |
| - Original model: https://huggingface.co/ise-uiuc/Magicoder-S-DS-6.7B/ | |
| Original model description: | |
| --- | |
| license: other | |
| library_name: transformers | |
| datasets: | |
| - ise-uiuc/Magicoder-OSS-Instruct-75K | |
| - ise-uiuc/Magicoder-Evol-Instruct-110K | |
| license_name: deepseek | |
| pipeline_tag: text-generation | |
| --- | |
| # 🎩 Magicoder: Source Code Is All You Need | |
| > Refer to our GitHub repo [ise-uiuc/magicoder](https://github.com/ise-uiuc/magicoder/) for an up-to-date introduction to the Magicoder family! | |
| * 🎩**Magicoder** is a model family empowered by 🪄**OSS-Instruct**, a novel approach to enlightening LLMs with open-source code snippets for generating *low-bias* and *high-quality* instruction data for code. | |
| * 🪄**OSS-Instruct** mitigates the *inherent bias* of the LLM-synthesized instruction data by empowering them with *a wealth of open-source references* to produce more diverse, realistic, and controllable data. | |
|  | |
|  | |
| ## Model Details | |
| ### Model Description | |
| * **Developed by:** | |
| [Yuxiang Wei](https://yuxiang.cs.illinois.edu), | |
| [Zhe Wang](https://github.com/zhewang2001), | |
| [Jiawei Liu](https://jiawei-site.github.io), | |
| [Yifeng Ding](https://yifeng-ding.com), | |
| [Lingming Zhang](https://lingming.cs.illinois.edu) | |
| * **License:** [DeepSeek](https://github.com/deepseek-ai/DeepSeek-Coder/blob/main/LICENSE-MODEL) | |
| * **Finetuned from model:** [deepseek-coder-6.7b-base](https://huggingface.co/deepseek-ai/deepseek-coder-6.7b-base) | |
| ### Model Sources | |
| * **Repository:** <https://github.com/ise-uiuc/magicoder> | |
| * **Paper:** <https://arxiv.org/abs/2312.02120> | |
| * **Demo (powered by [Gradio](https://www.gradio.app)):** | |
| <https://github.com/ise-uiuc/magicoder/tree/main/demo> | |
| ### Training Data | |
| * [Magicoder-OSS-Instruct-75K](https://huggingface.co/datasets/ise-uiuc/Magicoder_oss_instruct_75k): generated through **OSS-Instruct** using `gpt-3.5-turbo-1106` and used to train both Magicoder and Magicoder-S series. | |
| * [Magicoder-Evol-Instruct-110K](https://huggingface.co/datasets/ise-uiuc/Magicoder_evol_instruct_110k): decontaminated and redistributed from [theblackcat102/evol-codealpaca-v1](https://huggingface.co/datasets/theblackcat102/evol-codealpaca-v1), used to further finetune Magicoder series and obtain Magicoder-S models. | |
| ## Uses | |
| ### Direct Use | |
| Magicoders are designed and best suited for **coding tasks**. | |
| ### Out-of-Scope Use | |
| Magicoders may not work well in non-coding tasks. | |
| ## Bias, Risks, and Limitations | |
| Magicoders may sometimes make errors, producing misleading contents, or struggle to manage tasks that are not related to coding. | |
| ### Recommendations | |
| Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. | |
| ## How to Get Started with the Model | |
| Use the code below to get started with the model. Make sure you installed the [transformers](https://huggingface.co/docs/transformers/index) library. | |
| ```python | |
| from transformers import pipeline | |
| import torch | |
| MAGICODER_PROMPT = """You are an exceptionally intelligent coding assistant that consistently delivers accurate and reliable responses to user instructions. | |
| @@ Instruction | |
| {instruction} | |
| @@ Response | |
| """ | |
| instruction = <Your code instruction here> | |
| prompt = MAGICODER_PROMPT.format(instruction=instruction) | |
| generator = pipeline( | |
| model="ise-uiuc/Magicoder-S-DS-6.7B", | |
| task="text-generation", | |
| torch_dtype=torch.bfloat16, | |
| device_map="auto", | |
| ) | |
| result = generator(prompt, max_length=1024, num_return_sequences=1, temperature=0.0) | |
| print(result[0]["generated_text"]) | |
| ``` | |
| ## Technical Details | |
| Refer to our GitHub repo: [ise-uiuc/magicoder](https://github.com/ise-uiuc/magicoder/). | |
| ## Citation | |
| ```bibtex | |
| @misc{magicoder, | |
| title={Magicoder: Source Code Is All You Need}, | |
| author={Yuxiang Wei and Zhe Wang and Jiawei Liu and Yifeng Ding and Lingming Zhang}, | |
| year={2023}, | |
| eprint={2312.02120}, | |
| archivePrefix={arXiv}, | |
| primaryClass={cs.CL} | |
| } | |
| ``` | |
| ## Acknowledgements | |
| * [WizardCoder](https://github.com/nlpxucan/WizardLM/tree/main/WizardCoder): Evol-Instruct | |
| * [DeepSeek-Coder](https://github.com/deepseek-ai/DeepSeek-Coder): Base model for Magicoder-DS | |
| * [CodeLlama](https://ai.meta.com/research/publications/code-llama-open-foundation-models-for-code/): Base model for Magicoder-CL | |
| * [StarCoder](https://arxiv.org/abs/2305.06161): Data decontamination | |
| ## Important Note | |
| Magicoder models are trained on the synthetic data generated by OpenAI models. Please pay attention to OpenAI's [terms of use](https://openai.com/policies/terms-of-use) when using the models and the datasets. Magicoders will not compete with OpenAI's commercial products. | |