Update README.md
Browse files
README.md
CHANGED
|
@@ -19,6 +19,12 @@ or join our Deep
|
|
| 19 |
Learning Daily [Discord
|
| 20 |
community](https://discord.com/invite/p9ecgRhDR8/).
|
| 21 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 22 |
|
| 23 |
## **[DeciLM 6B](https://huggingface.co/Deci/DeciLM-6b):**
|
| 24 |
A permissively licensed, 5.7 billion-parameter pretrained text generation model using variable Grouped Query Attention (GQA) to achieve an optimal balance between performance and computational efficiency. Generated by Deci's proprietary Neural Architecture Search technology, AutoNAC™, DeciLM 6B delivers 15x the throughput of Llama 2 7B while maintaining comparable quality.
|
|
@@ -28,10 +34,6 @@ DeciLM-6B was fine-tuned using LoRA for instruction-following on a subset of the
|
|
| 28 |
A permissively licensed, text-to-image latent diffusion model generated by Deci's Neural Architecture Search technology.
|
| 29 |
DeciDiffusion generates Stable Diffusion-caliber images 3x faster.
|
| 30 |
|
| 31 |
-
## **[DeciCoder 1B](https://huggingface.co/Deci/DeciCoder-1b):**
|
| 32 |
-
A permissively licensed 1.1 billion-parameter code generation model generated by Deci's Neural Architecture Search technology.
|
| 33 |
-
Equipped with a 2048-context window, DeciCoder 1B delivers a 3.5x increase in throughput, improved accuracy on the HumanEval benchmark, and reduced memory usage compared to widely-used code generation LLMs such as SantaCoder.
|
| 34 |
-
|
| 35 |
## **[Infery-LLM](https://deci.ai/infery-llm-book-a-demo/):**
|
| 36 |
The most advanced inference SDK for LLM optimization and deployment, Infery-LLM includes unique features such as optimized kernels, continuous batching, advanced selective quantization, ultra-efficient beam search, parallel execution, and more.
|
| 37 |
To explore the full capabilities of Infery-LLM, we invite you to try it out [here](https://hubs.ly/Q02cz_pB0).
|
|
|
|
| 19 |
Learning Daily [Discord
|
| 20 |
community](https://discord.com/invite/p9ecgRhDR8/).
|
| 21 |
|
| 22 |
+
## **[DeciLM-7B](https://huggingface.co/Deci/DeciLM-7B):**
|
| 23 |
+
A 7.04 billion-parameter decoder-only text generation model, licensed under Apache 2.0. DeciLM-7B is not only the most accurate 7B base model to date, but it also currently outpaces all models in its class with a throughput that is up to 4.4x that of Mistral-7B's. DeciLM-7B’s architecture is the result of Deci's Neural Architecture Search technology. The model was fine-tuned using LoRA on the SlimOrca dataset, creating [DeciLM-7B-instruct](https://huggingface.co/Deci/DeciLM-7B-instruct)
|
| 24 |
+
|
| 25 |
+
## **[DeciCoder 1B](https://huggingface.co/Deci/DeciCoder-1b):**
|
| 26 |
+
A permissively licensed 1.1 billion-parameter code generation model generated by Deci's Neural Architecture Search technology.
|
| 27 |
+
Equipped with a 2048-context window, DeciCoder 1B delivers a 3.5x increase in throughput, improved accuracy on the HumanEval benchmark, and reduced memory usage compared to widely-used code generation LLMs such as SantaCoder.
|
| 28 |
|
| 29 |
## **[DeciLM 6B](https://huggingface.co/Deci/DeciLM-6b):**
|
| 30 |
A permissively licensed, 5.7 billion-parameter pretrained text generation model using variable Grouped Query Attention (GQA) to achieve an optimal balance between performance and computational efficiency. Generated by Deci's proprietary Neural Architecture Search technology, AutoNAC™, DeciLM 6B delivers 15x the throughput of Llama 2 7B while maintaining comparable quality.
|
|
|
|
| 34 |
A permissively licensed, text-to-image latent diffusion model generated by Deci's Neural Architecture Search technology.
|
| 35 |
DeciDiffusion generates Stable Diffusion-caliber images 3x faster.
|
| 36 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 37 |
## **[Infery-LLM](https://deci.ai/infery-llm-book-a-demo/):**
|
| 38 |
The most advanced inference SDK for LLM optimization and deployment, Infery-LLM includes unique features such as optimized kernels, continuous batching, advanced selective quantization, ultra-efficient beam search, parallel execution, and more.
|
| 39 |
To explore the full capabilities of Infery-LLM, we invite you to try it out [here](https://hubs.ly/Q02cz_pB0).
|