danaevan commited on
Commit
cb8f0b5
·
1 Parent(s): c1a8b5e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -4
README.md CHANGED
@@ -19,6 +19,12 @@ or join our Deep
19
  Learning Daily [Discord
20
  community](https://discord.com/invite/p9ecgRhDR8/).
21
 
 
 
 
 
 
 
22
 
23
  ## **[DeciLM 6B](https://huggingface.co/Deci/DeciLM-6b):**
24
  A permissively licensed, 5.7 billion-parameter pretrained text generation model using variable Grouped Query Attention (GQA) to achieve an optimal balance between performance and computational efficiency. Generated by Deci's proprietary Neural Architecture Search technology, AutoNAC™, DeciLM 6B delivers 15x the throughput of Llama 2 7B while maintaining comparable quality.
@@ -28,10 +34,6 @@ DeciLM-6B was fine-tuned using LoRA for instruction-following on a subset of the
28
  A permissively licensed, text-to-image latent diffusion model generated by Deci's Neural Architecture Search technology.
29
  DeciDiffusion generates Stable Diffusion-caliber images 3x faster.
30
 
31
- ## **[DeciCoder 1B](https://huggingface.co/Deci/DeciCoder-1b):**
32
- A permissively licensed 1.1 billion-parameter code generation model generated by Deci's Neural Architecture Search technology.
33
- Equipped with a 2048-context window, DeciCoder 1B delivers a 3.5x increase in throughput, improved accuracy on the HumanEval benchmark, and reduced memory usage compared to widely-used code generation LLMs such as SantaCoder.
34
-
35
  ## **[Infery-LLM](https://deci.ai/infery-llm-book-a-demo/):**
36
  The most advanced inference SDK for LLM optimization and deployment, Infery-LLM includes unique features such as optimized kernels, continuous batching, advanced selective quantization, ultra-efficient beam search, parallel execution, and more.
37
  To explore the full capabilities of Infery-LLM, we invite you to try it out [here](https://hubs.ly/Q02cz_pB0).
 
19
  Learning Daily [Discord
20
  community](https://discord.com/invite/p9ecgRhDR8/).
21
 
22
+ ## **[DeciLM-7B](https://huggingface.co/Deci/DeciLM-7B):**
23
+ A 7.04 billion-parameter decoder-only text generation model, licensed under Apache 2.0. DeciLM-7B is not only the most accurate 7B base model to date, but it also currently outpaces all models in its class with a throughput that is up to 4.4x that of Mistral-7B's. DeciLM-7B’s architecture is the result of Deci's Neural Architecture Search technology. The model was fine-tuned using LoRA on the SlimOrca dataset, creating [DeciLM-7B-instruct](https://huggingface.co/Deci/DeciLM-7B-instruct)
24
+
25
+ ## **[DeciCoder 1B](https://huggingface.co/Deci/DeciCoder-1b):**
26
+ A permissively licensed 1.1 billion-parameter code generation model generated by Deci's Neural Architecture Search technology.
27
+ Equipped with a 2048-context window, DeciCoder 1B delivers a 3.5x increase in throughput, improved accuracy on the HumanEval benchmark, and reduced memory usage compared to widely-used code generation LLMs such as SantaCoder.
28
 
29
  ## **[DeciLM 6B](https://huggingface.co/Deci/DeciLM-6b):**
30
  A permissively licensed, 5.7 billion-parameter pretrained text generation model using variable Grouped Query Attention (GQA) to achieve an optimal balance between performance and computational efficiency. Generated by Deci's proprietary Neural Architecture Search technology, AutoNAC™, DeciLM 6B delivers 15x the throughput of Llama 2 7B while maintaining comparable quality.
 
34
  A permissively licensed, text-to-image latent diffusion model generated by Deci's Neural Architecture Search technology.
35
  DeciDiffusion generates Stable Diffusion-caliber images 3x faster.
36
 
 
 
 
 
37
  ## **[Infery-LLM](https://deci.ai/infery-llm-book-a-demo/):**
38
  The most advanced inference SDK for LLM optimization and deployment, Infery-LLM includes unique features such as optimized kernels, continuous batching, advanced selective quantization, ultra-efficient beam search, parallel execution, and more.
39
  To explore the full capabilities of Infery-LLM, we invite you to try it out [here](https://hubs.ly/Q02cz_pB0).