Spaces:
No application file
No application file
Update README.md
Browse files
README.md
CHANGED
@@ -20,7 +20,7 @@ We believe the future of AI is open. That’s why we’re sharing our latest mod
|
|
20 |
🔧 **With Red Hat AI, you can:**
|
21 |
- **Use or build optimized foundation models**, including Llama, Mistral, Qwen, Gemma, DeepSeek, and others, tailored for performance and accuracy in real-world deployments.
|
22 |
- **Customize and fine-tune models for your workflows**, from experimentation to production, with tools and frameworks built to support reproducible research and enterprise AI pipelines.
|
23 |
-
- **Maximize inference efficiency across hardware** using production-grade compression and optimization techniques like quantization (FP8, INT8, INT4), structured/unstructured sparsity, distillation, and more, ready for cost-efficient deployments with vLLM.
|
24 |
- **[**Validated models**](http://www.redhat.com/en/products/ai/validated-models) by Red Hat AI offer confidence, predictability, and flexibility when deploying third-party generative AI models across the Red Hat AI platform.** Red Hat AI validates models by running a series of capacity planning scenarios with [GuideLLM](https://github.com/neuralmagic/guidellm) for benchmarking, [Language Model Evaluation Harness](https://github.com/EleutherAI/lm-evaluation-harness) for accuracy evaluations, and [vLLM](https://github.com/vllm-project/vllm) for inference serving across a wide variety of AI acclerators.
|
25 |
|
26 |
🔗 **Explore relevant open-source tools**:
|
|
|
20 |
🔧 **With Red Hat AI, you can:**
|
21 |
- **Use or build optimized foundation models**, including Llama, Mistral, Qwen, Gemma, DeepSeek, and others, tailored for performance and accuracy in real-world deployments.
|
22 |
- **Customize and fine-tune models for your workflows**, from experimentation to production, with tools and frameworks built to support reproducible research and enterprise AI pipelines.
|
23 |
+
- **Maximize inference efficiency across hardware** using production-grade compression and optimization techniques like quantization (FP8-dynamic, INT8, INT4), structured/unstructured sparsity, distillation, and more, ready for cost-efficient deployments with vLLM.
|
24 |
- **[**Validated models**](http://www.redhat.com/en/products/ai/validated-models) by Red Hat AI offer confidence, predictability, and flexibility when deploying third-party generative AI models across the Red Hat AI platform.** Red Hat AI validates models by running a series of capacity planning scenarios with [GuideLLM](https://github.com/neuralmagic/guidellm) for benchmarking, [Language Model Evaluation Harness](https://github.com/EleutherAI/lm-evaluation-harness) for accuracy evaluations, and [vLLM](https://github.com/vllm-project/vllm) for inference serving across a wide variety of AI acclerators.
|
25 |
|
26 |
🔗 **Explore relevant open-source tools**:
|