Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,35 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
# Model card metadata following Hugging Face specification:
|
3 |
+
# https://github.com/huggingface/hub-docs/blob/main/modelcard.md?plain=1
|
4 |
+
# Documentation: https://huggingface.co/docs/hub/model-cards
|
5 |
+
license: mit
|
6 |
+
tags:
|
7 |
+
- llama-cpp-python
|
8 |
+
- cuda
|
9 |
+
- nvidia
|
10 |
+
- blackwell
|
11 |
+
- windows
|
12 |
+
- prebuilt-wheels
|
13 |
+
- python
|
14 |
+
- machine-learning
|
15 |
+
- large-language-models
|
16 |
+
- gpu-acceleration
|
17 |
+
---
|
18 |
+
|
19 |
+
# llama-cpp-python 0.3.9 Prebuilt Wheel with CUDA Support for Windows
|
20 |
+
|
21 |
+
This repository provides a prebuilt Python wheel for **llama-cpp-python** (version 0.3.9) with NVIDIA CUDA support, optimized for Windows 10/11 (x64) systems. This wheel enables GPU-accelerated inference for large language models (LLMs) using the `llama.cpp` library, simplifying setup by eliminating the need to compile from source. The wheel is compatible with Python 3.10 and supports NVIDIA GPUs, including the latest Blackwell architecture.
|
22 |
+
|
23 |
+
## Available Wheel
|
24 |
+
- `llama_cpp_python-0.3.9-cp310-cp310-win_amd64.whl` (Python 3.10, CUDA 12.8)
|
25 |
+
|
26 |
+
## Compatibility
|
27 |
+
The prebuilt wheels are designed for NVIDIA Blackwell GPUs but have been tested and confirmed compatible with previous-generation NVIDIA GPUs, including:
|
28 |
+
- NVIDIA RTX 5090
|
29 |
+
- NVIDIA RTX 3090
|
30 |
+
|
31 |
+
## Installation
|
32 |
+
To install the wheel, use the following command in your Python 3.10 environment:
|
33 |
+
|
34 |
+
```bash
|
35 |
+
pip install llama_cpp_python-0.3.9-cp310-cp310-win_amd64.whl
|