use transformers from git
Browse files- Dockerfile +1 -2
- server/entrypoint.sh +2 -2
- server/pdm.lock +4 -22
- server/pyproject.toml +3 -1
Dockerfile
CHANGED
|
@@ -31,7 +31,6 @@ ENV PDM_CACHE_DIR $HOME/code/.pdm_cache/
|
|
| 31 |
ENV HF_HOME $HOME/code/.hf_home/
|
| 32 |
|
| 33 |
ENV DEVICE cuda:0
|
| 34 |
-
|
| 35 |
-
ENV ATTN_IMPLEMENTATION flash_attention
|
| 36 |
|
| 37 |
ENTRYPOINT [ "./entrypoint.sh" ]
|
|
|
|
| 31 |
ENV HF_HOME $HOME/code/.hf_home/
|
| 32 |
|
| 33 |
ENV DEVICE cuda:0
|
| 34 |
+
ENV ATTN_IMPLEMENTATION sdpa
|
|
|
|
| 35 |
|
| 36 |
ENTRYPOINT [ "./entrypoint.sh" ]
|
server/entrypoint.sh
CHANGED
|
@@ -1,7 +1,7 @@
|
|
| 1 |
#!/bin/bash
|
| 2 |
|
| 3 |
-
curl -sS https://bootstrap.pypa.io/get-pip.py | ./.venv/bin/python
|
| 4 |
-
./.venv/bin/python -m pip install flash_attn==1.0.9 --no-build-isolation
|
| 5 |
|
| 6 |
echo "Installation done"
|
| 7 |
|
|
|
|
| 1 |
#!/bin/bash
|
| 2 |
|
| 3 |
+
# curl -sS https://bootstrap.pypa.io/get-pip.py | ./.venv/bin/python
|
| 4 |
+
# ./.venv/bin/python -m pip install flash_attn==1.0.9 --no-build-isolation
|
| 5 |
|
| 6 |
echo "Installation done"
|
| 7 |
|
server/pdm.lock
CHANGED
|
@@ -5,7 +5,7 @@
|
|
| 5 |
groups = ["default"]
|
| 6 |
strategy = ["cross_platform"]
|
| 7 |
lock_version = "4.4"
|
| 8 |
-
content_hash = "sha256:
|
| 9 |
|
| 10 |
[[package]]
|
| 11 |
name = "accelerate"
|
|
@@ -859,8 +859,10 @@ files = [
|
|
| 859 |
|
| 860 |
[[package]]
|
| 861 |
name = "transformers"
|
| 862 |
-
version = "4.
|
| 863 |
requires_python = ">=3.8.0"
|
|
|
|
|
|
|
| 864 |
summary = "State-of-the-art Machine Learning for JAX, PyTorch and TensorFlow"
|
| 865 |
dependencies = [
|
| 866 |
"filelock",
|
|
@@ -874,26 +876,6 @@ dependencies = [
|
|
| 874 |
"tokenizers<0.19,>=0.14",
|
| 875 |
"tqdm>=4.27",
|
| 876 |
]
|
| 877 |
-
files = [
|
| 878 |
-
{file = "transformers-4.36.2-py3-none-any.whl", hash = "sha256:462066c4f74ee52516f12890dcc9ec71d1a5e97998db621668455117a54330f6"},
|
| 879 |
-
{file = "transformers-4.36.2.tar.gz", hash = "sha256:d8068e897e47793281501e547d2bbdfc5b8556409c2cb6c3d9e2ca77d4c0b4ec"},
|
| 880 |
-
]
|
| 881 |
-
|
| 882 |
-
[[package]]
|
| 883 |
-
name = "transformers"
|
| 884 |
-
version = "4.36.2"
|
| 885 |
-
extras = ["torch"]
|
| 886 |
-
requires_python = ">=3.8.0"
|
| 887 |
-
summary = "State-of-the-art Machine Learning for JAX, PyTorch and TensorFlow"
|
| 888 |
-
dependencies = [
|
| 889 |
-
"accelerate>=0.21.0",
|
| 890 |
-
"torch!=1.12.0,>=1.10",
|
| 891 |
-
"transformers==4.36.2",
|
| 892 |
-
]
|
| 893 |
-
files = [
|
| 894 |
-
{file = "transformers-4.36.2-py3-none-any.whl", hash = "sha256:462066c4f74ee52516f12890dcc9ec71d1a5e97998db621668455117a54330f6"},
|
| 895 |
-
{file = "transformers-4.36.2.tar.gz", hash = "sha256:d8068e897e47793281501e547d2bbdfc5b8556409c2cb6c3d9e2ca77d4c0b4ec"},
|
| 896 |
-
]
|
| 897 |
|
| 898 |
[[package]]
|
| 899 |
name = "triton"
|
|
|
|
| 5 |
groups = ["default"]
|
| 6 |
strategy = ["cross_platform"]
|
| 7 |
lock_version = "4.4"
|
| 8 |
+
content_hash = "sha256:a7e1894df31010cec5c072c8c7c796e2124f68acffe168327b11e80d5ab0f775"
|
| 9 |
|
| 10 |
[[package]]
|
| 11 |
name = "accelerate"
|
|
|
|
| 859 |
|
| 860 |
[[package]]
|
| 861 |
name = "transformers"
|
| 862 |
+
version = "4.37.0.dev0"
|
| 863 |
requires_python = ">=3.8.0"
|
| 864 |
+
git = "https://github.com/huggingface/transformers.git"
|
| 865 |
+
revision = "772307be7649e1333a933cfaa229dc0dec2fd331"
|
| 866 |
summary = "State-of-the-art Machine Learning for JAX, PyTorch and TensorFlow"
|
| 867 |
dependencies = [
|
| 868 |
"filelock",
|
|
|
|
| 876 |
"tokenizers<0.19,>=0.14",
|
| 877 |
"tqdm>=4.27",
|
| 878 |
]
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 879 |
|
| 880 |
[[package]]
|
| 881 |
name = "triton"
|
server/pyproject.toml
CHANGED
|
@@ -9,9 +9,11 @@ dependencies = [
|
|
| 9 |
"fastapi>=0.109.0",
|
| 10 |
"uvicorn[standard]>=0.25.0",
|
| 11 |
"numpy>=1.26.3",
|
| 12 |
-
"transformers
|
| 13 |
"ninja>=1.11.1.1",
|
| 14 |
"packaging>=23.2",
|
|
|
|
|
|
|
| 15 |
]
|
| 16 |
requires-python = ">=3.11"
|
| 17 |
readme = "README.md"
|
|
|
|
| 9 |
"fastapi>=0.109.0",
|
| 10 |
"uvicorn[standard]>=0.25.0",
|
| 11 |
"numpy>=1.26.3",
|
| 12 |
+
"transformers @ git+https://github.com/huggingface/transformers.git",
|
| 13 |
"ninja>=1.11.1.1",
|
| 14 |
"packaging>=23.2",
|
| 15 |
+
"torch>=2.1.2",
|
| 16 |
+
"accelerate>=0.26.1",
|
| 17 |
]
|
| 18 |
requires-python = ">=3.11"
|
| 19 |
readme = "README.md"
|