mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2024-12-24 10:24:35 +00:00
py : use cpu-only torch in requirements.txt (#8335)
This commit is contained in:
parent
b81ba1f96b
commit
d39130a398
@ -1,3 +1,4 @@
|
|||||||
-r ../../requirements/requirements-convert_legacy_llama.txt
|
-r ../../requirements/requirements-convert_legacy_llama.txt
|
||||||
|
--extra-index-url https://download.pytorch.org/whl/cpu
|
||||||
pillow~=10.2.0
|
pillow~=10.2.0
|
||||||
torch~=2.2.1
|
torch~=2.2.1
|
||||||
|
@ -1,2 +1,3 @@
|
|||||||
-r ./requirements-convert_legacy_llama.txt
|
-r ./requirements-convert_legacy_llama.txt
|
||||||
|
--extra-index-url https://download.pytorch.org/whl/cpu
|
||||||
torch~=2.2.1
|
torch~=2.2.1
|
||||||
|
@ -1,2 +1,3 @@
|
|||||||
-r ./requirements-convert_legacy_llama.txt
|
-r ./requirements-convert_legacy_llama.txt
|
||||||
|
--extra-index-url https://download.pytorch.org/whl/cpu
|
||||||
torch~=2.2.1
|
torch~=2.2.1
|
||||||
|
Loading…
Reference in New Issue
Block a user