Sync from upstream llama.cpp repository
This commit is contained in:
18
requirements/requirements-all.txt
Normal file
18
requirements/requirements-all.txt
Normal file
@@ -0,0 +1,18 @@
|
||||
-r ../tools/mtmd/requirements.txt
|
||||
-r ../tools/server/bench/requirements.txt
|
||||
-r ../tools/server/tests/requirements.txt
|
||||
|
||||
-r ./requirements-compare-llama-bench.txt
|
||||
-r ./requirements-server-bench.txt
|
||||
-r ./requirements-pydantic.txt
|
||||
-r ./requirements-test-tokenizer-random.txt
|
||||
|
||||
-r ./requirements-convert_hf_to_gguf.txt
|
||||
-r ./requirements-convert_hf_to_gguf_update.txt
|
||||
-r ./requirements-convert_legacy_llama.txt
|
||||
-r ./requirements-convert_llama_ggml_to_gguf.txt
|
||||
-r ./requirements-tool_bench.txt
|
||||
|
||||
-r ./requirements-gguf_editor_gui.txt
|
||||
|
||||
-r ../examples/model-conversion/requirements.txt
|
||||
3
requirements/requirements-compare-llama-bench.txt
Normal file
3
requirements/requirements-compare-llama-bench.txt
Normal file
@@ -0,0 +1,3 @@
|
||||
tabulate~=0.9.0
|
||||
GitPython~=3.1.43
|
||||
matplotlib~=3.10.0
|
||||
9
requirements/requirements-convert_hf_to_gguf.txt
Normal file
9
requirements/requirements-convert_hf_to_gguf.txt
Normal file
@@ -0,0 +1,9 @@
|
||||
-r ./requirements-convert_legacy_llama.txt
|
||||
--extra-index-url https://download.pytorch.org/whl/cpu
|
||||
|
||||
## Embedding Gemma requires PyTorch 2.6.0 or later
|
||||
torch~=2.6.0; platform_machine != "s390x"
|
||||
|
||||
# torch s390x packages can only be found from nightly builds
|
||||
--extra-index-url https://download.pytorch.org/whl/nightly
|
||||
torch>=0.0.0.dev0; platform_machine == "s390x"
|
||||
1
requirements/requirements-convert_hf_to_gguf_update.txt
Normal file
1
requirements/requirements-convert_hf_to_gguf_update.txt
Normal file
@@ -0,0 +1 @@
|
||||
-r ./requirements-convert_legacy_llama.txt
|
||||
7
requirements/requirements-convert_legacy_llama.txt
Normal file
7
requirements/requirements-convert_legacy_llama.txt
Normal file
@@ -0,0 +1,7 @@
|
||||
numpy~=1.26.4
|
||||
sentencepiece~=0.2.0
|
||||
|
||||
transformers>=4.57.1,<5.0.0
|
||||
|
||||
gguf>=0.1.0
|
||||
protobuf>=4.21.0,<5.0.0
|
||||
1
requirements/requirements-convert_llama_ggml_to_gguf.txt
Normal file
1
requirements/requirements-convert_llama_ggml_to_gguf.txt
Normal file
@@ -0,0 +1 @@
|
||||
-r ./requirements-convert_legacy_llama.txt
|
||||
4
requirements/requirements-convert_lora_to_gguf.txt
Normal file
4
requirements/requirements-convert_lora_to_gguf.txt
Normal file
@@ -0,0 +1,4 @@
|
||||
-r ./requirements-convert_hf_to_gguf.txt
|
||||
--extra-index-url https://download.pytorch.org/whl/cpu
|
||||
# torch s390x packages can only be found from nightly builds
|
||||
--extra-index-url https://download.pytorch.org/whl/nightly
|
||||
3
requirements/requirements-gguf_editor_gui.txt
Normal file
3
requirements/requirements-gguf_editor_gui.txt
Normal file
@@ -0,0 +1,3 @@
|
||||
numpy~=1.26.4
|
||||
PySide6~=6.9.0
|
||||
gguf>=0.17.0
|
||||
3
requirements/requirements-pydantic.txt
Normal file
3
requirements/requirements-pydantic.txt
Normal file
@@ -0,0 +1,3 @@
|
||||
docstring_parser~=0.15
|
||||
pydantic~=2.11.7
|
||||
requests
|
||||
5
requirements/requirements-server-bench.txt
Normal file
5
requirements/requirements-server-bench.txt
Normal file
@@ -0,0 +1,5 @@
|
||||
datasets~=3.2.0
|
||||
matplotlib~=3.10.0
|
||||
numpy~=1.26.4
|
||||
requests~=2.32.3
|
||||
tqdm~=4.67.1
|
||||
1
requirements/requirements-test-tokenizer-random.txt
Normal file
1
requirements/requirements-test-tokenizer-random.txt
Normal file
@@ -0,0 +1 @@
|
||||
cffi~=1.16.0
|
||||
12
requirements/requirements-tool_bench.txt
Normal file
12
requirements/requirements-tool_bench.txt
Normal file
@@ -0,0 +1,12 @@
|
||||
aiohttp~=3.9.3
|
||||
pytest~=8.3.3
|
||||
huggingface_hub>=0.34.0,<1.0
|
||||
matplotlib~=3.10.0
|
||||
numpy~=1.26.4
|
||||
openai~=1.55.3
|
||||
pandas~=2.2.3
|
||||
prometheus-client~=0.20.0
|
||||
requests~=2.32.3
|
||||
wget~=3.2
|
||||
typer~=0.15.1
|
||||
seaborn~=0.13.2
|
||||
Reference in New Issue
Block a user