mirror of https://github.com/vllm-project/vllm
[OpenVINO] migrate to latest dependencies versions (#7251)
This commit is contained in:
parent
b764547616
commit
80cbe10c59
|
@ -21,7 +21,7 @@ COPY setup.py /workspace/vllm/
|
||||||
# install build requirements
|
# install build requirements
|
||||||
RUN PIP_EXTRA_INDEX_URL="https://download.pytorch.org/whl/cpu" python3 -m pip install -r /workspace/vllm/requirements-build.txt
|
RUN PIP_EXTRA_INDEX_URL="https://download.pytorch.org/whl/cpu" python3 -m pip install -r /workspace/vllm/requirements-build.txt
|
||||||
# build vLLM with OpenVINO backend
|
# build vLLM with OpenVINO backend
|
||||||
RUN PIP_EXTRA_INDEX_URL="https://download.pytorch.org/whl/cpu https://storage.openvinotoolkit.org/simple/wheels/pre-release" VLLM_TARGET_DEVICE="openvino" python3 -m pip install /workspace/vllm/
|
RUN PIP_EXTRA_INDEX_URL="https://download.pytorch.org/whl/cpu" VLLM_TARGET_DEVICE="openvino" python3 -m pip install /workspace/vllm/
|
||||||
|
|
||||||
COPY examples/ /workspace/vllm/examples
|
COPY examples/ /workspace/vllm/examples
|
||||||
COPY benchmarks/ /workspace/vllm/benchmarks
|
COPY benchmarks/ /workspace/vllm/benchmarks
|
||||||
|
|
|
@ -57,7 +57,7 @@ Install from source
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
|
|
||||||
$ PIP_EXTRA_INDEX_URL="https://download.pytorch.org/whl/cpu https://storage.openvinotoolkit.org/simple/wheels/pre-release" VLLM_TARGET_DEVICE=openvino python -m pip install -v .
|
$ PIP_EXTRA_INDEX_URL="https://download.pytorch.org/whl/cpu" VLLM_TARGET_DEVICE=openvino python -m pip install -v .
|
||||||
|
|
||||||
.. _openvino_backend_performance_tips:
|
.. _openvino_backend_performance_tips:
|
||||||
|
|
||||||
|
|
|
@ -1,34 +1,7 @@
|
||||||
# Common dependencies
|
# Common dependencies
|
||||||
# -r requirements-common.txt
|
-r requirements-common.txt
|
||||||
# TODO: remove temporary copy of all common dependencies once Optimum Intel will support Transformers >= 4.43.2
|
|
||||||
cmake >= 3.21
|
|
||||||
ninja # For faster builds.
|
|
||||||
psutil
|
|
||||||
sentencepiece # Required for LLaMA tokenizer.
|
|
||||||
numpy < 2.0.0
|
|
||||||
requests
|
|
||||||
tqdm
|
|
||||||
py-cpuinfo
|
|
||||||
transformers < 4.43
|
|
||||||
tokenizers >= 0.19.1 # Required for Llama 3.
|
|
||||||
fastapi
|
|
||||||
aiohttp
|
|
||||||
openai
|
|
||||||
uvicorn[standard]
|
|
||||||
pydantic >= 2.0 # Required for OpenAI server.
|
|
||||||
pillow # Required for image processing
|
|
||||||
prometheus_client >= 0.18.0
|
|
||||||
prometheus-fastapi-instrumentator >= 7.0.0
|
|
||||||
tiktoken >= 0.6.0 # Required for DBRX tokenizer
|
|
||||||
lm-format-enforcer == 0.10.3
|
|
||||||
outlines >= 0.0.43, < 0.1 # Requires torch >= 2.1.0
|
|
||||||
typing_extensions
|
|
||||||
filelock >= 3.10.4 # filelock starts to support `mode` argument from 3.10.4
|
|
||||||
pyzmq
|
|
||||||
gguf == 0.9.1
|
|
||||||
|
|
||||||
# OpenVINO dependencies
|
# OpenVINO dependencies
|
||||||
torch >= 2.1.2
|
torch >= 2.1.2
|
||||||
openvino ~= 2024.3.0.dev
|
openvino ~= 2024.3.0
|
||||||
openvino-tokenizers[transformers] ~= 2024.3.0.0.dev
|
optimum-intel[openvino] >= 1.18.2
|
||||||
optimum-intel[openvino] >= 1.18.1
|
|
||||||
|
|
2
setup.py
2
setup.py
|
@ -272,7 +272,7 @@ def _build_custom_ops() -> bool:
|
||||||
|
|
||||||
|
|
||||||
def _build_core_ext() -> bool:
|
def _build_core_ext() -> bool:
|
||||||
return not _is_neuron() and not _is_tpu()
|
return not _is_neuron() and not _is_tpu() and not _is_openvino()
|
||||||
|
|
||||||
|
|
||||||
def get_hipcc_rocm_version():
|
def get_hipcc_rocm_version():
|
||||||
|
|
Loading…
Reference in New Issue