[CI] Revert speedup image building and CI Installation related PRs (#6891)
### What this PR does / why we need it? Revert speedup image building and CI Installation related PRs git revert8835236181git revert64fba51275git revert263c2f8e8dgit revert84b00695f8### Does this PR introduce _any_ user-facing change? ### How was this patch tested? - vLLM version: v0.16.0 - vLLM main:15d76f74e2--------- Signed-off-by: wjunLu <wjunlu217@gmail.com>
This commit is contained in:
340
.github/workflows/_e2e_test.yaml
vendored
340
.github/workflows/_e2e_test.yaml
vendored
@@ -17,71 +17,16 @@ on:
|
||||
type: boolean
|
||||
|
||||
jobs:
|
||||
select-image:
|
||||
runs-on: linux-aarch64-a2b3-0
|
||||
outputs:
|
||||
image: ${{ steps.select.outputs.image }}
|
||||
image_a3: ${{ steps.select.outputs.image_a3 }}
|
||||
image_310p: ${{ steps.select.outputs.image_310p }}
|
||||
steps:
|
||||
- name: Select image based on base branch
|
||||
id: select
|
||||
env:
|
||||
IMAGE_NAMESPACE: swr.cn-southwest-2.myhuaweicloud.com/base_image/ascend-ci/vllm-ascend
|
||||
run: |
|
||||
BRANCH="${{ github.base_ref }}"
|
||||
BRANCH_TAG="${BRANCH//\//-}"
|
||||
MAIN_IMAGE="${{ inputs.image }}"
|
||||
MAIN_IMAGE_A3="${{ inputs.image }}-a3"
|
||||
MAIN_IMAGE_310P="${{ inputs.image }}-310p"
|
||||
if [ "$BRANCH_TAG" = "main" ]; then
|
||||
echo "Target branch is main, using main images: ${MAIN_IMAGE} / ${MAIN_IMAGE_A3} / ${MAIN_IMAGE_310P}"
|
||||
echo "image=${MAIN_IMAGE}" >> $GITHUB_OUTPUT
|
||||
echo "image_a3=${MAIN_IMAGE_A3}" >> $GITHUB_OUTPUT
|
||||
echo "image_310p=${MAIN_IMAGE_310P}" >> $GITHUB_OUTPUT
|
||||
exit 0
|
||||
fi
|
||||
|
||||
# while target branch is not main
|
||||
BRANCH_IMAGE="${IMAGE_NAMESPACE}:${BRANCH_TAG}"
|
||||
BRANCH_IMAGE_A3="${IMAGE_NAMESPACE}:${BRANCH_TAG}-a3"
|
||||
BRANCH_IMAGE_310P="${IMAGE_NAMESPACE}:${BRANCH_TAG}-310p"
|
||||
# Check if branch-specific A2 image exists via IMAGE_NAMESPACE, fallback to main if not
|
||||
if docker manifest inspect "${BRANCH_IMAGE}" > /dev/null 2>&1; then
|
||||
echo "Using branch image: ${BRANCH_IMAGE}"
|
||||
echo "image=${BRANCH_IMAGE}" >> $GITHUB_OUTPUT
|
||||
else
|
||||
echo "Branch image not found, falling back to ${MAIN_IMAGE}"
|
||||
echo "image=${MAIN_IMAGE}" >> $GITHUB_OUTPUT
|
||||
fi
|
||||
# Check if branch-specific A3 image exists via IMAGE_NAMESPACE, fallback to main if not
|
||||
if docker manifest inspect "${BRANCH_IMAGE_A3}" > /dev/null 2>&1; then
|
||||
echo "Using branch A3 image: ${BRANCH_IMAGE_A3}"
|
||||
echo "image_a3=${BRANCH_IMAGE_A3}" >> $GITHUB_OUTPUT
|
||||
else
|
||||
echo "Branch A3 image not found, falling back to ${MAIN_IMAGE_A3}"
|
||||
echo "image_a3=${MAIN_IMAGE_A3}" >> $GITHUB_OUTPUT
|
||||
fi
|
||||
# Check if branch-specific 310P image exists via IMAGE_NAMESPACE, fallback to main if not
|
||||
if docker manifest inspect "${BRANCH_IMAGE_310P}" > /dev/null 2>&1; then
|
||||
echo "Using branch 310P image: ${BRANCH_IMAGE_310P}"
|
||||
echo "image_310p=${BRANCH_IMAGE_310P}" >> $GITHUB_OUTPUT
|
||||
else
|
||||
echo "Branch 310P image not found, falling back to ${MAIN_IMAGE_310P}"
|
||||
echo "image_310p=${MAIN_IMAGE_310P}" >> $GITHUB_OUTPUT
|
||||
fi
|
||||
|
||||
e2e-light:
|
||||
name: singlecard-light
|
||||
if: ${{ inputs.type == 'light' }}
|
||||
needs: [select-image]
|
||||
runs-on: linux-aarch64-a2b3-1
|
||||
strategy:
|
||||
fail-fast: false
|
||||
matrix:
|
||||
part: [0]
|
||||
container:
|
||||
image: ${{ needs.select-image.outputs.image }}
|
||||
image: ${{ inputs.image }}
|
||||
env:
|
||||
VLLM_LOGGING_LEVEL: ERROR
|
||||
VLLM_USE_MODELSCOPE: True
|
||||
@@ -89,8 +34,6 @@ jobs:
|
||||
steps:
|
||||
- name: Checkout vllm-project/vllm-ascend repo
|
||||
uses: actions/checkout@v6
|
||||
with:
|
||||
fetch-depth: 0
|
||||
- name: Check npu and CANN info
|
||||
run: |
|
||||
npu-smi info
|
||||
@@ -103,11 +46,14 @@ jobs:
|
||||
pip config set global.trusted-host cache-service.nginx-pypi-cache.svc.cluster.local
|
||||
apt-get update -y
|
||||
apt install git -y
|
||||
git config --global --add safe.directory "${GITHUB_WORKSPACE}"
|
||||
|
||||
- name: Install system dependencies
|
||||
run: |
|
||||
apt-get -y install `cat packages.txt`
|
||||
apt-get -y install gcc g++ cmake libnuma-dev clang-15
|
||||
|
||||
update-alternatives --install /usr/bin/clang clang /usr/bin/clang-15 20
|
||||
update-alternatives --install /usr/bin/clang++ clang++ /usr/bin/clang++-15 20
|
||||
|
||||
- name: Checkout vllm-project/vllm repo
|
||||
uses: actions/checkout@v6
|
||||
@@ -117,39 +63,19 @@ jobs:
|
||||
path: ./vllm-empty
|
||||
fetch-depth: 1
|
||||
|
||||
- name: Install vllm-project/vllm
|
||||
- name: Install vllm-project/vllm from source
|
||||
working-directory: ./vllm-empty
|
||||
run: |
|
||||
pip uninstall -y vllm
|
||||
rm -rf /vllm-workspace/vllm
|
||||
cp -r ./vllm-empty /vllm-workspace/vllm
|
||||
VLLM_TARGET_DEVICE=empty pip install -v -e /vllm-workspace/vllm/
|
||||
VLLM_TARGET_DEVICE=empty pip install -e .
|
||||
|
||||
- name: Install vllm-project/vllm-ascend
|
||||
env:
|
||||
PIP_EXTRA_INDEX_URL: https://mirrors.huaweicloud.com/ascend/repos/pypi
|
||||
run: |
|
||||
DEST="/vllm-workspace/vllm-ascend"
|
||||
IMAGE_SHA=$(git -C "${DEST}" log -1 --format=%H 2>/dev/null || echo "")
|
||||
cp -rT . "${DEST}/"
|
||||
if [ -n "$IMAGE_SHA" ] && git cat-file -e "${IMAGE_SHA}" 2>/dev/null; then
|
||||
C_CHANGES=$(git diff "${IMAGE_SHA}"..HEAD --name-only -- \
|
||||
csrc/ cmake/ CMakeLists.txt setup.py requirements.txt requirements-dev.txt)
|
||||
echo "[debug] C_CHANGES=${C_CHANGES:-<empty>}"
|
||||
else
|
||||
echo "[debug] IMAGE_SHA not found in local history (empty or unreachable), forcing reinstall"
|
||||
C_CHANGES="yes"
|
||||
fi
|
||||
|
||||
pip install -r ${DEST}/requirements-dev.txt
|
||||
if [ -n "$C_CHANGES" ]; then
|
||||
echo "[debug] C code / build changes detected, reinstalling vllm-ascend..."
|
||||
pip install -v -e "${DEST}/"
|
||||
else
|
||||
echo "[debug] No C code / build changes detected, skipping reinstall."
|
||||
fi
|
||||
pip install -r requirements-dev.txt
|
||||
pip install -v -e .
|
||||
|
||||
- name: Run vllm-project/vllm-ascend test
|
||||
working-directory: /vllm-workspace/vllm-ascend
|
||||
env:
|
||||
PYTORCH_NPU_ALLOC_CONF: max_split_size_mb:256
|
||||
VLLM_WORKER_MULTIPROC_METHOD: spawn
|
||||
@@ -159,14 +85,13 @@ jobs:
|
||||
e2e-full:
|
||||
name: singlecard-full
|
||||
if: ${{ inputs.type == 'full' }}
|
||||
needs: [select-image]
|
||||
runs-on: linux-aarch64-a2b3-1
|
||||
strategy:
|
||||
fail-fast: false
|
||||
matrix:
|
||||
part: [0, 1]
|
||||
container:
|
||||
image: ${{ needs.select-image.outputs.image }}
|
||||
image: ${{ inputs.image }}
|
||||
env:
|
||||
VLLM_LOGGING_LEVEL: ERROR
|
||||
VLLM_USE_MODELSCOPE: True
|
||||
@@ -175,8 +100,6 @@ jobs:
|
||||
steps:
|
||||
- name: Checkout vllm-project/vllm-ascend repo
|
||||
uses: actions/checkout@v6
|
||||
with:
|
||||
fetch-depth: 0
|
||||
|
||||
- name: Check npu and CANN info
|
||||
run: |
|
||||
@@ -190,11 +113,14 @@ jobs:
|
||||
pip config set global.trusted-host cache-service.nginx-pypi-cache.svc.cluster.local
|
||||
apt-get update -y
|
||||
apt install git -y
|
||||
git config --global --add safe.directory "${GITHUB_WORKSPACE}"
|
||||
|
||||
- name: Install system dependencies
|
||||
run: |
|
||||
apt-get -y install `cat packages.txt`
|
||||
apt-get -y install gcc g++ cmake libnuma-dev clang-15
|
||||
|
||||
update-alternatives --install /usr/bin/clang clang /usr/bin/clang-15 20
|
||||
update-alternatives --install /usr/bin/clang++ clang++ /usr/bin/clang++-15 20
|
||||
|
||||
- name: Checkout vllm-project/vllm repo
|
||||
uses: actions/checkout@v6
|
||||
@@ -204,38 +130,18 @@ jobs:
|
||||
path: ./vllm-empty
|
||||
fetch-depth: 1
|
||||
|
||||
- name: Install vllm-project/vllm
|
||||
- name: Install vllm-project/vllm from source
|
||||
working-directory: ./vllm-empty
|
||||
run: |
|
||||
pip uninstall -y vllm
|
||||
rm -rf /vllm-workspace/vllm
|
||||
cp -r ./vllm-empty /vllm-workspace/vllm
|
||||
VLLM_TARGET_DEVICE=empty pip install -v -e /vllm-workspace/vllm/
|
||||
VLLM_TARGET_DEVICE=empty pip install -e .
|
||||
|
||||
- name: Install vllm-project/vllm-ascend
|
||||
env:
|
||||
PIP_EXTRA_INDEX_URL: https://mirrors.huaweicloud.com/ascend/repos/pypi
|
||||
run: |
|
||||
DEST="/vllm-workspace/vllm-ascend"
|
||||
IMAGE_SHA=$(git -C "${DEST}" log -1 --format=%H 2>/dev/null || echo "")
|
||||
cp -rT . "${DEST}/"
|
||||
if [ -n "$IMAGE_SHA" ] && git cat-file -e "${IMAGE_SHA}" 2>/dev/null; then
|
||||
C_CHANGES=$(git diff "${IMAGE_SHA}"..HEAD --name-only -- \
|
||||
csrc/ cmake/ CMakeLists.txt setup.py requirements.txt requirements-dev.txt)
|
||||
echo "[debug] C_CHANGES=${C_CHANGES:-<empty>}"
|
||||
else
|
||||
echo "[debug] IMAGE_SHA not found in local history (empty or unreachable), forcing reinstall"
|
||||
C_CHANGES="yes"
|
||||
fi
|
||||
pip install -r ${DEST}/requirements-dev.txt
|
||||
if [ -n "$C_CHANGES" ]; then
|
||||
echo "[debug] C code / build changes detected, reinstalling vllm-ascend..."
|
||||
pip install -v -e "${DEST}/"
|
||||
else
|
||||
echo "[debug] No C code / build changes detected, skipping reinstall."
|
||||
fi
|
||||
|
||||
pip install -r requirements-dev.txt
|
||||
pip install -v -e .
|
||||
- name: Run e2e test
|
||||
working-directory: /vllm-workspace/vllm-ascend
|
||||
env:
|
||||
VLLM_WORKER_MULTIPROC_METHOD: spawn
|
||||
PYTORCH_NPU_ALLOC_CONF: max_split_size_mb:256
|
||||
@@ -245,14 +151,13 @@ jobs:
|
||||
e2e-2-cards-light:
|
||||
name: multicard-2-light
|
||||
if: ${{ inputs.type == 'light' }}
|
||||
needs: [select-image]
|
||||
runs-on: linux-aarch64-a3-2
|
||||
strategy:
|
||||
fail-fast: false
|
||||
matrix:
|
||||
part: [0]
|
||||
container:
|
||||
image: ${{ needs.select-image.outputs.image_a3 }}
|
||||
image: swr.cn-southwest-2.myhuaweicloud.com/base_image/ascend-ci/cann:8.5.0-a3-ubuntu22.04-py3.11
|
||||
env:
|
||||
VLLM_LOGGING_LEVEL: ERROR
|
||||
VLLM_USE_MODELSCOPE: True
|
||||
@@ -261,8 +166,6 @@ jobs:
|
||||
steps:
|
||||
- name: Checkout vllm-project/vllm-ascend repo
|
||||
uses: actions/checkout@v6
|
||||
with:
|
||||
fetch-depth: 0
|
||||
- name: Check npu and CANN info
|
||||
run: |
|
||||
npu-smi info
|
||||
@@ -275,11 +178,14 @@ jobs:
|
||||
pip config set global.trusted-host cache-service.nginx-pypi-cache.svc.cluster.local
|
||||
apt-get update -y
|
||||
apt install git -y
|
||||
git config --global --add safe.directory "${GITHUB_WORKSPACE}"
|
||||
|
||||
- name: Install system dependencies
|
||||
run: |
|
||||
apt-get -y install `cat packages.txt`
|
||||
apt-get -y install gcc g++ cmake libnuma-dev clang-15
|
||||
|
||||
update-alternatives --install /usr/bin/clang clang /usr/bin/clang-15 20
|
||||
update-alternatives --install /usr/bin/clang++ clang++ /usr/bin/clang++-15 20
|
||||
|
||||
- name: Checkout vllm-project/vllm repo
|
||||
uses: actions/checkout@v6
|
||||
@@ -289,38 +195,18 @@ jobs:
|
||||
path: ./vllm-empty
|
||||
fetch-depth: 1
|
||||
|
||||
- name: Install vllm-project/vllm
|
||||
- name: Install vllm-project/vllm from source
|
||||
working-directory: ./vllm-empty
|
||||
run: |
|
||||
pip uninstall -y vllm
|
||||
rm -rf /vllm-workspace/vllm
|
||||
cp -r ./vllm-empty /vllm-workspace/vllm
|
||||
VLLM_TARGET_DEVICE=empty pip install -v -e /vllm-workspace/vllm/
|
||||
VLLM_TARGET_DEVICE=empty pip install -e .
|
||||
|
||||
- name: Install vllm-project/vllm-ascend
|
||||
env:
|
||||
PIP_EXTRA_INDEX_URL: https://mirrors.huaweicloud.com/ascend/repos/pypi
|
||||
run: |
|
||||
DEST="/vllm-workspace/vllm-ascend"
|
||||
IMAGE_SHA=$(git -C "${DEST}" log -1 --format=%H 2>/dev/null || echo "")
|
||||
cp -rT . "${DEST}/"
|
||||
if [ -n "$IMAGE_SHA" ] && git cat-file -e "${IMAGE_SHA}" 2>/dev/null; then
|
||||
C_CHANGES=$(git diff "${IMAGE_SHA}"..HEAD --name-only -- \
|
||||
csrc/ cmake/ CMakeLists.txt setup.py requirements.txt requirements-dev.txt)
|
||||
echo "[debug] C_CHANGES=${C_CHANGES:-<empty>}"
|
||||
else
|
||||
echo "[debug] IMAGE_SHA not found in local history (empty or unreachable), forcing reinstall"
|
||||
C_CHANGES="yes"
|
||||
fi
|
||||
pip install -r ${DEST}/requirements-dev.txt
|
||||
if [ -n "$C_CHANGES" ]; then
|
||||
echo "[debug] C code / build changes detected, reinstalling vllm-ascend..."
|
||||
pip install -v -e "${DEST}/"
|
||||
else
|
||||
echo "[debug] No C code / build changes detected, skipping reinstall."
|
||||
fi
|
||||
|
||||
pip install -r requirements-dev.txt
|
||||
pip install -v -e .
|
||||
- name: Run vllm-project/vllm-ascend test (light)
|
||||
working-directory: /vllm-workspace/vllm-ascend
|
||||
env:
|
||||
VLLM_WORKER_MULTIPROC_METHOD: spawn
|
||||
run: |
|
||||
@@ -329,14 +215,13 @@ jobs:
|
||||
e2e-2-cards-full:
|
||||
name: multicard-2-full
|
||||
if: ${{ inputs.type == 'full' }}
|
||||
needs: [select-image]
|
||||
runs-on: linux-aarch64-a3-2
|
||||
strategy:
|
||||
fail-fast: false
|
||||
matrix:
|
||||
part: [0]
|
||||
container:
|
||||
image: ${{ needs.select-image.outputs.image_a3 }}
|
||||
image: swr.cn-southwest-2.myhuaweicloud.com/base_image/ascend-ci/cann:8.5.0-a3-ubuntu22.04-py3.11
|
||||
env:
|
||||
VLLM_LOGGING_LEVEL: ERROR
|
||||
VLLM_USE_MODELSCOPE: True
|
||||
@@ -345,8 +230,6 @@ jobs:
|
||||
steps:
|
||||
- name: Checkout vllm-project/vllm-ascend repo
|
||||
uses: actions/checkout@v6
|
||||
with:
|
||||
fetch-depth: 0
|
||||
- name: Check npu and CANN info
|
||||
run: |
|
||||
npu-smi info
|
||||
@@ -359,11 +242,14 @@ jobs:
|
||||
pip config set global.trusted-host cache-service.nginx-pypi-cache.svc.cluster.local
|
||||
apt-get update -y
|
||||
apt install git -y
|
||||
git config --global --add safe.directory "${GITHUB_WORKSPACE}"
|
||||
|
||||
- name: Install system dependencies
|
||||
run: |
|
||||
apt-get -y install `cat packages.txt`
|
||||
apt-get -y install gcc g++ cmake libnuma-dev clang-15
|
||||
|
||||
update-alternatives --install /usr/bin/clang clang /usr/bin/clang-15 20
|
||||
update-alternatives --install /usr/bin/clang++ clang++ /usr/bin/clang++-15 20
|
||||
|
||||
- name: Checkout vllm-project/vllm repo
|
||||
uses: actions/checkout@v6
|
||||
@@ -373,38 +259,18 @@ jobs:
|
||||
path: ./vllm-empty
|
||||
fetch-depth: 1
|
||||
|
||||
- name: Install vllm-project/vllm
|
||||
- name: Install vllm-project/vllm from source
|
||||
working-directory: ./vllm-empty
|
||||
run: |
|
||||
pip uninstall -y vllm
|
||||
rm -rf /vllm-workspace/vllm
|
||||
cp -r ./vllm-empty /vllm-workspace/vllm
|
||||
VLLM_TARGET_DEVICE=empty pip install -v -e /vllm-workspace/vllm/
|
||||
VLLM_TARGET_DEVICE=empty pip install -e .
|
||||
|
||||
- name: Install vllm-project/vllm-ascend
|
||||
env:
|
||||
PIP_EXTRA_INDEX_URL: https://mirrors.huaweicloud.com/ascend/repos/pypi
|
||||
run: |
|
||||
DEST="/vllm-workspace/vllm-ascend"
|
||||
IMAGE_SHA=$(git -C "${DEST}" log -1 --format=%H 2>/dev/null || echo "")
|
||||
cp -rT . "${DEST}/"
|
||||
if [ -n "$IMAGE_SHA" ] && git cat-file -e "${IMAGE_SHA}" 2>/dev/null; then
|
||||
C_CHANGES=$(git diff "${IMAGE_SHA}"..HEAD --name-only -- \
|
||||
csrc/ cmake/ CMakeLists.txt setup.py requirements.txt requirements-dev.txt)
|
||||
echo "[debug] C_CHANGES=${C_CHANGES:-<empty>}"
|
||||
else
|
||||
echo "[debug] IMAGE_SHA not found in local history (empty or unreachable), forcing reinstall"
|
||||
C_CHANGES="yes"
|
||||
fi
|
||||
pip install -r ${DEST}/requirements-dev.txt
|
||||
if [ -n "$C_CHANGES" ]; then
|
||||
echo "[debug] C code / build changes detected, reinstalling vllm-ascend..."
|
||||
pip install -v -e "${DEST}/"
|
||||
else
|
||||
echo "[debug] No C code / build changes detected, skipping reinstall."
|
||||
fi
|
||||
|
||||
pip install -r requirements-dev.txt
|
||||
pip install -v -e .
|
||||
- name: Run vllm-project/vllm-ascend test (full)
|
||||
working-directory: /vllm-workspace/vllm-ascend
|
||||
env:
|
||||
VLLM_WORKER_MULTIPROC_METHOD: spawn
|
||||
run: |
|
||||
@@ -412,7 +278,6 @@ jobs:
|
||||
|
||||
- name: Run vllm-project/vllm-ascend test (non triton)
|
||||
if: ${{ inputs.type == 'full' && matrix.part == 0 }}
|
||||
working-directory: /vllm-workspace/vllm-ascend
|
||||
env:
|
||||
VLLM_WORKER_MULTIPROC_METHOD: spawn
|
||||
run: |
|
||||
@@ -422,14 +287,13 @@ jobs:
|
||||
e2e-4-cards-full:
|
||||
name: multicard-4-full
|
||||
if: ${{ inputs.type == 'full' }}
|
||||
needs: [select-image]
|
||||
runs-on: linux-aarch64-a3-4
|
||||
strategy:
|
||||
fail-fast: false
|
||||
matrix:
|
||||
part: [0]
|
||||
container:
|
||||
image: ${{ needs.select-image.outputs.image_a3 }}
|
||||
image: m.daocloud.io/quay.io/ascend/cann:8.5.0-a3-ubuntu22.04-py3.11
|
||||
env:
|
||||
VLLM_LOGGING_LEVEL: ERROR
|
||||
VLLM_USE_MODELSCOPE: True
|
||||
@@ -437,8 +301,6 @@ jobs:
|
||||
steps:
|
||||
- name: Checkout vllm-project/vllm-ascend repo
|
||||
uses: actions/checkout@v6
|
||||
with:
|
||||
fetch-depth: 0
|
||||
- name: Check npu and CANN info
|
||||
run: |
|
||||
npu-smi info
|
||||
@@ -451,11 +313,14 @@ jobs:
|
||||
pip config set global.trusted-host cache-service.nginx-pypi-cache.svc.cluster.local
|
||||
apt-get update -y
|
||||
apt install git -y
|
||||
git config --global --add safe.directory "${GITHUB_WORKSPACE}"
|
||||
|
||||
- name: Install system dependencies
|
||||
run: |
|
||||
apt-get -y install `cat packages.txt`
|
||||
apt-get -y install gcc g++ cmake libnuma-dev clang-15
|
||||
|
||||
update-alternatives --install /usr/bin/clang clang /usr/bin/clang-15 20
|
||||
update-alternatives --install /usr/bin/clang++ clang++ /usr/bin/clang++-15 20
|
||||
|
||||
- name: Checkout vllm-project/vllm repo
|
||||
uses: actions/checkout@v6
|
||||
@@ -465,38 +330,19 @@ jobs:
|
||||
path: ./vllm-empty
|
||||
fetch-depth: 1
|
||||
|
||||
- name: Install vllm-project/vllm
|
||||
- name: Install vllm-project/vllm from source
|
||||
working-directory: ./vllm-empty
|
||||
run: |
|
||||
pip uninstall -y vllm
|
||||
rm -rf /vllm-workspace/vllm
|
||||
cp -r ./vllm-empty /vllm-workspace/vllm
|
||||
VLLM_TARGET_DEVICE=empty pip install -v -e /vllm-workspace/vllm/
|
||||
VLLM_TARGET_DEVICE=empty pip install -e .
|
||||
|
||||
- name: Install vllm-project/vllm-ascend
|
||||
env:
|
||||
PIP_EXTRA_INDEX_URL: https://mirrors.huaweicloud.com/ascend/repos/pypi
|
||||
run: |
|
||||
DEST="/vllm-workspace/vllm-ascend"
|
||||
IMAGE_SHA=$(git -C "${DEST}" log -1 --format=%H 2>/dev/null || echo "")
|
||||
cp -rT . "${DEST}/"
|
||||
if [ -n "$IMAGE_SHA" ] && git cat-file -e "${IMAGE_SHA}" 2>/dev/null; then
|
||||
C_CHANGES=$(git diff "${IMAGE_SHA}"..HEAD --name-only -- \
|
||||
csrc/ cmake/ CMakeLists.txt setup.py requirements.txt requirements-dev.txt)
|
||||
echo "[debug] C_CHANGES=${C_CHANGES:-<empty>}"
|
||||
else
|
||||
echo "[debug] IMAGE_SHA not found in local history (empty or unreachable), forcing reinstall"
|
||||
C_CHANGES="yes"
|
||||
fi
|
||||
pip install -r ${DEST}/requirements-dev.txt
|
||||
if [ -n "$C_CHANGES" ]; then
|
||||
echo "[debug] C code / build changes detected, reinstalling vllm-ascend..."
|
||||
pip install -v -e "${DEST}/"
|
||||
else
|
||||
echo "[debug] No C code / build changes detected, skipping reinstall."
|
||||
fi
|
||||
pip install -r requirements-dev.txt
|
||||
pip install -v -e .
|
||||
|
||||
- name: Run vllm-project/vllm-ascend test for V1 Engine
|
||||
working-directory: /vllm-workspace/vllm-ascend
|
||||
env:
|
||||
VLLM_WORKER_MULTIPROC_METHOD: spawn
|
||||
run: |
|
||||
@@ -506,24 +352,17 @@ jobs:
|
||||
name: 310p singlecard
|
||||
runs-on: linux-aarch64-310p-1
|
||||
if: ${{ inputs.contains_310 }}
|
||||
needs: [select-image]
|
||||
container:
|
||||
image: ${{ needs.select-image.outputs.image_310p }}
|
||||
image: swr.cn-southwest-2.myhuaweicloud.com/base_image/ascend-ci/cann:8.5.0-310p-ubuntu22.04-py3.11
|
||||
env:
|
||||
VLLM_LOGGING_LEVEL: ERROR
|
||||
VLLM_USE_MODELSCOPE: True
|
||||
HF_HUB_OFFLINE: 1
|
||||
steps:
|
||||
- name: Checkout vllm-project/vllm-ascend repo
|
||||
uses: actions/checkout@v6
|
||||
with:
|
||||
fetch-depth: 0
|
||||
|
||||
- name: Check npu and CANN info
|
||||
run: |
|
||||
npu-smi info
|
||||
cat /usr/local/Ascend/ascend-toolkit/latest/"$(uname -i)"-linux/ascend_toolkit_install.info
|
||||
|
||||
- name: Config mirrors
|
||||
run: |
|
||||
sed -Ei 's@(ports|archive).ubuntu.com@cache-service.nginx-pypi-cache.svc.cluster.local:8081@g' /etc/apt/sources.list
|
||||
@@ -531,11 +370,14 @@ jobs:
|
||||
pip config set global.trusted-host cache-service.nginx-pypi-cache.svc.cluster.local
|
||||
apt-get update -y
|
||||
apt install git -y
|
||||
git config --global --add safe.directory "${GITHUB_WORKSPACE}"
|
||||
|
||||
- name: Checkout vllm-project/vllm-ascend repo
|
||||
uses: actions/checkout@v6
|
||||
|
||||
- name: Install system dependencies
|
||||
run: |
|
||||
apt-get -y install `cat packages.txt`
|
||||
apt-get -y install gcc g++ cmake libnuma-dev
|
||||
|
||||
- name: Checkout vllm-project/vllm repo
|
||||
uses: actions/checkout@v6
|
||||
@@ -545,38 +387,19 @@ jobs:
|
||||
path: ./vllm-empty
|
||||
fetch-depth: 1
|
||||
|
||||
- name: Install vllm-project/vllm
|
||||
- name: Install vllm-project/vllm from source
|
||||
working-directory: ./vllm-empty
|
||||
run: |
|
||||
pip uninstall -y vllm
|
||||
rm -rf /vllm-workspace/vllm
|
||||
cp -r ./vllm-empty /vllm-workspace/vllm
|
||||
VLLM_TARGET_DEVICE=empty pip install -v -e /vllm-workspace/vllm/
|
||||
VLLM_TARGET_DEVICE=empty pip install -e .
|
||||
|
||||
- name: Install vllm-project/vllm-ascend
|
||||
env:
|
||||
PIP_EXTRA_INDEX_URL: https://mirrors.huaweicloud.com/ascend/repos/pypi
|
||||
run: |
|
||||
DEST="/vllm-workspace/vllm-ascend"
|
||||
IMAGE_SHA=$(git -C "${DEST}" log -1 --format=%H 2>/dev/null || echo "")
|
||||
cp -rT . "${DEST}/"
|
||||
if [ -n "$IMAGE_SHA" ] && git cat-file -e "${IMAGE_SHA}" 2>/dev/null; then
|
||||
C_CHANGES=$(git diff "${IMAGE_SHA}"..HEAD --name-only -- \
|
||||
csrc/ cmake/ CMakeLists.txt setup.py requirements.txt requirements-dev.txt)
|
||||
echo "[debug] C_CHANGES=${C_CHANGES:-<empty>}"
|
||||
else
|
||||
echo "[debug] IMAGE_SHA not found in local history (empty or unreachable), forcing reinstall"
|
||||
C_CHANGES="yes"
|
||||
fi
|
||||
pip install -r ${DEST}/requirements-dev.txt
|
||||
if [ -n "$C_CHANGES" ]; then
|
||||
echo "[debug] C code / build changes detected, reinstalling vllm-ascend..."
|
||||
pip install -v -e "${DEST}/"
|
||||
else
|
||||
echo "[debug] No C code / build changes detected, skipping reinstall."
|
||||
fi
|
||||
pip install -r requirements-dev.txt
|
||||
pip install -v -e .
|
||||
|
||||
- name: Run vllm-project/vllm-ascend test
|
||||
working-directory: /vllm-workspace/vllm-ascend
|
||||
env:
|
||||
PYTORCH_NPU_ALLOC_CONF: max_split_size_mb:256
|
||||
VLLM_WORKER_MULTIPROC_METHOD: spawn
|
||||
@@ -587,24 +410,17 @@ jobs:
|
||||
name: 310p multicards 4cards
|
||||
runs-on: linux-aarch64-310p-4
|
||||
if: ${{ inputs.contains_310 }}
|
||||
needs: [select-image]
|
||||
container:
|
||||
image: ${{ needs.select-image.outputs.image_310p }}
|
||||
image: swr.cn-southwest-2.myhuaweicloud.com/base_image/ascend-ci/cann:8.5.0-310p-ubuntu22.04-py3.11
|
||||
env:
|
||||
VLLM_LOGGING_LEVEL: ERROR
|
||||
VLLM_USE_MODELSCOPE: True
|
||||
HF_HUB_OFFLINE: 1
|
||||
steps:
|
||||
- name: Checkout vllm-project/vllm-ascend repo
|
||||
uses: actions/checkout@v6
|
||||
with:
|
||||
fetch-depth: 0
|
||||
|
||||
- name: Check npu and CANN info
|
||||
run: |
|
||||
npu-smi info
|
||||
cat /usr/local/Ascend/ascend-toolkit/latest/"$(uname -i)"-linux/ascend_toolkit_install.info
|
||||
|
||||
- name: Config mirrors
|
||||
run: |
|
||||
sed -Ei 's@(ports|archive).ubuntu.com@cache-service.nginx-pypi-cache.svc.cluster.local:8081@g' /etc/apt/sources.list
|
||||
@@ -612,11 +428,14 @@ jobs:
|
||||
pip config set global.trusted-host cache-service.nginx-pypi-cache.svc.cluster.local
|
||||
apt-get update -y
|
||||
apt install git -y
|
||||
git config --global --add safe.directory "${GITHUB_WORKSPACE}"
|
||||
|
||||
- name: Checkout vllm-project/vllm-ascend repo
|
||||
uses: actions/checkout@v6
|
||||
|
||||
- name: Install system dependencies
|
||||
run: |
|
||||
apt-get -y install `cat packages.txt`
|
||||
apt-get -y install gcc g++ cmake libnuma-dev
|
||||
|
||||
- name: Checkout vllm-project/vllm repo
|
||||
uses: actions/checkout@v6
|
||||
@@ -626,38 +445,19 @@ jobs:
|
||||
path: ./vllm-empty
|
||||
fetch-depth: 1
|
||||
|
||||
- name: Install vllm-project/vllm
|
||||
- name: Install vllm-project/vllm from source
|
||||
working-directory: ./vllm-empty
|
||||
run: |
|
||||
pip uninstall -y vllm
|
||||
rm -rf /vllm-workspace/vllm
|
||||
cp -r ./vllm-empty /vllm-workspace/vllm
|
||||
VLLM_TARGET_DEVICE=empty pip install -v -e /vllm-workspace/vllm/
|
||||
VLLM_TARGET_DEVICE=empty pip install -e .
|
||||
|
||||
- name: Install vllm-project/vllm-ascend
|
||||
env:
|
||||
PIP_EXTRA_INDEX_URL: https://mirrors.huaweicloud.com/ascend/repos/pypi
|
||||
run: |
|
||||
DEST="/vllm-workspace/vllm-ascend"
|
||||
IMAGE_SHA=$(git -C "${DEST}" log -1 --format=%H 2>/dev/null || echo "")
|
||||
cp -rT . "${DEST}/"
|
||||
if [ -n "$IMAGE_SHA" ] && git cat-file -e "${IMAGE_SHA}" 2>/dev/null; then
|
||||
C_CHANGES=$(git diff "${IMAGE_SHA}"..HEAD --name-only -- \
|
||||
csrc/ cmake/ CMakeLists.txt setup.py requirements.txt requirements-dev.txt)
|
||||
echo "[debug] C_CHANGES=${C_CHANGES:-<empty>}"
|
||||
else
|
||||
echo "[debug] IMAGE_SHA not found in local history (empty or unreachable), forcing reinstall"
|
||||
C_CHANGES="yes"
|
||||
fi
|
||||
pip install -r ${DEST}/requirements-dev.txt
|
||||
if [ -n "$C_CHANGES" ]; then
|
||||
echo "[debug] C code / build changes detected, reinstalling vllm-ascend..."
|
||||
pip install -v -e "${DEST}/"
|
||||
else
|
||||
echo "[debug] No C code / build changes detected, skipping reinstall."
|
||||
fi
|
||||
pip install -r requirements-dev.txt
|
||||
pip install -v -e .
|
||||
|
||||
- name: Run vllm-project/vllm-ascend test
|
||||
working-directory: /vllm-workspace/vllm-ascend
|
||||
env:
|
||||
PYTORCH_NPU_ALLOC_CONF: max_split_size_mb:256
|
||||
VLLM_WORKER_MULTIPROC_METHOD: spawn
|
||||
|
||||
25
.github/workflows/_schedule_image_build.yaml
vendored
25
.github/workflows/_schedule_image_build.yaml
vendored
@@ -76,28 +76,6 @@ jobs:
|
||||
driver: docker-container
|
||||
use: true
|
||||
|
||||
- name: Set cache ref
|
||||
id: cache
|
||||
run: |
|
||||
if [ "${{ github.ref_type }}" = "tag" ]; then
|
||||
# For tag events, use the images built from source branch as cache (the tag image doesn't exist yet).
|
||||
if [ -z "$branch" ]; then
|
||||
branch=$(git branch -r --contains HEAD \
|
||||
| grep -v 'HEAD' \
|
||||
| sed 's|[[:space:]]*origin/||' \
|
||||
| head -1)
|
||||
fi
|
||||
branch="${branch:-main}"
|
||||
else
|
||||
# For PR events github.ref_name is "<PR_number>/merge" which has no cached image;
|
||||
# use base_ref (target branch) instead. For push/schedule, base_ref is empty so
|
||||
# fall back to ref_name which is the actual branch name.
|
||||
branch="${{ github.base_ref || github.ref_name }}"
|
||||
fi
|
||||
# Replace / with - for use in image tags
|
||||
branch="${branch//\//-}"
|
||||
echo "ref=quay.io/ascend/vllm-ascend:${branch}-${{ inputs.suffix }}" >> $GITHUB_OUTPUT
|
||||
|
||||
- name: Build and push
|
||||
uses: docker/build-push-action@v6
|
||||
id: build
|
||||
@@ -111,8 +89,6 @@ jobs:
|
||||
outputs: type=image,name=quay.io/ascend/vllm-ascend,push-by-digest=true,name-canonical=true,push=${{ inputs.should_push }}
|
||||
build-args: |
|
||||
PIP_INDEX_URL=https://pypi.org/simple
|
||||
# use previously pushed multi-arch image as cache to speed up builds
|
||||
cache-from: type=registry,ref=${{ steps.cache.outputs.ref }}
|
||||
provenance: false
|
||||
|
||||
- name: Export digest
|
||||
@@ -178,7 +154,6 @@ jobs:
|
||||
# which follow the rule from vLLM with prefix v
|
||||
# TODO(yikun): the post release might be considered as latest release
|
||||
tags: |
|
||||
type=ref,event=branch,suffix=${{ env.SUFFIX }}
|
||||
type=pep440,pattern={{raw}},suffix=${{ env.SUFFIX }}
|
||||
type=schedule,pattern=main,suffix=${{ env.SUFFIX }}
|
||||
type=raw,value=${{ inputs.workflow_dispatch_tag }},enable=${{ github.event_name == 'workflow_dispatch' }},suffix=${{ env.SUFFIX }}
|
||||
|
||||
2
.github/workflows/pr_test_full.yaml
vendored
2
.github/workflows/pr_test_full.yaml
vendored
@@ -81,6 +81,6 @@ jobs:
|
||||
uses: ./.github/workflows/_e2e_test.yaml
|
||||
with:
|
||||
vllm: ${{ matrix.vllm_version }}
|
||||
image: swr.cn-southwest-2.myhuaweicloud.com/base_image/ascend-ci/vllm-ascend:main
|
||||
image: swr.cn-southwest-2.myhuaweicloud.com/base_image/ascend-ci/cann:8.5.0-910b-ubuntu22.04-py3.11
|
||||
contains_310: false
|
||||
type: full
|
||||
|
||||
2
.github/workflows/pr_test_light.yaml
vendored
2
.github/workflows/pr_test_light.yaml
vendored
@@ -107,6 +107,6 @@ jobs:
|
||||
uses: ./.github/workflows/_e2e_test.yaml
|
||||
with:
|
||||
vllm: ${{ matrix.vllm_version }}
|
||||
image: swr.cn-southwest-2.myhuaweicloud.com/base_image/ascend-ci/vllm-ascend:main
|
||||
image: swr.cn-southwest-2.myhuaweicloud.com/base_image/ascend-ci/cann:8.5.0-910b-ubuntu22.04-py3.11
|
||||
contains_310: ${{ needs.changes.outputs._310_tracker == 'true' }}
|
||||
type: light
|
||||
|
||||
@@ -12,15 +12,12 @@
|
||||
name: Image Build and Push
|
||||
on:
|
||||
schedule:
|
||||
# UTC+8: 8am, 10am, 12pm, 14pm, 16pm, 18pm, 22pm
|
||||
- cron: '0 0,2,4,6,8,10,14 * * *'
|
||||
# UTC+8: 8am, 12pm, 16pm, 22pm
|
||||
- cron: '0 0,4,8,14 * * *'
|
||||
push:
|
||||
branches:
|
||||
# Build release branch images proactively so cache is warm when the tag is pushed
|
||||
- 'releases/*'
|
||||
tags:
|
||||
- 'v*'
|
||||
pull_request:
|
||||
pull_request:
|
||||
branches:
|
||||
- 'main'
|
||||
types: [ labeled, synchronize ]
|
||||
|
||||
@@ -50,7 +50,6 @@ jobs:
|
||||
--build-arg CANN_VERSION="8.5.0" \
|
||||
--build-arg UBUNTU_VERSION="22.04" \
|
||||
--build-arg PYTHON_VERSION="3.11" \
|
||||
--cache-from "$IMAGE_TAG" \
|
||||
-t "$IMAGE_TAG" .
|
||||
|
||||
echo "image-tag=$IMAGE_TAG" >> $GITHUB_OUTPUT
|
||||
|
||||
@@ -34,6 +34,6 @@ jobs:
|
||||
uses: ./.github/workflows/_e2e_test.yaml
|
||||
with:
|
||||
vllm: main
|
||||
image: swr.cn-southwest-2.myhuaweicloud.com/base_image/ascend-ci/vllm-ascend:main
|
||||
image: swr.cn-southwest-2.myhuaweicloud.com/base_image/ascend-ci/cann:8.5.0-910b-ubuntu22.04-py3.11
|
||||
contains_310: false
|
||||
type: full
|
||||
|
||||
Reference in New Issue
Block a user