Publish pre-built wheels with CUDA support for Linux aarch64. (#1507)
This commit is contained in:
120
.github/workflows/build-wheels-aarch64-cuda.yaml
vendored
Normal file
120
.github/workflows/build-wheels-aarch64-cuda.yaml
vendored
Normal file
@@ -0,0 +1,120 @@
|
|||||||
|
name: build-wheels-aarch64-cuda
|
||||||
|
|
||||||
|
on:
|
||||||
|
push:
|
||||||
|
branches:
|
||||||
|
- wheel
|
||||||
|
workflow_dispatch:
|
||||||
|
|
||||||
|
env:
|
||||||
|
SHERPA_ONNX_IS_IN_GITHUB_ACTIONS: 1
|
||||||
|
|
||||||
|
concurrency:
|
||||||
|
group: build-wheels-aarch64-cuda-${{ github.ref }}
|
||||||
|
cancel-in-progress: true
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
build_wheels_aarch64_cuda:
|
||||||
|
name: ${{ matrix.manylinux }} ${{ matrix.python-version }}
|
||||||
|
runs-on: ${{ matrix.os }}
|
||||||
|
strategy:
|
||||||
|
fail-fast: false
|
||||||
|
matrix:
|
||||||
|
os: [ubuntu-latest]
|
||||||
|
python-version: ["cp37", "cp38", "cp39", "cp310", "cp311", "cp312", "cp313"]
|
||||||
|
manylinux: [manylinux2014] #, manylinux_2_28]
|
||||||
|
|
||||||
|
steps:
|
||||||
|
- uses: actions/checkout@v4
|
||||||
|
|
||||||
|
- name: Set up QEMU
|
||||||
|
uses: docker/setup-qemu-action@v2
|
||||||
|
with:
|
||||||
|
platforms: all
|
||||||
|
|
||||||
|
# see https://cibuildwheel.readthedocs.io/en/stable/changelog/
|
||||||
|
# for a list of versions
|
||||||
|
- name: Build wheels
|
||||||
|
uses: pypa/cibuildwheel@v2.21.3
|
||||||
|
env:
|
||||||
|
CIBW_BEFORE_ALL: |
|
||||||
|
git clone --depth 1 --branch v1.2.12 https://github.com/alsa-project/alsa-lib
|
||||||
|
cd alsa-lib
|
||||||
|
./gitcompile
|
||||||
|
cd ..
|
||||||
|
echo "PWD"
|
||||||
|
ls -lh /project/alsa-lib/src/.libs
|
||||||
|
|
||||||
|
CIBW_ENVIRONMENT: CPLUS_INCLUDE_PATH=/project/alsa-lib/include:$CPLUS_INCLUDE_PATH SHERPA_ONNX_ALSA_LIB_DIR=/project/alsa-lib/src/.libs LD_LIBRARY_PATH=/project/build/bdist.linux-x86_64/wheel/sherpa_onnx/lib:$SHERPA_ONNX_ALSA_LIB_DIR SHERPA_ONNX_MAKE_ARGS="VERBOSE=1" SHERPA_ONNX_ENABLE_ALSA=1 SHERPA_ONNX_ENABLE_GPU=ON
|
||||||
|
CIBW_BUILD: "${{ matrix.python-version}}-* "
|
||||||
|
CIBW_SKIP: "cp27-* cp35-* cp36-* *-win32 pp* *-musllinux* *-manylinux_i686"
|
||||||
|
CIBW_BUILD_VERBOSITY: 3
|
||||||
|
CIBW_ARCHS_LINUX: aarch64
|
||||||
|
CIBW_MANYLINUX_AARCH64_IMAGE: quay.io/pypa/${{ matrix.manylinux }}_aarch64
|
||||||
|
# From onnxruntime >= 1.17.0, it drops support for CentOS 7.0 and it supports only manylinux_2_28.
|
||||||
|
# manylinux_2_24 is no longer supported
|
||||||
|
|
||||||
|
- name: Display wheels
|
||||||
|
shell: bash
|
||||||
|
run: |
|
||||||
|
ls -lh ./wheelhouse/
|
||||||
|
|
||||||
|
- name: Install patchelf
|
||||||
|
if: matrix.os == 'ubuntu-latest'
|
||||||
|
shell: bash
|
||||||
|
run: |
|
||||||
|
sudo apt-get update -q
|
||||||
|
sudo apt-get install -q -y patchelf
|
||||||
|
patchelf --help
|
||||||
|
|
||||||
|
- name: Patch wheels
|
||||||
|
shell: bash
|
||||||
|
if: matrix.os == 'ubuntu-latest'
|
||||||
|
run: |
|
||||||
|
mkdir ./wheels
|
||||||
|
sudo ./scripts/wheel/patch_wheel.py --in-dir ./wheelhouse --out-dir ./wheels
|
||||||
|
|
||||||
|
ls -lh ./wheels/
|
||||||
|
rm -rf ./wheelhouse
|
||||||
|
mv ./wheels ./wheelhouse
|
||||||
|
|
||||||
|
- name: Publish to huggingface
|
||||||
|
env:
|
||||||
|
HF_TOKEN: ${{ secrets.HF_TOKEN }}
|
||||||
|
uses: nick-fields/retry@v3
|
||||||
|
with:
|
||||||
|
max_attempts: 20
|
||||||
|
timeout_seconds: 200
|
||||||
|
shell: bash
|
||||||
|
command: |
|
||||||
|
git config --global user.email "csukuangfj@gmail.com"
|
||||||
|
git config --global user.name "Fangjun Kuang"
|
||||||
|
|
||||||
|
rm -rf huggingface
|
||||||
|
export GIT_LFS_SKIP_SMUDGE=1
|
||||||
|
export GIT_CLONE_PROTECTION_ACTIVE=false
|
||||||
|
|
||||||
|
SHERPA_ONNX_VERSION=$(grep "SHERPA_ONNX_VERSION" ./CMakeLists.txt | cut -d " " -f 2 | cut -d '"' -f 2)
|
||||||
|
echo "SHERPA_ONNX_VERSION $SHERPA_ONNX_VERSION"
|
||||||
|
|
||||||
|
d=cuda/$SHERPA_ONNX_VERSION
|
||||||
|
|
||||||
|
git clone https://csukuangfj:$HF_TOKEN@huggingface.co/csukuangfj/sherpa-onnx-wheels huggingface
|
||||||
|
cd huggingface
|
||||||
|
git fetch
|
||||||
|
git pull
|
||||||
|
git merge -m "merge remote" --ff origin main
|
||||||
|
|
||||||
|
mkdir -p $d
|
||||||
|
|
||||||
|
cp -v ../wheelhouse/*.whl $d/
|
||||||
|
|
||||||
|
git status
|
||||||
|
git add .
|
||||||
|
git commit -m "add more wheels"
|
||||||
|
git push https://csukuangfj:$HF_TOKEN@huggingface.co/csukuangfj/sherpa-onnx-wheels main
|
||||||
|
|
||||||
|
- uses: actions/upload-artifact@v4
|
||||||
|
with:
|
||||||
|
name: wheel-${{ matrix.python-version }}-${{ matrix.manylinux }}
|
||||||
|
path: ./wheelhouse/*.whl
|
||||||
@@ -7,10 +7,28 @@
|
|||||||
//
|
//
|
||||||
// clang-format off
|
// clang-format off
|
||||||
//
|
//
|
||||||
|
// cd /path/sherpa-onnx/
|
||||||
|
// mkdir build
|
||||||
|
// cd build
|
||||||
|
// cmake ..
|
||||||
|
// make
|
||||||
|
//
|
||||||
// wget https://github.com/k2-fsa/sherpa-onnx/releases/download/asr-models/sherpa-onnx-streaming-zipformer-bilingual-zh-en-2023-02-20.tar.bz2
|
// wget https://github.com/k2-fsa/sherpa-onnx/releases/download/asr-models/sherpa-onnx-streaming-zipformer-bilingual-zh-en-2023-02-20.tar.bz2
|
||||||
// tar xvf sherpa-onnx-streaming-zipformer-bilingual-zh-en-2023-02-20.tar.bz2
|
// tar xvf sherpa-onnx-streaming-zipformer-bilingual-zh-en-2023-02-20.tar.bz2
|
||||||
// rm sherpa-onnx-streaming-zipformer-bilingual-zh-en-2023-02-20.tar.bz2
|
// rm sherpa-onnx-streaming-zipformer-bilingual-zh-en-2023-02-20.tar.bz2
|
||||||
//
|
//
|
||||||
|
// # 1. Test on CPU, run once
|
||||||
|
//
|
||||||
|
// ./bin/streaming-zipformer-rtf-cxx-api
|
||||||
|
//
|
||||||
|
// # 2. Test on CPU, run 10 times
|
||||||
|
//
|
||||||
|
// ./bin/streaming-zipformer-rtf-cxx-api 10
|
||||||
|
//
|
||||||
|
// # 3. Test on GPU, run 10 times
|
||||||
|
//
|
||||||
|
// ./bin/streaming-zipformer-rtf-cxx-api 10 cuda
|
||||||
|
//
|
||||||
// clang-format on
|
// clang-format on
|
||||||
|
|
||||||
#include <chrono> // NOLINT
|
#include <chrono> // NOLINT
|
||||||
@@ -21,13 +39,15 @@
|
|||||||
|
|
||||||
int32_t main(int argc, char *argv[]) {
|
int32_t main(int argc, char *argv[]) {
|
||||||
int32_t num_runs = 1;
|
int32_t num_runs = 1;
|
||||||
if (argc == 2) {
|
if (argc >= 2) {
|
||||||
num_runs = atoi(argv[1]);
|
num_runs = atoi(argv[1]);
|
||||||
if (num_runs < 0) {
|
if (num_runs < 0) {
|
||||||
num_runs = 1;
|
num_runs = 1;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
bool use_gpu = (argc == 3);
|
||||||
|
|
||||||
using namespace sherpa_onnx::cxx; // NOLINT
|
using namespace sherpa_onnx::cxx; // NOLINT
|
||||||
OnlineRecognizerConfig config;
|
OnlineRecognizerConfig config;
|
||||||
|
|
||||||
@@ -50,6 +70,7 @@ int32_t main(int argc, char *argv[]) {
|
|||||||
"./sherpa-onnx-streaming-zipformer-bilingual-zh-en-2023-02-20/tokens.txt";
|
"./sherpa-onnx-streaming-zipformer-bilingual-zh-en-2023-02-20/tokens.txt";
|
||||||
|
|
||||||
config.model_config.num_threads = 1;
|
config.model_config.num_threads = 1;
|
||||||
|
config.model_config.provider = use_gpu ? "cuda" : "cpu";
|
||||||
|
|
||||||
std::cout << "Loading model\n";
|
std::cout << "Loading model\n";
|
||||||
OnlineRecognizer recongizer = OnlineRecognizer::Create(config);
|
OnlineRecognizer recongizer = OnlineRecognizer::Create(config);
|
||||||
|
|||||||
Reference in New Issue
Block a user