Skip to content

Added a sample integration for LLM models in huggingface #660

Added a sample integration for LLM models in huggingface

Added a sample integration for LLM models in huggingface #660

Workflow file for this run

# This workflow will install Python dependencies, run tests and lint with a single version of Python
# For more information see: https://help.github.com/actions/language-and-framework-guides/using-python-with-github-actions
name: build
on:
push:
branches: [ main ]
pull_request:
branches: [ main ]
jobs:
test:
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v2
- name: Set up Python 3.8
uses: actions/setup-python@v2
with:
python-version: "3.8"
- name: Setup Dev Environment
run: |
pip install virtualenv
virtualenv .venv
source .venv/bin/activate
python3 -m pip install nvidia-cuda-runtime-cu12
./run setup
- name: Check formatting
run: |
source .venv/bin/activate
python3 -m pip install nvidia-cuda-runtime-cu12
python3 -c 'import sys; print(sys.executable)'
python3 -c 'import site; print(site.getsitepackages())'
python3 -m pip freeze
export CUDA_WHL_LIB_DIR=$(python3 -c 'import nvidia.cuda_runtime; print(nvidia.cuda_runtime.__path__[0])')/lib
export LD_LIBRARY_PATH="$CUDA_WHL_LIB_DIR:$LD_LIBRARY_PATH"
python3 -c 'from holoscan.core import *'
./run check -f
- name: Run Unit tests
run: |
source .venv/bin/activate
python3 -m pip install nvidia-cuda-runtime-cu12
export CUDA_WHL_LIB_DIR=$(python3 -c 'import nvidia.cuda_runtime; print(nvidia.cuda_runtime.__path__[0])')/lib
export LD_LIBRARY_PATH="$CUDA_WHL_LIB_DIR:$LD_LIBRARY_PATH"
./run test all unit
- name: Coverage
run: |
source .venv/bin/activate
python3 -m pip install nvidia-cuda-runtime-cu12
export CUDA_WHL_LIB_DIR=$(python3 -c 'import nvidia.cuda_runtime; print(nvidia.cuda_runtime.__path__[0])')/lib
export LD_LIBRARY_PATH="$CUDA_WHL_LIB_DIR:$LD_LIBRARY_PATH"
coverage xml
- name: Upload coverage
uses: codecov/codecov-action@v2
with:
fail_ci_if_error: false
files: ./coverage.xml