Enhancements to PIP install scripts (#2219) #9
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
# Copyright (c) Meta Platforms, Inc. and affiliates. | |
# All rights reserved. | |
# This source code is licensed under the BSD-style license found in the | |
# LICENSE file in the root directory of this source tree. | |
# This workflow is used for FBGEMM_GPU-CUDA CI as well as nightly builds of | |
# FBGEMM_GPU-CUDA against PyTorch-CUDA Nightly. | |
name: FBGEMM_GPU-CUDA CI | |
on: | |
# PR Trigger (enabled for regression checks and debugging) | |
# | |
pull_request: | |
branches: | |
- main | |
# Push Trigger (enable to catch errors coming out of multiple merges) | |
# | |
push: | |
branches: | |
- main | |
# Cron Trigger (UTC) | |
# | |
# Based on the Conda page for PyTorch-nightly, the GPU nightly releases appear | |
# around 02:30 PST every day (roughly 2 hours after the CPU releases) | |
# | |
schedule: | |
- cron: '45 12 * * *' | |
# Manual Trigger | |
# | |
workflow_dispatch: | |
inputs: | |
publish_to_pypi: | |
description: Publish Artifact to PyPI | |
type: boolean | |
required: false | |
default: false | |
concurrency: | |
# Cancel previous runs in the PR if a new commit is pushed | |
group: ${{ github.workflow }}-${{ github.event.pull_request.number || github.ref }} | |
cancel-in-progress: true | |
jobs: | |
# Build on CPU hosts and upload to GHA | |
build_artifact: | |
runs-on: ${{ matrix.host-machine.instance }} | |
container: | |
image: amazonlinux:2023 | |
options: --user root | |
defaults: | |
run: | |
shell: bash | |
env: | |
PRELUDE: .github/scripts/setup_env.bash | |
BUILD_ENV: build_binary | |
continue-on-error: true | |
strategy: | |
# Don't fast-fail all the other builds if one of the them fails | |
fail-fast: false | |
matrix: | |
host-machine: [ | |
{ arch: x86, instance: "linux.24xlarge" }, | |
] | |
python-version: [ "3.8", "3.9", "3.10", "3.11", "3.12" ] | |
cuda-version: [ "11.8.0", "12.1.1" ] | |
steps: | |
- name: Setup Build Container | |
run: yum update -y; yum install -y binutils findutils git pciutils sudo tar wget which | |
- name: Checkout the Repository | |
uses: actions/checkout@v4 | |
with: | |
submodules: true | |
- name: Display System Info | |
run: . $PRELUDE; print_system_info | |
- name: Display GPU Info | |
run: . $PRELUDE; print_gpu_info | |
- name: Setup Miniconda | |
run: . $PRELUDE; setup_miniconda $HOME/miniconda | |
- name: Create Conda Environment | |
run: . $PRELUDE; create_conda_environment $BUILD_ENV ${{ matrix.python-version }} | |
- name: Install C/C++ Compilers | |
run: . $PRELUDE; install_cxx_compiler $BUILD_ENV | |
- name: Install Build Tools | |
run: . $PRELUDE; install_build_tools $BUILD_ENV | |
- name: Install CUDA | |
run: . $PRELUDE; install_cuda $BUILD_ENV ${{ matrix.cuda-version }} | |
# Install via PIP to avoid defaulting to the CPU variant if the GPU variant of the day is not ready | |
- name: Install PyTorch Nightly | |
run: . $PRELUDE; install_pytorch_pip $BUILD_ENV nightly cuda ${{ matrix.cuda-version }} | |
- name: Collect PyTorch Environment Info | |
if: ${{ success() || failure() }} | |
run: . $PRELUDE; collect_pytorch_env_info $BUILD_ENV | |
- name: Install cuDNN | |
run: . $PRELUDE; install_cudnn $BUILD_ENV "$(pwd)/build_only/cudnn" ${{ matrix.cuda-version }} | |
- name: Prepare FBGEMM_GPU Build | |
run: . $PRELUDE; cd fbgemm_gpu; prepare_fbgemm_gpu_build $BUILD_ENV | |
- name: Build FBGEMM_GPU Wheel | |
run: . $PRELUDE; cd fbgemm_gpu; build_fbgemm_gpu_package $BUILD_ENV nightly cuda | |
- name: Upload Built Wheel as GHA Artifact | |
uses: actions/upload-artifact@v3 | |
with: | |
name: fbgemm_gpu_nightly_cuda_${{ matrix.host-machine.arch }}_${{ matrix.python-version }}_cu${{ matrix.cuda-version }}.whl | |
path: fbgemm_gpu/dist/fbgemm_gpu_nightly-*.whl | |
# Download the built artifact from GHA, test on GPU, and push to PyPI | |
test_and_publish_artifact: | |
# runs-on: linux.4xlarge.nvidia.gpu | |
# Use available instance types - https://github.com/pytorch/test-infra/blob/main/.github/scale-config.yml | |
runs-on: ${{ matrix.host-machine.instance }} | |
defaults: | |
run: | |
shell: bash | |
env: | |
PRELUDE: .github/scripts/setup_env.bash | |
BUILD_ENV: build_binary | |
ENFORCE_NVIDIA_GPU: 1 | |
strategy: | |
fail-fast: false | |
matrix: | |
host-machine: [ | |
{ arch: x86, instance: "linux.g5.4xlarge.nvidia.gpu" }, | |
] | |
python-version: [ "3.8", "3.9", "3.10", "3.11", "3.12" ] | |
cuda-version: [ "11.8.0", "12.1.1" ] | |
# Specify exactly ONE CUDA version for artifact publish | |
cuda-version-publish: [ "12.1.1" ] | |
needs: build_artifact | |
steps: | |
# Cannot upgrade to actions/checkout@v4 yet because GLIBC on the instance is too old | |
- name: Checkout the Repository | |
uses: actions/checkout@v3 | |
with: | |
submodules: true | |
- name: Download Wheel Artifact from GHA | |
uses: actions/download-artifact@v3 | |
with: | |
name: fbgemm_gpu_nightly_cuda_${{ matrix.host-machine.arch }}_${{ matrix.python-version }}_cu${{ matrix.cuda-version }}.whl | |
# Use PyTorch test infrastructure action - https://github.com/pytorch/test-infra/blob/main/.github/actions/setup-nvidia/action.yml | |
- name: Install NVIDIA Drivers and NVIDIA-Docker Runtime | |
uses: pytorch/test-infra/.github/actions/setup-nvidia@main | |
- name: Display System Info | |
run: . $PRELUDE; print_system_info; print_ec2_info | |
- name: Display GPU Info | |
run: . $PRELUDE; print_gpu_info | |
- name: Setup Miniconda | |
run: . $PRELUDE; setup_miniconda $HOME/miniconda | |
- name: Create Conda Environment | |
run: . $PRELUDE; create_conda_environment $BUILD_ENV ${{ matrix.python-version }} | |
- name: Install CUDA | |
run: . $PRELUDE; install_cuda $BUILD_ENV ${{ matrix.cuda-version }} | |
# Install via PIP to avoid defaulting to the CPU variant if the GPU variant of the day is not ready | |
- name: Install PyTorch Nightly | |
run: . $PRELUDE; install_pytorch_pip $BUILD_ENV nightly cuda ${{ matrix.cuda-version }} | |
- name: Collect PyTorch Environment Info | |
if: ${{ success() || failure() }} | |
run: . $PRELUDE; collect_pytorch_env_info $BUILD_ENV | |
- name: Prepare FBGEMM_GPU Build | |
run: . $PRELUDE; cd fbgemm_gpu; prepare_fbgemm_gpu_build $BUILD_ENV | |
- name: Install FBGEMM_GPU Wheel | |
run: . $PRELUDE; install_fbgemm_gpu_wheel $BUILD_ENV *.whl | |
- name: Test with PyTest | |
timeout-minutes: 15 | |
run: . $PRELUDE; cd fbgemm_gpu/test; run_fbgemm_gpu_tests $BUILD_ENV cuda | |
- name: Push Wheel to PyPI | |
if: ${{ (github.event_name == 'schedule' && matrix.cuda-version == matrix.cuda-version-publish) || (github.event_name == 'workflow_dispatch' && github.event.inputs.publish_to_pypi == 'true' && matrix.cuda-version == matrix.cuda-version-publish) }} | |
env: | |
PYPI_TOKEN: ${{ secrets.PYPI_TOKEN }} | |
run: . $PRELUDE; publish_to_pypi $BUILD_ENV fbgemm_gpu_nightly-*.whl "$PYPI_TOKEN" |