Skip to content

remove invalid trino configs #4413

remove invalid trino configs

remove invalid trino configs #4413

Workflow file for this run

# NOTE: This name appears in GitHub's Checks API and in workflow's status badge.
name: ci-default
env:
# CI variables
DOCKER_PLATFORM: "amd64"
TURBO_TEAM: ${{ secrets.TURBO_TEAM }}
TURBO_TOKEN: ${{ secrets.TURBO_TOKEN }}
# Frontend variables
NODE_ENV: ${{ vars.NODE_ENV }}
PLASMIC_PROJECT_ID: ${{ vars.PLASMIC_PROJECT_ID }}
PLASMIC_PROJECT_API_TOKEN: ${{ vars.PLASMIC_PROJECT_API_TOKEN }}
NEXT_PUBLIC_DOMAIN: "www.opensource.observer"
NEXT_PUBLIC_DB_GRAPHQL_URL: ${{ vars.NEXT_PUBLIC_DB_GRAPHQL_URL }}
HASURA_URL: ${{ vars.HASURA_URL }}
OSO_API_KEY: ${{ secrets.OSO_API_KEY }}
# Docs variables
DOCS_URL: "https://docs.opensource.observer"
DOCS_ALGOLIA_APP_ID: "test"
DOCS_ALGOLIA_API_KEY: "test"
DOCS_ALGOLIA_INDEX: "test"
DOCS_SEGMENT_WRITE_KEY: "test"
# Hasura variables
DBT_TARGET: "production"
# Google variables
GOOGLE_PROJECT_ID: "opensource-observer"
GOOGLE_TEST_DUMMY_CREDENTIALS_JSON: ${{ vars.GOOGLE_TEST_DUMMY_CREDENTIALS_JSON }}
CLICKHOUSE_URL: "http://127.0.0.1:8123" # dummy value
CLICKHOUSE_USERNAME: "username"
CLICKHOUSE_PASSWORD: "password"
# should not be set to a legitimate value for testing. This will use up API
# quota otherwise
DUNE_API_KEY: "none"
# Trigger the workflow when:
on:
# A push occurs to one of the matched branches.
push:
branches:
- main
# Or when a pull request event occurs for a pull request against one of the
# matched branches.
pull_request:
branches:
- main
# Allows you to run this workflow manually from the Actions tab
workflow_dispatch:
merge_group:
# Cancel in progress jobs on new pushes.
concurrency:
group: ${{ github.workflow }}-${{ github.ref }}
cancel-in-progress: true
jobs:
lint-and-test:
# NOTE: This name appears in GitHub's Checks API.
name: test
environment: testing
runs-on: ubuntu-latest
strategy:
matrix:
component: ["node", "python", "dbt", "sqlmesh"]
steps:
- name: Checkout code
uses: actions/checkout@v4
with:
# Check out pull request's HEAD commit instead of the merge commit to
# prevent gitlint from failing due to too long commit message titles,
# e.g. "Merge 3e621938d65caaa67f8e35d145335d889d470fc8 into 19a39b2f66cd7a165082d1486b2f1eb36ec2354a".
ref: ${{ github.event.pull_request.head.sha }}
# Fetch all history so gitlint can check the relevant commits.
fetch-depth: "0"
- name: Load public vars
run: |
bash .github/scripts/load-public-vars.sh $GITHUB_ENV \
NODE_ENV \
PLASMIC_PROJECT_ID \
PLASMIC_PROJECT_API_TOKEN \
NEXT_PUBLIC_DOMAIN \
NEXT_PUBLIC_DB_GRAPHQL_URL \
HASURA_URL \
OSO_API_KEY \
GOOGLE_PROJECT_ID \
GOOGLE_TEST_DUMMY_CREDENTIALS_JSON \
PUBLIC_VARS_TEST
- name: Check if algolia is empty, set the variable to some dummy value if it is
shell: bash
run: |
echo "NEXT_PUBLIC_ALGOLIA_API_KEY=${NEXT_PUBLIC_ALGOLIA_API_KEY:-xxxx}" >> "$GITHUB_ENV"
- name: Setup pnpm
uses: pnpm/action-setup@v4
with:
version: 9
run_install: |
- recursive: true
args: [--frozen-lockfile, --strict-peer-dependencies]
- name: Set up Node.js
uses: actions/setup-node@v4
with:
node-version: "20.x"
cache: "pnpm"
- name: "Setup Python, Poetry and Dependencies"
uses: packetcoders/action-setup-cache-python-poetry@main
with:
python-version: 3.12
poetry-version: 1.8.2
- name: Authenticate to google with a test-dummy user
uses: "google-github-actions/auth@v2"
with:
credentials_json: "${{ env.GOOGLE_TEST_DUMMY_CREDENTIALS_JSON }}"
create_credentials_file: true
- name: Setup dbt profile
run: |
bash .github/scripts/create-dbt-profile.sh ${GOOGLE_APPLICATION_CREDENTIALS}
- name: Install Foundry
uses: foundry-rs/foundry-toolchain@v1
with:
version: nightly-87bc53fc6c874bd4c92d97ed180b949e3a36d78c
if: ${{ matrix.component == 'node' }}
- name: Run supabase local
id: supabase
run: |
bash .github/scripts/run-supabase-local.sh apps/frontend
if: ${{ matrix.component == 'node' }}
# Automatically retry if the build fails
- name: Build (Node)
uses: nick-fields/retry@v3
with:
timeout_minutes: 5
max_attempts: 3
command: |
pnpm build
# Always run this step so that all linting errors can be seen at once.
if: ${{ matrix.component == 'node' }}
- name: Lint (Node)
run: |
pnpm lint
# Always run this step so that all linting errors can be seen at once.
if: ${{ always() && steps.supabase.conclusion == 'success' && matrix.component == 'node' }}
- name: Test (Node)
run: |
pnpm test
# Always run this step so that all linting errors can be seen at once.
if: ${{ always() && steps.supabase.conclusion == 'success' && matrix.component == 'node' }}
# Temporarily include this separately. Eventually this should all be part of lint
- name: Lint SQL (dbt)
run: |
poetry run sqlfluff lint --dialect bigquery ./warehouse/dbt/models
if: ${{ always() && matrix.component == 'dbt' }}
- name: Type Check (Python)
run: |
pnpm pyright
if: ${{ always() && matrix.component == 'python' }}
- name: Lint (Python)
run: |
poetry run ruff check
if: ${{ always() && matrix.component == 'python' }}
- name: Test (Python)
run: |
poetry run pytest
if: ${{ always() && matrix.component == 'python' }}
- name: Test SQLMesh
run: |
poetry install && cd warehouse/metrics_mesh && poetry run -C ../../ sqlmesh test
if: ${{ always() && matrix.component == 'sqlmesh' }}