Skip to content

Remove _pipeline from linker and refactor CTE pipeline #1524

Remove _pipeline from linker and refactor CTE pipeline

Remove _pipeline from linker and refactor CTE pipeline #1524

name: Run splink example notebooks
on:
pull_request:
branches:
- master
- "**dev"
paths:
- "splink/**"
- "docs/demos/examples/**"
- "pyproject.toml"
workflow_dispatch:
jobs:
test-notebooks:
runs-on: ubuntu-latest
strategy:
matrix:
python-version: ["3.9"]
test-group: ["duckdb", "spark", "sqlite"]
steps:
#----------------------------------------------
# check-out repo and set-up python
#----------------------------------------------
- uses: actions/checkout@v3
- name: Set up Python ${{ matrix.python-version }}
uses: actions/setup-python@v4
with:
python-version: ${{ matrix.python-version }}
#----------------------------------------------
# -- save a few section by caching poetry --
#----------------------------------------------
- name: Load cached Poetry installation
uses: actions/cache@v2
with:
path: ~/.local # the path depends on the OS
key: poetry-0 # increment to reset cache
#----------------------------------------------
# ----- install & configure poetry -----
#----------------------------------------------
- name: Install Poetry
uses: snok/install-poetry@v1
with:
version: "1.7.0"
virtualenvs-create: true
virtualenvs-in-project: true
installer-parallel: true
#----------------------------------------------
# load cached venv if cache exists
#----------------------------------------------
- name: Load cached venv
id: cached-poetry-dependencies
uses: actions/cache@v2
with:
path: .venv
key: venv-demos-${{ runner.os }}-${{ steps.setup-python.outputs.python-version }}-${{ hashFiles('**/poetry.lock') }}-00
#----------------------------------------------
# install dependencies if cache does not exist
#----------------------------------------------
- name: Install dependencies
if: steps.cached-poetry-dependencies.outputs.cache-hit != 'true'
run: poetry install --no-interaction
#----------------------------------------------
# pip install additional requirements
#----------------------------------------------
- name: Install additional requirements from demos_requirements.txt
run: |
poetry run pip install -r docs/demos/data/demos_requirements.txt
#----------------------------------------------
# Modify the notebooks to reduce the amount of data processed and speed up the tests
#----------------------------------------------
- name: Modify Notebooks to reduce data size
run: |
python scripts/reduce_notebook_runtime.py
#----------------------------------------------
# Test each notebook group in parallel
#----------------------------------------------
- name: Test ${{ matrix.test-group }} example notebooks with pytest
run: |
source .venv/bin/activate
python -m pytest -vv --nbmake -n=auto --nbmake-kernel=python3 --durations=0 docs/demos/examples/${{ matrix.test-group }}/*ipynb