Skip to content

Chunk combined datasets to limit memory usage #214

Chunk combined datasets to limit memory usage

Chunk combined datasets to limit memory usage #214

Workflow file for this run

name: Python CI
on: push
jobs:
test:
runs-on: ubuntu-20.04
strategy:
matrix:
python-version: ["3.9", "3.10", "3.11"]
steps:
- uses: actions/checkout@v2
- name: Set up Python ${{ matrix.python-version }}
uses: actions/setup-python@v2
with:
python-version: ${{ matrix.python-version }}
- name: Install poetry
run: |
curl -sSL https://install.python-poetry.org | python3 -
echo "$HOME/.local/bin" >> $GITHUB_PATH
- id: cache-poetry
uses: actions/cache@v2
with:
path: ~/.cache/pypoetry/virtualenvs
key: ${{ runner.os }}-poetry-${{ hashFiles('**/poetry.lock') }}
- name: Install dependencies if changed
if: ${{ steps.cache-poetry.outputs.cache-hit != 'true' }}
run: |
poetry env use python${{ matrix.python-version }}
poetry install --with=dev
poetry run pip install -e .
- name: Test with pytest (full)
if: github.ref == 'refs/heads/master'
run: |
poetry run py.test -m "not online" -v
- name: Test with pytest (fast)
if: github.ref != 'refs/heads/master'
run: |
poetry run py.test -m "not online and not slow" -v