Merge pull request #3387 from swryan/2711_set_input_defaults #11
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
# Run OpenMDAO Tests | |
name: OpenMDAO Tests | |
on: | |
# Trigger on push or pull request events for the master branch | |
push: | |
branches: [ master ] | |
pull_request: | |
branches: [ master ] | |
# Allow running the workflow manually from the Actions tab | |
# All jobs are excluded by default, desired jobs must be selected | |
workflow_dispatch: | |
inputs: | |
run_name: | |
type: string | |
description: 'Name of workflow run as it will appear under Actions tab:' | |
required: false | |
default: "" | |
Ubuntu_Baseline: | |
type: boolean | |
description: 'Include Ubuntu Baseline in test matrix' | |
required: false | |
default: false | |
MacOS_Baseline: | |
type: boolean | |
description: 'Include MacOS Baseline in test matrix' | |
required: false | |
default: false | |
MacOS_ARM: | |
type: boolean | |
description: 'Include MacOS ARM in test matrix' | |
required: false | |
default: false | |
Windows_Baseline: | |
type: boolean | |
description: 'Include Windows Baseline in test matrix' | |
required: false | |
default: false | |
Ubuntu_Minimal: | |
type: boolean | |
description: 'Include Ubuntu Minimal in test matrix' | |
required: false | |
default: false | |
Ubuntu_Oldest: | |
type: boolean | |
description: 'Include Ubuntu Oldest in test matrix' | |
required: false | |
default: false | |
Build_Docs: | |
type: boolean | |
description: 'Build docs' | |
required: false | |
default: false | |
debug_enabled: | |
type: boolean | |
description: 'Run the build with tmate debugging enabled (https://github.com/marketplace/actions/debugging-with-tmate)' | |
required: false | |
default: false | |
run-name: ${{ inputs.run_name }} | |
permissions: {} | |
jobs: | |
tests: | |
timeout-minutes: 120 | |
strategy: | |
fail-fast: false | |
matrix: | |
include: | |
# test baseline versions on Ubuntu | |
- NAME: Ubuntu Baseline | |
OS: ubuntu-latest | |
PY: '3.12' | |
NUMPY: '1.26' | |
SCIPY: '1.14' | |
PETSc: '3.20' | |
PYOPTSPARSE: 'v2.11.0' | |
# PAROPT: true | |
SNOPT: '7.7' | |
OPTIONAL: '[all]' | |
BANDIT: true | |
PEP517: true | |
TESTS: true | |
EXCLUDE: ${{ github.event_name == 'workflow_dispatch' && ! inputs.Ubuntu_Baseline }} | |
# test baseline versions on MacOS | |
- NAME: MacOS Baseline | |
OS: macos-13 | |
PY: '3.12' | |
NUMPY: '1.26' | |
SCIPY: '1.13' | |
PETSc: '3.20' | |
PYOPTSPARSE: 'v2.11.0' | |
# PAROPT: true | |
SNOPT: '7.7' | |
OPTIONAL: '[docs,doe,jax,notebooks,test]' | |
TESTS: true | |
EXCLUDE: ${{ github.event_name == 'workflow_dispatch' && ! inputs.MacOS_Baseline }} | |
# test baseline versions on MacOS/ARM | |
- NAME: MacOS ARM | |
OS: macos-14 | |
PY: '3.12' | |
NUMPY: '1.26' | |
SCIPY: '1.14' | |
PETSc: '3.20' | |
# PYOPTSPARSE: 'v2.11.0' | |
# PAROPT: true | |
# SNOPT: '7.7' | |
OPTIONAL: '[all]' | |
TESTS: true | |
EXCLUDE: ${{ github.event_name == 'workflow_dispatch' && ! inputs.MacOS_ARM }} | |
# test minimal install | |
- NAME: Ubuntu Minimal | |
OS: ubuntu-latest | |
PY: '3.12' | |
NUMPY: '1.26' | |
SCIPY: '1.14' | |
OPTIONAL: '[test]' | |
TESTS: true | |
EXCLUDE: ${{ github.event_name == 'workflow_dispatch' && ! inputs.Ubuntu_Minimal }} | |
# test oldest supported versions | |
- NAME: Ubuntu Oldest | |
OS: ubuntu-latest | |
PY: '3.9' | |
NUMPY: '1.24' | |
SCIPY: '1.9' | |
OPENMPI: '4.0' | |
MPI4PY: '3.0' | |
PETSc: '3.13' | |
PYOPTSPARSE: 'v2.10.2' | |
SNOPT: '7.2' | |
OPTIONAL: '[all]' | |
TESTS: true | |
EXCLUDE: ${{ github.event_name == 'workflow_dispatch' && ! inputs.Ubuntu_Oldest }} | |
# build docs (baseline versions) | |
- NAME: Build Docs | |
OS: ubuntu-latest | |
PY: '3.11' | |
NUMPY: '1.26' | |
SCIPY: '1.13' | |
PETSc: '3.19' | |
PYOPTSPARSE: 'v2.11.0' | |
SNOPT: '7.7' | |
OPTIONAL: '[all]' | |
BUILD_DOCS: true | |
EXCLUDE: ${{ github.event_name == 'workflow_dispatch' && ! inputs.Build_Docs }} | |
runs-on: ${{ matrix.OS }} | |
name: ${{ matrix.NAME }} | |
defaults: | |
run: | |
shell: bash -l {0} | |
steps: | |
- name: Display run details | |
run: | | |
echo "=============================================================" | |
echo "Run #${GITHUB_RUN_NUMBER}" | |
echo "Run ID: ${GITHUB_RUN_ID}" | |
echo "Testing: ${GITHUB_REPOSITORY}" | |
echo "Triggered by: ${GITHUB_EVENT_NAME}" | |
echo "Initiated by: ${GITHUB_ACTOR}" | |
echo "=============================================================" | |
- name: Exit if this job was excluded | |
if: matrix.EXCLUDE | |
uses: actions/github-script@v7 | |
with: | |
script: core.setFailed('The ${{ matrix.NAME }} job was excluded from the run, exiting...'); | |
- name: Create SSH key | |
if: (matrix.SNOPT || matrix.BUILD_DOCS) | |
env: | |
SSH_PRIVATE_KEY: ${{ secrets.SSH_PRIVATE_KEY }} | |
SSH_KNOWN_HOSTS: ${{ secrets.SSH_KNOWN_HOSTS }} | |
run: | | |
mkdir -p ~/.ssh/ | |
echo "$SSH_PRIVATE_KEY" > ~/.ssh/id_rsa | |
sudo chmod 600 ~/.ssh/id_rsa | |
echo "$SSH_KNOWN_HOSTS" > ~/.ssh/known_hosts | |
- name: Checkout code | |
uses: actions/checkout@v4 | |
- name: Setup conda | |
uses: conda-incubator/setup-miniconda@v3 | |
with: | |
python-version: ${{ matrix.PY }} | |
channels: conda-forge | |
- name: Install OpenMDAO | |
run: | | |
echo "Make sure we are not using anaconda packages" | |
conda config --remove channels defaults | |
conda install numpy=${{ matrix.NUMPY }} scipy=${{ matrix.SCIPY }} -q -y | |
python -m pip install --upgrade pip | |
echo "=============================================================" | |
echo "Install OpenMDAO" | |
echo "=============================================================" | |
if [[ "${{ matrix.PEP517 }}" == "true" ]]; then | |
pip wheel --no-deps --use-pep517 . | |
WHEEL=`find openmdao-*.whl` | |
echo "-----------------------------------------------------------" | |
echo "Installing from wheel: $WHEEL" | |
echo "-----------------------------------------------------------" | |
python -m pip install $WHEEL${{ matrix.OPTIONAL }} | |
else | |
python -m pip install .${{ matrix.OPTIONAL }} | |
fi | |
- name: Install MacOS-specific dependencies | |
if: matrix.OS == 'macos-13' | |
run: | | |
conda install swig | |
- name: Install PETSc | |
if: matrix.PETSc | |
run: | | |
echo "=============================================================" | |
echo "Install compilers for PETSc" | |
echo "=============================================================" | |
if [[ "${{ matrix.OPENMPI }}" ]]; then | |
COMPILERS="cython compilers openmpi-mpicc=${{ matrix.OPENMPI }}" | |
else | |
COMPILERS="cython compilers openmpi-mpicc" | |
fi | |
conda install $COMPILERS -q -y | |
echo "=============================================================" | |
echo "Install PETSc" | |
echo "=============================================================" | |
if [[ "${{ matrix.PETSc }}" == "3" ]]; then | |
python -m pip install git+https://github.com/mpi4py/mpi4py | |
python -m pip install petsc petsc4py | |
else | |
if [[ "${{ matrix.MPI4PY }}" ]]; then | |
conda install mpi4py=${{ matrix.MPI4PY }} petsc4py=${{ matrix.PETSc }} -q -y | |
else | |
conda install mpi4py petsc4py=${{ matrix.PETSc }} -q -y | |
fi | |
fi | |
echo "=============================================================" | |
echo "Check MPI and PETSc installation" | |
echo "=============================================================" | |
export OMPI_MCA_rmaps_base_oversubscribe=1 | |
export OMPI_MCA_btl=^openib | |
echo "-----------------------" | |
echo "Quick test of mpi4py:" | |
mpirun -n 3 python -c "from mpi4py import MPI; print(f'Rank: {MPI.COMM_WORLD.rank}')" | |
echo "-----------------------" | |
echo "Quick test of petsc4py:" | |
mpirun -n 3 python -c "import numpy; from mpi4py import MPI; comm = MPI.COMM_WORLD; \ | |
import petsc4py; petsc4py.init(); \ | |
x = petsc4py.PETSc.Vec().createWithArray(numpy.ones(5)*comm.rank, comm=comm); \ | |
print(x.getArray())" | |
echo "-----------------------" | |
echo "OMPI_MCA_rmaps_base_oversubscribe=1" >> $GITHUB_ENV | |
echo "OMPI_MCA_btl=^openib" >> $GITHUB_ENV | |
echo "Workaround for intermittent failures with OMPI https://github.com/open-mpi/ompi/issues/7393" | |
echo "TMPDIR=/tmp" >> $GITHUB_ENV | |
- name: Install pyOptSparse | |
if: matrix.PYOPTSPARSE | |
run: | | |
echo "=============================================================" | |
echo "Install pyoptsparse" | |
echo "=============================================================" | |
if [[ "${{ matrix.PYOPTSPARSE }}" == "conda-forge" ]]; then | |
if [[ "${{ matrix.SNOPT }}" ]]; then | |
echo "SNOPT ${{ matrix.SNOPT }} was requested but is not available on conda-forge" | |
fi | |
conda install -c conda-forge pyoptsparse | |
else | |
python -m pip install git+https://github.com/OpenMDAO/build_pyoptsparse | |
if [[ "${{ matrix.PYOPTSPARSE }}" == "latest" ]]; then | |
function latest_version() { | |
local REPO_URL=$1/releases/latest | |
local LATEST_URL=`curl -fsSLI -o /dev/null -w %{url_effective} $REPO_URL` | |
local LATEST_VER=`echo $LATEST_URL | awk '{split($0,a,"/tag/"); print a[2]}'` | |
echo $LATEST_VER | |
} | |
BRANCH="-b $(latest_version https://github.com/mdolab/pyoptsparse)" | |
else | |
BRANCH="-b ${{ matrix.PYOPTSPARSE }}" | |
fi | |
if [[ "${{ matrix.PAROPT }}" ]]; then | |
PAROPT="-a" | |
fi | |
if [[ "${{ matrix.SNOPT }}" == "7.7" && "${{ secrets.SNOPT_LOCATION_77 }}" ]]; then | |
echo " > Secure copying SNOPT 7.7 over SSH" | |
mkdir SNOPT | |
scp -qr ${{ secrets.SNOPT_LOCATION_77 }} SNOPT | |
SNOPT="-s SNOPT/src" | |
elif [[ "${{ matrix.SNOPT }}" == "7.2" && "${{ secrets.SNOPT_LOCATION_72 }}" ]]; then | |
echo " > Secure copying SNOPT 7.2 over SSH" | |
mkdir SNOPT | |
scp -qr ${{ secrets.SNOPT_LOCATION_72 }} SNOPT | |
SNOPT="-s SNOPT/source" | |
elif [[ "${{ matrix.SNOPT }}" ]]; then | |
echo "SNOPT version ${{ matrix.SNOPT }} was requested but source is not available" | |
fi | |
build_pyoptsparse -v $BRANCH $PAROPT $SNOPT | |
fi | |
- name: Install optional dependencies | |
if: matrix.OPTIONAL == '[all]' | |
run: | | |
echo "=============================================================" | |
echo "Install additional packages for testing/coverage" | |
echo "=============================================================" | |
conda install graphviz | |
python -m pip install pyparsing psutil objgraph pyxdsm pydot | |
echo "Pre-install playwright dependencies to avoid 'Playwright Host validation warning'" | |
playwright install --with-deps | |
- name: Display environment info | |
id: env_info | |
continue-on-error: true | |
run: | | |
conda info | |
conda list | |
echo "=============================================================" | |
echo "Check installed versions of Python, Numpy and Scipy" | |
echo "=============================================================" | |
echo 'errors<<EOF' >> $GITHUB_OUTPUT | |
FINAL_VER=`python -c "import platform; print(platform.python_version())"` | |
if [[ ! "$FINAL_VER" == "${{ matrix.PY }}"* ]]; then | |
echo "Python version was changed from ${{ matrix.PY }} to $FINAL_VER" >> $GITHUB_OUTPUT | |
fi | |
FINAL_VER=`python -c "import numpy; print(numpy.__version__)"` | |
if [[ ! "$FINAL_VER" == "${{ matrix.NUMPY }}"* ]]; then | |
echo "NumPy version was changed from ${{ matrix.NUMPY }} to $FINAL_VER" >> $GITHUB_OUTPUT | |
fi | |
FINAL_VER=`python -c "import scipy; print(scipy.__version__)"` | |
if [[ ! "$FINAL_VER" == "${{ matrix.SCIPY }}"* ]]; then | |
echo "SciPy version was changed from ${{ matrix.SCIPY }} to $FINAL_VER" >> $GITHUB_OUTPUT | |
fi | |
echo 'EOF' >> $GITHUB_OUTPUT | |
grep changed $GITHUB_OUTPUT || echo "" | |
# Enable tmate debugging of manually-triggered workflows if the input option was provided | |
# | |
# To access the terminal through the web-interface: | |
# 1. Click on the web-browser link printed out in this action from the github | |
# workflow terminal | |
# 2. Press cntrl + c in the new tab that opens up to reveal the terminal | |
# 3. To activate the conda environment run: | |
# $ source $CONDA/etc/profile.d/conda.sh | |
# $ conda activate test | |
- name: Setup tmate session | |
if: ${{ inputs.debug_enabled }} | |
uses: mxschmitt/action-tmate@v3 | |
with: | |
limit-access-to-actor: true | |
- name: Run tests | |
id: run_tests | |
if: matrix.TESTS | |
env: | |
OPENMDAO_CHECK_ALL_PARTIALS: true | |
run: | | |
echo "=============================================================" | |
echo "Run tests with coverage (from directory other than repo root)" | |
echo "=============================================================" | |
echo "############ .coveragerc contents #################" | |
cat .coveragerc | |
echo "###################################################" | |
cp .coveragerc $HOME | |
cd $HOME | |
testflo -n 2 openmdao --timeout=240 --show_skipped --coverage --coverpkg openmdao --durations=20 | |
- name: Submit coverage | |
id: coveralls | |
if: matrix.TESTS | |
continue-on-error: true | |
env: | |
COVERALLS_REPO_TOKEN: ${{ secrets.GITHUB_TOKEN }} | |
COVERALLS_SERVICE_NAME: "github" | |
COVERALLS_PARALLEL: true | |
run: | | |
echo "=============================================================" | |
echo "Submit coverage" | |
echo "=============================================================" | |
cp $HOME/.coverage . | |
python -m pip install coveralls | |
SITE_DIR=`python -c 'import site; print(site.getsitepackages()[-1])'` | |
coveralls --basedir $SITE_DIR | |
- name: Slack failure to upload to coveralls.io | |
if: steps.coveralls.outcome == 'failure' | |
uses: act10ns/[email protected] | |
with: | |
webhook-url: ${{ secrets.SLACK_WEBHOOK_URL }} | |
status: 'warning' | |
message: | | |
Uploading of coverage data to coveralls.io failed. | |
${{ github.server_url }}/${{ github.repository }}/actions/runs/${{ github.run_id }} | |
- name: Build docs | |
if: matrix.BUILD_DOCS | |
id: build_docs | |
run: | | |
export OPENMDAO_REPORTS=0 | |
export PYDEVD_DISABLE_FILE_VALIDATION=1 | |
cd openmdao/docs | |
if [[ "${{ secrets.SNOPT_LOCATION_72 }}" || "${{ secrets.SNOPT_LOCATION_77 }}" ]]; then | |
echo "=============================================================" | |
echo "Building docs with SNOPT examples." | |
echo "=============================================================" | |
else | |
echo "=============================================================" | |
echo "Disabling SNOPT cells in notebooks." | |
echo "=============================================================" | |
python openmdao_book/other/disable_snopt_cells.py | |
fi | |
# start ipcluster to run MPI under notebooks | |
./ipcluster_start.sh | |
sleep 12 | |
echo "=============================================================" | |
echo "Build the docs" | |
echo "=============================================================" | |
python build_source_docs.py | |
jupyter-book build -W --keep-going openmdao_book | |
python copy_build_artifacts.py | |
- name: Display doc build reports | |
if: failure() && matrix.BUILD_DOCS && steps.build_docs.outcome == 'failure' | |
run: | | |
for f in $(find /home/runner/work/OpenMDAO/OpenMDAO/openmdao/docs/openmdao_book/_build/html/reports -name '*.log'); do | |
echo "=============================================================" | |
echo $f | |
echo "=============================================================" | |
cat $f | |
done | |
- name: Publish docs | |
if: ${{ github.event_name != 'pull_request' && matrix.BUILD_DOCS }} | |
env: | |
DOCS_LOCATION: ${{ secrets.DOCS_LOCATION }} | |
run: | | |
if [[ "${#DOCS_LOCATION}" ]]; then | |
echo "=============================================================" | |
echo "Install version of openssl compatible with hosting service" | |
echo "=============================================================" | |
conda install -c conda-forge 'openssl=3.0' | |
echo "=============================================================" | |
echo "Fetch tags to get docs version" | |
echo "=============================================================" | |
git fetch --prune --unshallow --tags | |
echo "=============================================================" | |
echo "Publish docs" | |
echo "=============================================================" | |
cd openmdao/docs | |
python upload_doc_version.py openmdao_book/_build/html/ ${{ secrets.DOCS_LOCATION }} | |
else | |
echo "Docs destination not available." | |
fi | |
- name: Scan for security issues | |
if: matrix.BANDIT | |
id: bandit | |
continue-on-error: true | |
run: | | |
python -m pip install bandit | |
echo "=============================================================" | |
echo "Run bandit scan for high/medium severity issues" | |
echo "=============================================================" | |
cd $GITHUB_WORKSPACE | |
python -m bandit -c bandit.yml -ll -r openmdao | |
- name: Check NumPy 2.0 Compatibility | |
run: | | |
echo "=============================================================" | |
echo "Check OpenMDAO code for NumPy 2.0 compatibility" | |
echo "See: https://numpy.org/devdocs/numpy_2_0_migration_guide.html" | |
echo "=============================================================" | |
python -m pip install ruff | |
ruff check . --select NPY201 | |
- name: Perform linting with Ruff | |
if: ${{ matrix.NAME == 'Ubuntu Baseline' }} | |
run: | | |
echo "=============================================================" | |
echo "Lint OpenMDAO code per settings in pyproject.toml" | |
echo "=============================================================" | |
python -m pip install ruff | |
ruff check . --config pyproject.toml | |
- name: Slack env change | |
if: steps.env_info.outputs.errors != '' | |
uses: act10ns/[email protected] | |
with: | |
webhook-url: ${{ secrets.SLACK_WEBHOOK_URL }} | |
status: 'warning' | |
message: | | |
Environment change detected on `${{ matrix.NAME }}` build. | |
Python, NumPy or SciPy was not the requested version: | |
```${{steps.env_info.outputs.errors}}``` | |
${{ github.server_url }}/${{ github.repository }}/actions/runs/${{ github.run_id }} | |
- name: Slack unit test failure | |
if: failure() && steps.run_tests.outcome == 'failure' | |
uses: act10ns/[email protected] | |
with: | |
webhook-url: ${{ secrets.SLACK_WEBHOOK_URL }} | |
status: ${{ steps.run_tests.outcome }} | |
message: | |
Unit testing failed on `${{ matrix.NAME }}` build. | |
${{ github.server_url }}/${{ github.repository }}/actions/runs/${{ github.run_id }} | |
- name: Slack doc build failure | |
if: failure() && steps.build_docs.outcome == 'failure' | |
uses: act10ns/[email protected] | |
with: | |
webhook-url: ${{ secrets.SLACK_WEBHOOK_URL }} | |
status: ${{ steps.build_docs.outcome }} | |
message: | | |
Doc build failed on `${{ matrix.NAME }}` build. | |
${{ github.server_url }}/${{ github.repository }}/actions/runs/${{ github.run_id }} | |
- name: Slack security issue | |
if: steps.bandit.outcome == 'failure' | |
uses: act10ns/[email protected] | |
with: | |
webhook-url: ${{ secrets.SLACK_WEBHOOK_URL }} | |
status: ${{ steps.bandit.outcome }} | |
message: | |
Security issue found on `${{ matrix.NAME }}` build. | |
${{ github.server_url }}/${{ github.repository }}/actions/runs/${{ github.run_id }} | |
- name: Fail the workflow if tests or doc build failed | |
if: steps.run_tests.outcome == 'failure' || steps.build_docs.outcome == 'failure' | |
uses: actions/github-script@v3 | |
with: | |
script: | | |
let test_fail = ${{ steps.run_tests.outcome == 'failure' }}; | |
let docs_fail = ${{ steps.build_docs.outcome == 'failure' }}; | |
if (test_fail && docs_fail) { | |
core.setFailed('Tests and doc build failed.'); | |
} | |
else if (test_fail) { | |
core.setFailed('Tests failed.'); | |
} | |
else if (docs_fail) { | |
core.setFailed('Doc build failed.'); | |
} | |
windows_tests: | |
runs-on: windows-latest | |
timeout-minutes: 60 | |
strategy: | |
fail-fast: false | |
matrix: | |
include: | |
# baseline versions | |
- NAME: Windows Baseline | |
PY: '3.12' | |
NUMPY: '1.26' | |
SCIPY: '1.13' | |
PYOPTSPARSE: '2.11.0' | |
BANDIT: true | |
EXCLUDE: ${{ github.event_name == 'workflow_dispatch' && ! inputs.Windows_Baseline }} | |
name: ${{ matrix.NAME }} | |
defaults: | |
run: | |
shell: pwsh | |
steps: | |
- name: Display run details | |
run: | | |
echo "=============================================================" | |
echo "Run #$env:GITHUB_RUN_NUMBER" | |
echo "Run ID: $env:GITHUB_RUN_ID" | |
echo "Testing: $env:GITHUB_REPOSITORY" | |
echo "Triggered by: $env:GITHUB_EVENT_NAME" | |
echo "Initiated by: $env:GITHUB_ACTOR" | |
echo "=============================================================" | |
- name: Exit if this job was excluded | |
if: matrix.EXCLUDE | |
uses: actions/github-script@v7 | |
with: | |
script: core.setFailed('The ${{ matrix.NAME }} job was excluded from the run, exiting...'); | |
- name: Checkout code | |
uses: actions/checkout@v4 | |
- name: Setup conda | |
uses: conda-incubator/setup-miniconda@v3 | |
with: | |
python-version: ${{ matrix.PY }} | |
channels: conda-forge | |
- name: Install OpenDMAO | |
run: | | |
conda install numpy=${{ matrix.NUMPY }} scipy=${{ matrix.SCIPY }} -q -y | |
python -m pip install --upgrade pip | |
echo "=============================================================" | |
echo "Install mkl 2024.1.0 (test failure with 2024.2.2)" | |
echo "=============================================================" | |
conda install mkl=2024.1.0 | |
echo "=============================================================" | |
echo "Install OpenMDAO" | |
echo "=============================================================" | |
python -m pip install .[all] | |
- name: Install pyOptSparse | |
if: matrix.PYOPTSPARSE | |
run: | | |
echo "=============================================================" | |
echo "Install pyoptsparse from conda-forge" | |
echo "=============================================================" | |
conda install -c conda-forge pyoptsparse=${{ matrix.PYOPTSPARSE }} -q -y | |
- name: Install optional dependencies | |
run: | | |
echo "=============================================================" | |
echo "Install additional packages for testing/coverage" | |
echo "=============================================================" | |
python -m pip install psutil objgraph git+https://github.com/mdolab/pyxdsm | |
- name: Display environment info | |
run: | | |
conda info | |
conda list | |
echo "=============================================================" | |
echo "Check installed versions of Python, Numpy and Scipy" | |
echo "=============================================================" | |
python -c "import sys; assert str(sys.version).startswith(str(${{ matrix.PY }})), \ | |
f'Python version {sys.version} is not the requested version (${{ matrix.PY }})'" | |
python -c "import numpy; assert str(numpy.__version__).startswith(str(${{ matrix.NUMPY }})), \ | |
f'Numpy version {numpy.__version__} is not the requested version (${{ matrix.NUMPY }})'" | |
python -c "import scipy; assert str(scipy.__version__).startswith(str(${{ matrix.SCIPY }})), \ | |
f'Scipy version {scipy.__version__} is not the requested version (${{ matrix.SCIPY }})'" | |
- name: Run tests | |
env: | |
OPENMDAO_CHECK_ALL_PARTIALS: true | |
run: | | |
echo "=============================================================" | |
echo "Run tests with coverage (from directory other than repo root)" | |
echo "=============================================================" | |
copy .coveragerc $HOME | |
cd $HOME | |
Set-DisplayResolution -Width 1920 -Height 1080 -Force | |
testflo -n 2 openmdao --timeout=240 --show_skipped --coverage --coverpkg openmdao --durations=20 | |
- name: Slack unit test failure | |
if: failure() && steps.run_tests.outcome == 'failure' | |
uses: act10ns/[email protected] | |
with: | |
webhook-url: ${{ secrets.SLACK_WEBHOOK_URL }} | |
status: ${{ steps.run_tests.outcome }} | |
message: | |
Unit testing failed on `${{ matrix.NAME }}` build. | |
${{ github.server_url }}/${{ github.repository }}/actions/runs/${{ github.run_id }} | |
- name: Submit coverage | |
id: coveralls | |
continue-on-error: true | |
env: | |
COVERALLS_REPO_TOKEN: ${{ secrets.GITHUB_TOKEN }} | |
COVERALLS_SERVICE_NAME: "github" | |
COVERALLS_PARALLEL: true | |
run: | | |
echo "=============================================================" | |
echo "Submit coverage" | |
echo "=============================================================" | |
copy $HOME\.coverage . | |
python -m pip install coveralls | |
$SITE_DIR=python -c "import site; print(site.getsitepackages()[-1].replace('lib\\site-', 'Lib\\site-'))" | |
coveralls --basedir $SITE_DIR | |
- name: Slack failure to upload to coveralls.io | |
if: steps.coveralls.outcome == 'failure' | |
uses: act10ns/[email protected] | |
with: | |
webhook-url: ${{ secrets.SLACK_WEBHOOK_URL }} | |
status: 'warning' | |
message: | | |
Uploading of coverage data to coveralls.io failed. | |
${{ github.server_url }}/${{ github.repository }}/actions/runs/${{ github.run_id }} | |
- name: Scan for security issues | |
if: matrix.BANDIT | |
id: bandit | |
run: | | |
python -m pip install bandit | |
echo "=============================================================" | |
echo "Run bandit scan for high/medium severity issues" | |
echo "=============================================================" | |
cd ${{ github.workspace }} | |
python -m bandit -c bandit.yml -ll -r openmdao | |
coveralls: | |
name: Finish coverage | |
needs: [tests, windows_tests] | |
runs-on: ubuntu-latest | |
continue-on-error: true | |
steps: | |
- uses: coverallsapp/github-action@master | |
with: | |
github-token: ${{ secrets.GITHUB_TOKEN }} | |
parallel-finished: true | |
related_tests: | |
name: Run Dymos & pyCycle Tests | |
needs: [tests, windows_tests] | |
runs-on: ubuntu-latest | |
steps: | |
- uses: benc-uk/workflow-dispatch@v1 | |
with: | |
workflow: Dymos Tests | |
repo: ${{ github.repository_owner }}/dymos | |
inputs: > | |
{ | |
"run_name": "Test Latest OpenMDAO Development Version", | |
"latest": true | |
} | |
token: ${{ secrets.ACCESS_TOKEN }} | |
if: github.event_name == 'push' | |
- uses: benc-uk/workflow-dispatch@v1 | |
with: | |
workflow: pyCycle Tests | |
repo: ${{ github.repository_owner }}/pycycle | |
inputs: > | |
{ | |
"run_name": "Test Latest OpenMDAO Development Version", | |
"Ubuntu_Baseline": false, | |
"MacOS_Baseline": false, | |
"Windows_Baseline": false, | |
"OpenMDAO_Dev": true | |
} | |
token: ${{ secrets.ACCESS_TOKEN }} | |
if: github.event_name == 'push' |