Skip to content

[INFRA] [DO NOT MERGE] Test publishing against Spark 4 #3796

[INFRA] [DO NOT MERGE] Test publishing against Spark 4

[INFRA] [DO NOT MERGE] Test publishing against Spark 4 #3796

name: "Delta Connectors Tests"
on: [push, pull_request]
jobs:
build:
name: "Run tests"
runs-on: ubuntu-20.04
strategy:
matrix:
# These Scala versions must match those in the build.sbt
scala: [2.13.13]
steps:
- uses: actions/checkout@v2
- name: install java
uses: actions/setup-java@v2
with:
distribution: 'zulu'
java-version: '17'
- name: Cache Scala, SBT
uses: actions/cache@v2
with:
path: |
~/.sbt
~/.ivy2
~/.cache/coursier
~/.m2
key: build-cache-3-with-scala_${{ matrix.scala }}
- name: Run Scala Style tests on test sources
run: build/sbt "++ ${{ matrix.scala }}" testScalastyle
- name: Run sqlDeltaImport tests
run: build/sbt "++ ${{ matrix.scala }}" sqlDeltaImport/test
# These tests are not working yet
# - name: Run Delta Standalone Compatibility tests (Scala 2.12 only)
# run: build/sbt "++ ${{ matrix.scala }}" compatibility/test
# if: startsWith(matrix.scala, '2.12.')
- name: Run Delta Standalone tests
run: build/sbt "++ ${{ matrix.scala }}" standalone/test testStandaloneCosmetic/test standaloneParquet/test testParquetUtilsWithStandaloneCosmetic/test
- name: Run Hive 3 tests
run: build/sbt "++ ${{ matrix.scala }}" hiveMR/test hiveTez/test
- name: Run Hive 2 tests
run: build/sbt "++ ${{ matrix.scala }}" hive2MR/test hive2Tez/test