Skip to content

Sparknlp 1287 vector db connector annotator #4340

Sparknlp 1287 vector db connector annotator

Sparknlp 1287 vector db connector annotator #4340

name: build
on:
push:
paths-ignore:
- "docs/**"
- "conda/**"
- "scripts/**"
- "examples/**"
- "**.md"
- "**.yaml"
- "**.yml"
- "**.ipynb"
branches:
- "master"
- "*release*"
- "release/**"
- "main"
pull_request:
paths-ignore:
- "docs/**"
- "conda/**"
- "scripts/**"
- "examples/**"
- "**.md"
- "**.yaml"
- "**.yml"
- "**.ipynb"
branches:
- "master"
- "*release*"
- "release/**"
- "main"
concurrency:
group: ${{ github. workflow }}-${{ github.event. pull_request.number || github. ref }}
cancel-in-progress: true
jobs:
spark34:
if: "! contains(toJSON(github.event.commits.*.message), '[skip test]')"
runs-on: ubuntu-22.04
env:
TF_CPP_MIN_LOG_LEVEL: 3
JAVA_OPTS: "-Xmx4096m -XX:+UseG1GC"
name: Build and Test on Apache Spark 3.4.x
steps:
- uses: actions/checkout@v3
- name: Setup openjdk Java
run: |
sudo apt-get update
sudo apt-get install -y openjdk-8-jdk
echo "JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64" >> $GITHUB_ENV
echo "/usr/lib/jvm/java-8-openjdk-amd64/bin" >> $GITHUB_PATH
java -version
- name: Install Python 3.8
uses: actions/setup-python@v2
with:
python-version: 3.8
architecture: x64
- name: Install Python packages (Python 3.8)
run: |
python -m pip install --upgrade pip
pip install pyspark==3.4.0 numpy pytest
- uses: sbt/setup-sbt@v1
- name: Build Spark NLP on Apache Spark 3.4.0
run: |
sbt -mem 4096 -Dis_spark34=true clean assemblyAndCopy
- name: Test Spark NLP in Scala - Apache Spark 3.4.x
run: |
sbt -mem 4096 coverage test
- name: Upload coverage data to Coveralls
run: sbt coverageReport coveralls
env:
COVERALLS_REPO_TOKEN: ${{ secrets.GITHUB_TOKEN }}
COVERALLS_FLAG_NAME: Apache Spark 3.4.x - Scala 2.12
- name: Test Spark NLP in Python - Apache Spark 3.4.x
run: |
cd python
python3.8 -m pytest -v -m fast
spark35:
if: "! contains(toJSON(github.event.commits.*.message), '[skip test]')"
runs-on: ubuntu-22.04
env:
TF_CPP_MIN_LOG_LEVEL: 3
JAVA_OPTS: "-Xmx4096m -XX:+UseG1GC"
name: Build and Test on Apache Spark 3.5.x
steps:
- uses: actions/checkout@v3
- name: Setup openjdk Java
run: |
sudo apt-get update
sudo apt-get install -y openjdk-8-jdk
echo "JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64" >> $GITHUB_ENV
echo "/usr/lib/jvm/java-8-openjdk-amd64/bin" >> $GITHUB_PATH
java -version
- name: Install Python 3.10
uses: actions/setup-python@v2
with:
python-version: 3.10.12
architecture: x64
- name: Install Python packages (Python 3.10)
run: |
python -m pip install --upgrade pip
pip install pyspark==3.5.0 numpy pytest
- uses: sbt/setup-sbt@v1
- name: Build Spark NLP on Apache Spark 3.5.0
run: |
sbt -mem 4096 -Dis_spark35=true clean assemblyAndCopy
- name: Test Spark NLP in Scala - Apache Spark 3.5.x
run: |
sbt -mem 4096 test
- name: Test Spark NLP in Python - Apache Spark 3.5.x
run: |
cd python
python3.10 -m pytest -v -m fast
spark33:
if: "! contains(toJSON(github.event.commits.*.message), '[skip test]')"
runs-on: ubuntu-22.04
env:
TF_CPP_MIN_LOG_LEVEL: 3
JAVA_OPTS: "-Xmx4096m -XX:+UseG1GC"
name: Build and Test on Apache Spark 3.3.x
steps:
- uses: actions/checkout@v3
- name: Setup openjdk Java
run: |
sudo apt-get update
sudo apt-get install -y openjdk-8-jdk
echo "JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64" >> $GITHUB_ENV
echo "/usr/lib/jvm/java-8-openjdk-amd64/bin" >> $GITHUB_PATH
java -version
- name: Install Python 3.8
uses: actions/setup-python@v2
with:
python-version: 3.8
architecture: x64
- name: Install Python packages (Python 3.8)
run: |
python -m pip install --upgrade pip
pip install pyspark==3.3.1 numpy pytest
- uses: sbt/setup-sbt@v1
- name: Build Spark NLP on Apache Spark 3.3.1
run: |
sbt -mem 4096 -Dis_spark33=true clean assemblyAndCopy
- name: Test Spark NLP in Scala - Apache Spark 3.3.x
run: |
sbt -mem 4096 test
- name: Test Spark NLP in Python - Apache Spark 3.3.x
run: |
cd python
python3.8 -m pytest -v -m fast