#!/usr/bin/env bash # # Licensed to the Apache Software Foundation (ASF) under one or more # contributor license agreements. See the NOTICE file distributed with # this work for additional information regarding copyright ownership. # The ASF licenses this file to You under the Apache License, Version 2.0 # (the "License"); you may not use this file except in compliance with # the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # # Figure out where the Spark framework is installed FWDIR="$(cd "`dirname "$0"`"; cd ../; pwd)" . "$FWDIR"/bin/load-spark-env.sh # CD into the python directory to find things on the right path cd "$FWDIR/python" FAILED=0 LOG_FILE=unit-tests.log START=$(date +"%s") rm -f $LOG_FILE # Remove the metastore and warehouse directory created by the HiveContext tests in Spark SQL rm -rf metastore warehouse function run_test() { echo -en "Running test: $1 ... " | tee -a $LOG_FILE start=$(date +"%s") SPARK_TESTING=1 time "$FWDIR"/bin/pyspark $1 > $LOG_FILE 2>&1 FAILED=$((PIPESTATUS[0]||$FAILED)) # Fail and exit on the first test failure. if [[ $FAILED != 0 ]]; then cat $LOG_FILE | grep -v "^[0-9][0-9]*" # filter all lines starting with a number. echo -en "\033[31m" # Red echo "Had test failures; see logs." echo -en "\033[0m" # No color exit -1 else now=$(date +"%s") echo "ok ($(($now - $start))s)" fi } function run_core_tests() { echo "Run core tests ..." run_test "pyspark.rdd" run_test "pyspark.context" run_test "pyspark.conf" run_test "pyspark.broadcast" run_test "pyspark.accumulators" run_test "pyspark.serializers" run_test "pyspark.profiler" run_test "pyspark.shuffle" run_test "pyspark.tests" } function run_sql_tests() { echo "Run sql tests ..." run_test "pyspark.sql.types" run_test "pyspark.sql.context" run_test "pyspark.sql.column" run_test "pyspark.sql.dataframe" run_test "pyspark.sql.group" run_test "pyspark.sql.functions" run_test "pyspark.sql.readwriter" run_test "pyspark.sql.window" run_test "pyspark.sql.tests" } function run_mllib_tests() { echo "Run mllib tests ..." run_test "pyspark.mllib.classification" run_test "pyspark.mllib.clustering" run_test "pyspark.mllib.evaluation" run_test "pyspark.mllib.feature" run_test "pyspark.mllib.fpm" run_test "pyspark.mllib.linalg" run_test "pyspark.mllib.random" run_test "pyspark.mllib.recommendation" run_test "pyspark.mllib.regression" run_test "pyspark.mllib.stat._statistics" run_test "pyspark.mllib.tree" run_test "pyspark.mllib.util" run_test "pyspark.mllib.tests" } function run_ml_tests() { echo "Run ml tests ..." run_test "pyspark.ml.feature" run_test "pyspark.ml.classification" run_test "pyspark.ml.recommendation" run_test "pyspark.ml.regression" run_test "pyspark.ml.tuning" run_test "pyspark.ml.tests" run_test "pyspark.ml.evaluation" } function run_streaming_tests() { echo "Run streaming tests ..." KAFKA_ASSEMBLY_DIR="$FWDIR"/external/kafka-assembly JAR_PATH="${KAFKA_ASSEMBLY_DIR}/target/scala-${SPARK_SCALA_VERSION}" for f in "${JAR_PATH}"/spark-streaming-kafka-assembly-*.jar; do if [[ ! -e "$f" ]]; then echo "Failed to find Spark Streaming Kafka assembly jar in $KAFKA_ASSEMBLY_DIR" 1>&2 echo "You need to build Spark with " \ "'build/sbt assembly/assembly streaming-kafka-assembly/assembly' or" \ "'build/mvn package' before running this program" 1>&2 exit 1 fi KAFKA_ASSEMBLY_JAR="$f" done export PYSPARK_SUBMIT_ARGS="--jars ${KAFKA_ASSEMBLY_JAR} pyspark-shell" run_test "pyspark.streaming.util" run_test "pyspark.streaming.tests" } echo "Running PySpark tests. Output is in python/$LOG_FILE." export PYSPARK_PYTHON="python" # Try to test with Python 2.6, since that's the minimum version that we support: if [ $(which python2.6) ]; then export PYSPARK_PYTHON="python2.6" fi echo "Testing with Python version:" $PYSPARK_PYTHON --version run_core_tests run_sql_tests run_mllib_tests run_ml_tests run_streaming_tests # Try to test with Python 3 if [ $(which python3.4) ]; then export PYSPARK_PYTHON="python3.4" echo "Testing with Python3.4 version:" $PYSPARK_PYTHON --version run_core_tests run_sql_tests run_mllib_tests run_ml_tests run_streaming_tests fi # Try to test with PyPy if [ $(which pypy) ]; then export PYSPARK_PYTHON="pypy" echo "Testing with PyPy version:" $PYSPARK_PYTHON --version run_core_tests run_sql_tests run_streaming_tests fi if [[ $FAILED == 0 ]]; then now=$(date +"%s") echo -e "\033[32mTests passed \033[0min $(($now - $START)) seconds" fi # TODO: in the long-run, it would be nice to use a test runner like `nose`. # The doctest fixtures are the current barrier to doing this.