From 1acffef81251419f3b2674a16d7fa1bcbf570dc2 Mon Sep 17 00:00:00 2001 From: Yanzhang Wang Date: Tue, 17 Dec 2019 20:35:43 +0800 Subject: [PATCH] fix: enable integration accuracy tests (#2976) --- .../src/test/integration-accuracy-test.robot | 40 +++++++++---------- 1 file changed, 20 insertions(+), 20 deletions(-) diff --git a/spark/dl/src/test/integration-accuracy-test.robot b/spark/dl/src/test/integration-accuracy-test.robot index 35c9a77f841..c92e34e0848 100644 --- a/spark/dl/src/test/integration-accuracy-test.robot +++ b/spark/dl/src/test/integration-accuracy-test.robot @@ -5,17 +5,15 @@ Suite Setup Prepare DataSource And Verticals Suite Teardown Delete All Sessions Test template BigDL Test -*** Variables *** -@{verticals} ${spark_200_3_vid} ${spark_210_3_vid} ${hdfs_264_3_vid} ${spark_tf_210_3_vid} ${spark_tf_163_3_vid} - -*** Test Cases *** SuiteName VerticalId -1 Spark2.0 Test Suite ${spark_200_3_vid} -2 Spark2.1 Test Suite ${spark_210_3_vid} -3 Hdfs Test Suite ${hdfs_264_3_vid} -4 Quantization Test Suite ${hdfs_264_3_vid} -5 PySpark2.1 Test Suite ${spark_tf_210_3_vid} -6 PySpark1.6 Test Suite ${spark_tf_163_3_vid} -7 Yarn Test Suite ${hdfs_264_3_vid} +*** Test Cases *** SuiteName +1 Spark2.0 Test Suite +2 Spark2.1 Test Suite +3 Hdfs Test Suite +4 Quantization Test Suite +5 PySpark2.1 Test Suite +6 PySpark1.6 Test Suite +7 Yarn Test Suite +8 Torch Test Suite # predefined service masters: # hdfs_264_3_master @@ -42,7 +40,7 @@ Build SparkJar Log To Console build jar finished DownLoad Input - ${hadoop}= Catenate SEPARATOR=/ /opt/work/hadoop-2.6.5/bin hadoop + ${hadoop}= Catenate SEPARATOR=/ /opt/work/hadoop-2.7.2/bin hadoop Run ${hadoop} fs -get ${mnist_data_source} ./ Log To Console got mnist data!! Run ${hadoop} fs -get ${cifar_data_source} ./ @@ -84,18 +82,18 @@ Run Spark Test Log To Console begin PTBWordLM Run Shell ${submit} --master ${spark_master} --conf "spark.serializer=org.apache.spark.serializer.JavaSerializer" --driver-memory 40g --executor-memory 100g --executor-cores 8 --total-executor-cores 8 --class com.intel.analytics.bigdl.example.languagemodel.PTBWordLM ${jar_path} -f ./simple-examples/data -b 120 --numLayers 2 --vocab 10001 --hidden 650 --numSteps 35 --learningRate 0.005 -e 1 --learningRateDecay 0.001 --keepProb 0.5 --overWrite > 3.txt Log To Console begin resnet Train - Run Shell ${submit} --master ${spark_master} --conf "spark.serializer=org.apache.spark.serializer.JavaSerializer" --driver-memory 5g --executor-memory 5g --executor-cores 8 --total-executor-cores 32 --class com.intel.analytics.bigdl.models.resnet.Train ${jar_path} -f ./cifar --batchSize 448 --optnet true --depth 20 --classes 10 --shortcutType A --nEpochs 1 --learningRate 0.1 > 4.txt + Run Shell ${submit} --master ${spark_master} --conf "spark.serializer=org.apache.spark.serializer.JavaSerializer" --driver-memory 5g --executor-memory 5g --executor-cores 8 --total-executor-cores 32 --class com.intel.analytics.bigdl.models.resnet.TrainCIFAR10 ${jar_path} -f ./cifar --batchSize 448 --optnet true --depth 20 --classes 10 --shortcutType A --nEpochs 1 --learningRate 0.1 > 4.txt Log To Console begin DLClassifierLeNet - Run Shell ${submit} --master ${spark_master} --executor-cores 24 --total-executor-cores 24 --driver-memory 60g --executor-memory 200g --class com.intel.analytics.bigdl.example.MLPipeline.DLClassifierLeNet ${jar_path} -b 1200 -f ./mnist --maxEpoch 1 > 5.txt + Run Shell ${submit} --master ${spark_master} --executor-cores 16 --total-executor-cores 16 --driver-memory 5g --executor-memory 30g --class com.intel.analytics.bigdl.example.MLPipeline.DLClassifierLeNet ${jar_path} -b 1200 -f ./mnist --maxEpoch 1 > 5.txt Log To Console begin rnn Train Run Shell ${submit} --master ${spark_master} --driver-memory 5g --executor-memory 5g --executor-cores 12 --total-executor-cores 12 --class com.intel.analytics.bigdl.models.rnn.Train ${jar_path} -f ./ -s ./models --nEpochs 1 --checkpoint ./model/ -b 12 > 6.txt Run Shell bash spark/dl/src/test/accuracy-judge.sh Log To Console begin inceptionV1 train - Run Shell ${submit} --master ${spark_master} --conf "spark.serializer=org.apache.spark.serializer.JavaSerializer" --driver-memory 60g --executor-memory 200g --executor-cores 24 --total-executor-cores 24 --class com.intel.analytics.bigdl.models.inception.TrainInceptionV1 ${jar_path} -b 24 -f ${imagenet_test_data_source} --learningRate 0.1 -e 1 + Run Shell ${submit} --master ${spark_master} --conf "spark.serializer=org.apache.spark.serializer.JavaSerializer" --driver-memory 20g --executor-memory 40g --executor-cores 4 --total-executor-cores 8 --class com.intel.analytics.bigdl.models.inception.TrainInceptionV1 ${jar_path} -b 24 -f ${imagenet_test_data_source} --learningRate 0.1 -i 200 Log To Console begin googlenet - Run Shell ${submit} --master ${spark_master} --conf "spark.serializer=org.apache.spark.serializer.JavaSerializer" --driver-class-path ${jar_path} --driver-memory 20g --executor-memory 100g --executor-cores 28 --total-executor-cores 112 --class com.intel.analytics.bigdl.example.loadmodel.ModelValidator ${jar_path} -b 448 -f ${imagenet_data_source}/val/ -t caffe -m inception --caffeDefPath ${public_hdfs_master}:9000/models/bvlc_googlenet/deploy.prototxt --modelPath ${public_hdfs_master}:9000/models/bvlc_googlenet.caffemodel + Run Shell ${submit} --master ${spark_master} --conf "spark.serializer=org.apache.spark.serializer.JavaSerializer" --driver-class-path ${jar_path} --driver-memory 20g --executor-memory 100g --executor-cores 10 --total-executor-cores 20 --class com.intel.analytics.bigdl.example.loadmodel.ModelValidator ${jar_path} -b 420 -f ${imagenet_data_source}/val/ -t caffe -m inception --caffeDefPath ${public_hdfs_master}:9000/models/bvlc_googlenet/deploy.prototxt --modelPath ${public_hdfs_master}:9000/models/bvlc_googlenet.caffemodel Log To Console begin alexnet - Run Shell ${submit} --master ${spark_master} --conf "spark.serializer=org.apache.spark.serializer.JavaSerializer" --driver-class-path ${jar_path} --driver-memory 20g --executor-memory 100g --executor-cores 28 --total-executor-cores 112 --class com.intel.analytics.bigdl.example.loadmodel.ModelValidator ${jar_path} -b 448 -f ${imagenet_data_source}/val/ -t caffe -m inception --caffeDefPath ${public_hdfs_master}:9000/models/bvlc_alexnet/deploy.prototxt --modelPath ${public_hdfs_master}:9000/models/bvlc_alexnet.caffemodel + Run Shell ${submit} --master ${spark_master} --conf "spark.serializer=org.apache.spark.serializer.JavaSerializer" --driver-class-path ${jar_path} --driver-memory 20g --executor-memory 100g --executor-cores 10 --total-executor-cores 20 --class com.intel.analytics.bigdl.example.loadmodel.ModelValidator ${jar_path} -b 420 -f ${imagenet_data_source}/val/ -t caffe -m inception --caffeDefPath ${public_hdfs_master}:9000/models/bvlc_alexnet/deploy.prototxt --modelPath ${public_hdfs_master}:9000/models/bvlc_alexnet.caffemodel Log To Console begin treeLSTM Run Shell ${submit} --master ${spark_master} --driver-memory 20g --executor-memory 10g --total-executor-cores 8 --executor-cores 8 --class com.intel.analytics.bigdl.example.treeLSTMSentiment.Train ${jar_path} --baseDir ${public_hdfs_master}:9000/dataset/ --epoch 1 Log To Console begin text classification @@ -125,7 +123,7 @@ Hdfs Test Suite Quantization Test Suite - ${hadoop}= Catenate SEPARATOR=/ /opt/work/hadoop-2.6.5/bin hadoop + ${hadoop}= Catenate SEPARATOR=/ /opt/work/hadoop-2.7.2/bin hadoop Run ${hadoop} fs -get ${mnist_data_source} /tmp/ Log To Console got mnist data!! Run ${hadoop} fs -get ${cifar_data_source} /tmp/ @@ -145,8 +143,6 @@ Yarn Test Suite Set Environment Variable https_proxy ${https_proxy} ${submit}= Catenate SEPARATOR=/ /opt/work/spark-2.0.0-bin-hadoop2.7/bin spark-submit Run Shell ${submit} --master yarn --deploy-mode client --conf "spark.serializer=org.apache.spark.serializer.JavaSerializer" --executor-cores 10 --num-executors 3 --driver-memory 150g --class com.intel.analytics.bigdl.models.lenet.Train ${jar_path} -f ${mnist_data_source} -b 120 -e 3 - Set Environment Variable PYSPARK_DRIVER_PYTHON /var/jenkins_home/venv/bin/python - Set Environment Variable PYSPARK_PYTHON ./venv.zip/venv/bin/python Run Shell ${submit} --master yarn --deploy-mode client --executor-memory 2g --driver-memory 2g --executor-cores 10 --num-executors 2 --properties-file ${curdir}/dist/conf/spark-bigdl.conf --jars ${jar_path} --py-files ${curdir}/dist/lib/bigdl-${version}-python-api.zip --archives /var/jenkins_home/venv.zip --conf spark.driver.extraClassPath=${jar_path} --conf spark.executor.extraClassPath=bigdl-${version}-jar-with-dependencies.jar ${curdir}/pyspark/bigdl/models/lenet/lenet5.py -b 200 --action train --endTriggerType epoch --endTriggerNum 1 Remove Environment Variable http_proxy https_proxy PYSPARK_DRIVER_PYTHON PYSPARK_PYTHON @@ -162,3 +158,7 @@ PySpark1.6 Test Suite Set Environment Variable SPARK_HOME /opt/work/spark-1.6.3-bin-hadoop2.6 ${submit}= Catenate SEPARATOR=/ /opt/work/spark-1.6.3-bin-hadoop2.6/bin spark-submit Run Shell ${submit} --master ${spark_tf_163_3_master} --conf "spark.serializer=org.apache.spark.serializer.JavaSerializer" --driver-memory 150g --executor-cores 28 --total-executor-cores 56 --py-files ${curdir}/dist/lib/bigdl-${version}-python-api.zip --jars ${jar_path} --properties-file ${curdir}/dist/conf/spark-bigdl.conf --conf spark.driver.extraClassPath=${jar_path} --conf spark.executor.extraClassPath=bigdl-${version}-jar-with-dependencies.jar ${curdir}/pyspark/bigdl/models/lenet/lenet5.py -b 224 --action train --endTriggerType epoch --endTriggerNum 1 + +Torch Test Suite + Build SparkJar spark_1.6 + Run Shell mvn clean test -Dsuites=com.intel.analytics.bigdl.integration.torch