Skip to content

Commit

Permalink
fix: enable integration accuracy tests (#2976)
Browse files Browse the repository at this point in the history
  • Loading branch information
i8run committed Dec 17, 2019
1 parent 067a37a commit 1acffef
Showing 1 changed file with 20 additions and 20 deletions.
40 changes: 20 additions & 20 deletions spark/dl/src/test/integration-accuracy-test.robot
Original file line number Diff line number Diff line change
Expand Up @@ -5,17 +5,15 @@ Suite Setup Prepare DataSource And Verticals
Suite Teardown Delete All Sessions
Test template BigDL Test

*** Variables ***
@{verticals} ${spark_200_3_vid} ${spark_210_3_vid} ${hdfs_264_3_vid} ${spark_tf_210_3_vid} ${spark_tf_163_3_vid}

*** Test Cases *** SuiteName VerticalId
1 Spark2.0 Test Suite ${spark_200_3_vid}
2 Spark2.1 Test Suite ${spark_210_3_vid}
3 Hdfs Test Suite ${hdfs_264_3_vid}
4 Quantization Test Suite ${hdfs_264_3_vid}
5 PySpark2.1 Test Suite ${spark_tf_210_3_vid}
6 PySpark1.6 Test Suite ${spark_tf_163_3_vid}
7 Yarn Test Suite ${hdfs_264_3_vid}
*** Test Cases *** SuiteName
1 Spark2.0 Test Suite
2 Spark2.1 Test Suite
3 Hdfs Test Suite
4 Quantization Test Suite
5 PySpark2.1 Test Suite
6 PySpark1.6 Test Suite
7 Yarn Test Suite
8 Torch Test Suite

# predefined service masters:
# hdfs_264_3_master
Expand All @@ -42,7 +40,7 @@ Build SparkJar
Log To Console build jar finished

DownLoad Input
${hadoop}= Catenate SEPARATOR=/ /opt/work/hadoop-2.6.5/bin hadoop
${hadoop}= Catenate SEPARATOR=/ /opt/work/hadoop-2.7.2/bin hadoop
Run ${hadoop} fs -get ${mnist_data_source} ./
Log To Console got mnist data!!
Run ${hadoop} fs -get ${cifar_data_source} ./
Expand Down Expand Up @@ -84,18 +82,18 @@ Run Spark Test
Log To Console begin PTBWordLM
Run Shell ${submit} --master ${spark_master} --conf "spark.serializer=org.apache.spark.serializer.JavaSerializer" --driver-memory 40g --executor-memory 100g --executor-cores 8 --total-executor-cores 8 --class com.intel.analytics.bigdl.example.languagemodel.PTBWordLM ${jar_path} -f ./simple-examples/data -b 120 --numLayers 2 --vocab 10001 --hidden 650 --numSteps 35 --learningRate 0.005 -e 1 --learningRateDecay 0.001 --keepProb 0.5 --overWrite > 3.txt
Log To Console begin resnet Train
Run Shell ${submit} --master ${spark_master} --conf "spark.serializer=org.apache.spark.serializer.JavaSerializer" --driver-memory 5g --executor-memory 5g --executor-cores 8 --total-executor-cores 32 --class com.intel.analytics.bigdl.models.resnet.Train ${jar_path} -f ./cifar --batchSize 448 --optnet true --depth 20 --classes 10 --shortcutType A --nEpochs 1 --learningRate 0.1 > 4.txt
Run Shell ${submit} --master ${spark_master} --conf "spark.serializer=org.apache.spark.serializer.JavaSerializer" --driver-memory 5g --executor-memory 5g --executor-cores 8 --total-executor-cores 32 --class com.intel.analytics.bigdl.models.resnet.TrainCIFAR10 ${jar_path} -f ./cifar --batchSize 448 --optnet true --depth 20 --classes 10 --shortcutType A --nEpochs 1 --learningRate 0.1 > 4.txt
Log To Console begin DLClassifierLeNet
Run Shell ${submit} --master ${spark_master} --executor-cores 24 --total-executor-cores 24 --driver-memory 60g --executor-memory 200g --class com.intel.analytics.bigdl.example.MLPipeline.DLClassifierLeNet ${jar_path} -b 1200 -f ./mnist --maxEpoch 1 > 5.txt
Run Shell ${submit} --master ${spark_master} --executor-cores 16 --total-executor-cores 16 --driver-memory 5g --executor-memory 30g --class com.intel.analytics.bigdl.example.MLPipeline.DLClassifierLeNet ${jar_path} -b 1200 -f ./mnist --maxEpoch 1 > 5.txt
Log To Console begin rnn Train
Run Shell ${submit} --master ${spark_master} --driver-memory 5g --executor-memory 5g --executor-cores 12 --total-executor-cores 12 --class com.intel.analytics.bigdl.models.rnn.Train ${jar_path} -f ./ -s ./models --nEpochs 1 --checkpoint ./model/ -b 12 > 6.txt
Run Shell bash spark/dl/src/test/accuracy-judge.sh
Log To Console begin inceptionV1 train
Run Shell ${submit} --master ${spark_master} --conf "spark.serializer=org.apache.spark.serializer.JavaSerializer" --driver-memory 60g --executor-memory 200g --executor-cores 24 --total-executor-cores 24 --class com.intel.analytics.bigdl.models.inception.TrainInceptionV1 ${jar_path} -b 24 -f ${imagenet_test_data_source} --learningRate 0.1 -e 1
Run Shell ${submit} --master ${spark_master} --conf "spark.serializer=org.apache.spark.serializer.JavaSerializer" --driver-memory 20g --executor-memory 40g --executor-cores 4 --total-executor-cores 8 --class com.intel.analytics.bigdl.models.inception.TrainInceptionV1 ${jar_path} -b 24 -f ${imagenet_test_data_source} --learningRate 0.1 -i 200
Log To Console begin googlenet
Run Shell ${submit} --master ${spark_master} --conf "spark.serializer=org.apache.spark.serializer.JavaSerializer" --driver-class-path ${jar_path} --driver-memory 20g --executor-memory 100g --executor-cores 28 --total-executor-cores 112 --class com.intel.analytics.bigdl.example.loadmodel.ModelValidator ${jar_path} -b 448 -f ${imagenet_data_source}/val/ -t caffe -m inception --caffeDefPath ${public_hdfs_master}:9000/models/bvlc_googlenet/deploy.prototxt --modelPath ${public_hdfs_master}:9000/models/bvlc_googlenet.caffemodel
Run Shell ${submit} --master ${spark_master} --conf "spark.serializer=org.apache.spark.serializer.JavaSerializer" --driver-class-path ${jar_path} --driver-memory 20g --executor-memory 100g --executor-cores 10 --total-executor-cores 20 --class com.intel.analytics.bigdl.example.loadmodel.ModelValidator ${jar_path} -b 420 -f ${imagenet_data_source}/val/ -t caffe -m inception --caffeDefPath ${public_hdfs_master}:9000/models/bvlc_googlenet/deploy.prototxt --modelPath ${public_hdfs_master}:9000/models/bvlc_googlenet.caffemodel
Log To Console begin alexnet
Run Shell ${submit} --master ${spark_master} --conf "spark.serializer=org.apache.spark.serializer.JavaSerializer" --driver-class-path ${jar_path} --driver-memory 20g --executor-memory 100g --executor-cores 28 --total-executor-cores 112 --class com.intel.analytics.bigdl.example.loadmodel.ModelValidator ${jar_path} -b 448 -f ${imagenet_data_source}/val/ -t caffe -m inception --caffeDefPath ${public_hdfs_master}:9000/models/bvlc_alexnet/deploy.prototxt --modelPath ${public_hdfs_master}:9000/models/bvlc_alexnet.caffemodel
Run Shell ${submit} --master ${spark_master} --conf "spark.serializer=org.apache.spark.serializer.JavaSerializer" --driver-class-path ${jar_path} --driver-memory 20g --executor-memory 100g --executor-cores 10 --total-executor-cores 20 --class com.intel.analytics.bigdl.example.loadmodel.ModelValidator ${jar_path} -b 420 -f ${imagenet_data_source}/val/ -t caffe -m inception --caffeDefPath ${public_hdfs_master}:9000/models/bvlc_alexnet/deploy.prototxt --modelPath ${public_hdfs_master}:9000/models/bvlc_alexnet.caffemodel
Log To Console begin treeLSTM
Run Shell ${submit} --master ${spark_master} --driver-memory 20g --executor-memory 10g --total-executor-cores 8 --executor-cores 8 --class com.intel.analytics.bigdl.example.treeLSTMSentiment.Train ${jar_path} --baseDir ${public_hdfs_master}:9000/dataset/ --epoch 1
Log To Console begin text classification
Expand Down Expand Up @@ -125,7 +123,7 @@ Hdfs Test Suite


Quantization Test Suite
${hadoop}= Catenate SEPARATOR=/ /opt/work/hadoop-2.6.5/bin hadoop
${hadoop}= Catenate SEPARATOR=/ /opt/work/hadoop-2.7.2/bin hadoop
Run ${hadoop} fs -get ${mnist_data_source} /tmp/
Log To Console got mnist data!!
Run ${hadoop} fs -get ${cifar_data_source} /tmp/
Expand All @@ -145,8 +143,6 @@ Yarn Test Suite
Set Environment Variable https_proxy ${https_proxy}
${submit}= Catenate SEPARATOR=/ /opt/work/spark-2.0.0-bin-hadoop2.7/bin spark-submit
Run Shell ${submit} --master yarn --deploy-mode client --conf "spark.serializer=org.apache.spark.serializer.JavaSerializer" --executor-cores 10 --num-executors 3 --driver-memory 150g --class com.intel.analytics.bigdl.models.lenet.Train ${jar_path} -f ${mnist_data_source} -b 120 -e 3
Set Environment Variable PYSPARK_DRIVER_PYTHON /var/jenkins_home/venv/bin/python
Set Environment Variable PYSPARK_PYTHON ./venv.zip/venv/bin/python
Run Shell ${submit} --master yarn --deploy-mode client --executor-memory 2g --driver-memory 2g --executor-cores 10 --num-executors 2 --properties-file ${curdir}/dist/conf/spark-bigdl.conf --jars ${jar_path} --py-files ${curdir}/dist/lib/bigdl-${version}-python-api.zip --archives /var/jenkins_home/venv.zip --conf spark.driver.extraClassPath=${jar_path} --conf spark.executor.extraClassPath=bigdl-${version}-jar-with-dependencies.jar ${curdir}/pyspark/bigdl/models/lenet/lenet5.py -b 200 --action train --endTriggerType epoch --endTriggerNum 1
Remove Environment Variable http_proxy https_proxy PYSPARK_DRIVER_PYTHON PYSPARK_PYTHON

Expand All @@ -162,3 +158,7 @@ PySpark1.6 Test Suite
Set Environment Variable SPARK_HOME /opt/work/spark-1.6.3-bin-hadoop2.6
${submit}= Catenate SEPARATOR=/ /opt/work/spark-1.6.3-bin-hadoop2.6/bin spark-submit
Run Shell ${submit} --master ${spark_tf_163_3_master} --conf "spark.serializer=org.apache.spark.serializer.JavaSerializer" --driver-memory 150g --executor-cores 28 --total-executor-cores 56 --py-files ${curdir}/dist/lib/bigdl-${version}-python-api.zip --jars ${jar_path} --properties-file ${curdir}/dist/conf/spark-bigdl.conf --conf spark.driver.extraClassPath=${jar_path} --conf spark.executor.extraClassPath=bigdl-${version}-jar-with-dependencies.jar ${curdir}/pyspark/bigdl/models/lenet/lenet5.py -b 224 --action train --endTriggerType epoch --endTriggerNum 1

Torch Test Suite
Build SparkJar spark_1.6
Run Shell mvn clean test -Dsuites=com.intel.analytics.bigdl.integration.torch

0 comments on commit 1acffef

Please sign in to comment.