Files
sahara-ci-config/slave-scripts/gate-sahara.sh
Denis Egorenko 2a97387496 Remove Vanilla 2.3 jobs
Change-Id: I8dea288435966c245859ebb1745462ba9bb5cde5
2014-12-23 15:15:05 +04:00

127 lines
3.5 KiB
Bash

#!/bin/bash
. $FUNCTION_PATH
check_openstack_host
sudo pip install .
WORKSPACE=${1:-$WORKSPACE}
ENGINE_TYPE=$(echo $JOB_NAME | awk -F '-' '{ print $4 }')
JOB_TYPE=$(echo $JOB_NAME | awk -F '-' '{ print $5 }')
hadoop_version=1
SKIP_CINDER_TEST=False
SKIP_CLUSTER_CONFIG_TEST=False
SKIP_EDP_TEST=False
SKIP_MAP_REDUCE_TEST=True
SKIP_SWIFT_TEST=True
SKIP_SCALING_TEST=False
SKIP_TRANSIENT_TEST=True
SKIP_ONLY_TRANSIENT_TEST=False
SKIP_ALL_TESTS_FOR_PLUGIN=False
HDP_IMAGE=sahara-itests-ci-hdp-image-jdk-iptables-off
HDP_TWO_IMAGE=centos-6_4-64-hdp-2-0-hw
VANILLA_IMAGE=sahara-itests-ci-vanilla-image
VANILLA_TWO_IMAGE=ubuntu-vanilla-2.4-latest
SPARK_IMAGE=sahara_spark_latest
HEAT_JOB=False
if [[ "$ENGINE_TYPE" == 'heat' ]]
then
HEAT_JOB=True
echo "Heat detected"
fi
case $JOB_TYPE in
hdp_1)
PLUGIN_TYPE=hdp1
echo "HDP1 detected"
;;
hdp_2)
PLUGIN_TYPE=hdp2
hadoop_version=2
echo "HDP2 detected"
;;
vanilla*)
hadoop_version=$(echo $JOB_TYPE | awk -F '_' '{ print $2}')
if [ "$hadoop_version" == "1" ]; then
PLUGIN_TYPE=vanilla1
echo "Vanilla detected"
else
PLUGIN_TYPE=vanilla2
if [ "$hadoop_version" == "2.4" ]; then
hadoop_version=2-4
[ "$ZUUL_BRANCH" == "stable/icehouse" ] && echo "Vanilla 2.4 plugin is not supported in stable/icehouse" && exit 0
else
hadoop_version=2-6
VANILLA_TWO_IMAGE=ubuntu-vanilla-2.6-latest
[ "$ZUUL_BRANCH" == "stable/icehouse" -o "$ZUUL_BRANCH" == "stable/juno" ] && echo "Vanilla 2.6 plugin is not supported in stable/icehouse and stable/juno" && exit 0
fi
echo "Vanilla2 detected"
fi
;;
transient)
PLUGIN_TYPE=transient
SKIP_EDP_TEST=False
SKIP_TRANSIENT_TEST=False
SKIP_ONLY_TRANSIENT_TEST=True
SKIP_TRANSIENT_JOB=True
TRANSIENT_JOB=True
[ "$HEAT_JOB" == "True" ] && [ "$ZUUL_BRANCH" == "stable/icehouse" ] && echo "Heat_Transient plugin is not supported in stable/icehouse" && exit 0
echo "Transient detected"
;;
cdh*)
os_version=$(echo $JOB_TYPE | awk -F '_' '{ print $2}')
if [ "$os_version" == "centos" ]; then
CDH_IMAGE=centos_cdh_latest
hadoop_version=2c
else
# temporary using native ubuntu image
#CDH_IMAGE=ubuntu_cdh_latest
CDH_IMAGE=ubuntu-12.04
hadoop_version=2u
fi
SKIP_SCALING_TEST=True
PLUGIN_TYPE=cdh
[ "$ZUUL_BRANCH" == "stable/icehouse" ] && echo "CDH plugin is not supported in stable/icehouse" && exit 0
echo "CDH detected"
;;
spark)
PLUGIN_TYPE=spark
SKIP_EDP_TEST=False
SKIP_SCALING_TEST=False
[ "$ZUUL_BRANCH" == "stable/icehouse" ] && echo "Spark plugin is not supported in stable/icehouse" && exit 0
echo "Spark detected"
;;
esac
cd $WORKSPACE
[ "$ZUUL_BRANCH" == "stable/icehouse" ] && git checkout stable/icehouse && sudo pip install -U -r requirements.txt
TOX_LOG=$WORKSPACE/.tox/venv/log/venv-1.log
create_database
enable_pypi
write_sahara_main_conf etc/sahara/sahara.conf
start_sahara etc/sahara/sahara.conf
cd $WORKSPACE
CLUSTER_NAME="$HOST-$hadoop_version-$BUILD_NUMBER-$ZUUL_CHANGE-$ZUUL_PATCHSET"
write_tests_conf $WORKSPACE/sahara/tests/integration/configs/itest.conf
run_tests
cat_logs $WORKSPACE
if [ "$FAILURE" != 0 ]; then
exit 1
fi
if [[ "$STATUS" != 0 ]]
then
exit 1
fi