X-Git-Url: https://git.opendaylight.org/gerrit/gitweb?a=blobdiff_plain;f=jjb%2Fintegration%2Fcommon-functions.sh;h=c3fdb54a800c99ed8f358dd5fbf1c8ff07a3fa77;hb=6b87d4dde1886655d8ecacd5f1bc4809d8639106;hp=9bb051d494d99a534f69a67d309aa6d287c668de;hpb=940102f7fca2495d3601dfbf714730db93cdb792;p=releng%2Fbuilder.git diff --git a/jjb/integration/common-functions.sh b/jjb/integration/common-functions.sh index 9bb051d49..c3fdb54a8 100644 --- a/jjb/integration/common-functions.sh +++ b/jjb/integration/common-functions.sh @@ -21,6 +21,7 @@ function print_common_env() { cat << EOF common-functions environment: MAVENCONF: ${MAVENCONF} +ACTUALFEATURES: ${ACTUALFEATURES} FEATURESCONF: ${FEATURESCONF} CUSTOMPROP: ${CUSTOMPROP} LOGCONF: ${LOGCONF} @@ -29,6 +30,7 @@ CONTROLLERMEM: ${CONTROLLERMEM} AKKACONF: ${AKKACONF} MODULESCONF: ${MODULESCONF} MODULESHARDSCONF: ${MODULESHARDSCONF} +SUITES: ${SUITES} EOF } @@ -77,6 +79,9 @@ function configure_karaf_log() { echo "Configuring the karaf log... karaf_version: ${karaf_version}, logapi: ${logapi}" if [ "${logapi}" == "log4j2" ]; then # FIXME: Make log size limit configurable from build parameter. + # From Neon the default karaf file size is 64 MB + sed -ie 's/log4j2.appender.rolling.policies.size.size = 64MB/log4j2.appender.rolling.policies.size.size = 1GB/g' ${LOGCONF} + # Flourine still uses 16 MB sed -ie 's/log4j2.appender.rolling.policies.size.size = 16MB/log4j2.appender.rolling.policies.size.size = 1GB/g' ${LOGCONF} orgmodule="org.opendaylight.yangtools.yang.parser.repo.YangTextSchemaContextResolver" orgmodule_="${orgmodule//./_}" @@ -90,8 +95,9 @@ function configure_karaf_log() { fi # Add custom logging levels - # CONTROLLERDEBUGMAP is expected to be a key:value map of space separated values like "module:level module2:level2" - # where module is abbreviated and does not include "org.opendaylight." + # CONTROLLERDEBUGMAP is expected to be a key:value map of space separated + # values like "module:level module2:level2" where module is abbreviated and + # does not include "org.opendaylight." unset IFS echo "controllerdebugmap: ${controllerdebugmap}" if [ -n "${controllerdebugmap}" ]; then @@ -117,6 +123,123 @@ function configure_karaf_log() { cat ${LOGCONF} } # function configure_karaf_log() +function configure_karaf_log_for_apex() { + # TODO: add the extra steps to this function to do any extra work + # in this apex environment like we do in our standard environment. + # EX: log size, rollover, etc. + + # Modify ODL Log Levels, if needed, for new distribution. This will modify + # the control nodes hiera data which will be used during the puppet deploy + # CONTROLLERDEBUGMAP is expected to be a key:value map of space separated + # values like "module:level module2:level2" where module is abbreviated and + # does not include "org.opendaylight." + + local -r controller_ip=$1 + + unset IFS + # shellcheck disable=SC2153 + echo "CONTROLLERDEBUGMAP: ${CONTROLLERDEBUGMAP}" + if [ -n "${CONTROLLERDEBUGMAP}" ]; then + logging_config='\"opendaylight::log_levels\": {' + for kv in ${CONTROLLERDEBUGMAP}; do + module="${kv%%:*}" + level="${kv#*:}" + echo "module: $module, level: $level" + # shellcheck disable=SC2157 + if [ -n "${module}" ] && [ -n "${level}" ]; then + orgmodule="org.opendaylight.${module}" + logging_config="${logging_config} \\\"${orgmodule}\\\": \\\"${level}\\\"," + fi + done + # replace the trailing comma with a closing brace followed by trailing comma + logging_config=${logging_config%,}" }," + echo $logging_config + + # fine a sane line number to inject the custom logging json + lineno=$(ssh $OPENSTACK_CONTROL_NODE_1_IP "sudo grep -Fn 'opendaylight::log_mechanism' /etc/puppet/hieradata/service_configs.json" | awk -F: '{print $1}') + ssh $controller_ip "sudo sed -i \"${lineno}i ${logging_config}\" /etc/puppet/hieradata/service_configs.json" + ssh $controller_ip "sudo cat /etc/puppet/hieradata/service_configs.json" + fi +} # function configure_karaf_log_for_apex() + +function configure_odl_features_for_apex() { + + # if the environment variable $ACTUALFEATURES is not null, then rewrite + # the puppet config file with the features given in that variable, otherwise + # this function is a noop + + local -r controller_ip=$1 + local -r config_file=/etc/puppet/hieradata/service_configs.json + +cat > /tmp/set_odl_features.sh << EOF +sudo jq '.["opendaylight::extra_features"] |= []' $config_file > tmp.json && mv tmp.json $config_file +for feature in $(echo $ACTUALFEATURES | sed "s/,/ /g"); do + sudo jq --arg jq_arg \$feature '.["opendaylight::extra_features"] |= . + [\$jq_arg]' $config_file > tmp && mv tmp $config_file; +done +echo "Modified puppet-opendaylight service_configs.json..." +cat $config_file +EOF + + echo "Feature configuration script..." + cat /tmp/set_odl_features.sh + + if [ -n "${ACTUALFEATURES}" ]; then + scp /tmp/set_odl_features.sh $controller_ip:/tmp/set_odl_features.sh + ssh $controller_ip "sudo bash /tmp/set_odl_features.sh" + fi + +} # function configure_odl_features_for_apex() + +function get_os_deploy() { + local -r num_systems=${1:-$NUM_OPENSTACK_SYSTEM} + case ${num_systems} in + 1) + OPENSTACK_TOPO="1cmb-0ctl-0cmp" + ;; + 2) + OPENSTACK_TOPO="1cmb-0ctl-1cmp" + ;; + 3|*) + OPENSTACK_TOPO="0cmb-1ctl-2cmp" + ;; + esac + export OPENSTACK_TOPO +} + +function get_test_suites() { + + #let the caller pick the name of the variable we will assign the suites to + local __suite_list=$1 + + echo "Locating test plan to use..." + testplan_filepath="${WORKSPACE}/test/csit/testplans/${STREAMTESTPLAN}" + if [ ! -f "${testplan_filepath}" ]; then + testplan_filepath="${WORKSPACE}/test/csit/testplans/${TESTPLAN}" + fi + + echo "Changing the testplan path..." + cat "${testplan_filepath}" | sed "s:integration:${WORKSPACE}:" > testplan.txt + cat testplan.txt + + # Use the testplan if specific SUITES are not defined. + if [ -z "${SUITES}" ]; then + suite_list=$(egrep -v '(^[[:space:]]*#|^[[:space:]]*$)' testplan.txt | tr '\012' ' ') + else + suite_list="" + workpath="${WORKSPACE}/test/csit/suites" + for suite in ${SUITES}; do + fullsuite="${workpath}/${suite}" + if [ -z "${suite_list}" ]; then + suite_list+=${fullsuite} + else + suite_list+=" "${fullsuite} + fi + done + fi + + eval $__suite_list="'$suite_list'" +} + function run_plan() { local -r type=$1 @@ -129,7 +252,7 @@ function run_plan() { ;; esac - printf "Locating ${type} plan to use...\n" + printf "Locating %s plan to use...\n" "${type}" plan_filepath="${WORKSPACE}/test/csit/${type}plans/$plan" if [ ! -f "${plan_filepath}" ]; then plan_filepath="${WORKSPACE}/test/csit/${type}plans/${STREAMTESTPLAN}" @@ -139,17 +262,17 @@ function run_plan() { fi if [ -f "${plan_filepath}" ]; then - printf "${type} plan exists!!!\n" - printf "Changing the ${type} plan path...\n" + printf "%s plan exists!!!\n" "${type}" + printf "Changing the %s plan path...\n" "${type}" cat ${plan_filepath} | sed "s:integration:${WORKSPACE}:" > ${type}plan.txt cat ${type}plan.txt for line in $( egrep -v '(^[[:space:]]*#|^[[:space:]]*$)' ${type}plan.txt ); do - printf "Executing ${line}...\n" + printf "Executing %s...\n" "${line}" # shellcheck source=${line} disable=SC1091 source ${line} done fi - printf "Finished running ${type} plans\n" + printf "Finished running %s plans\n" "${type}" } # function run_plan() # Return elapsed time. Usage: @@ -159,7 +282,7 @@ function timer() { if [ $# -eq 0 ]; then # return the current time - printf "$(date "+%s")" + printf "%s" "$(date "+%s")" else local start_time=$1 end_time=$(date "+%s") @@ -240,13 +363,34 @@ PUBLIC_PHYSICAL_NETWORK: ${PUBLIC_PHYSICAL_NETWORK} ENABLE_NETWORKING_L2GW: ${ENABLE_NETWORKING_L2GW} CREATE_INITIAL_NETWORKS: ${CREATE_INITIAL_NETWORKS} LBAAS_SERVICE_PROVIDER: ${LBAAS_SERVICE_PROVIDER} -NUM_OPENSTACK_SITES: ${NUM_OPENSTACK_SITES} ODL_SFC_DRIVER: ${ODL_SFC_DRIVER} ODL_SNAT_MODE: ${ODL_SNAT_MODE} EOF } +function tcpdump_start() { + local -r prefix=$1 + local -r ip=$2 + local -r filter=$3 + filter_=${filter// /_} + + printf "node %s, %s_%s__%s: starting tcpdump\n" "${ip}" "${prefix}" "${ip}" "${filter}" + ssh ${ip} "nohup sudo /usr/sbin/tcpdump -vvv -ni eth0 ${filter} -w /tmp/tcpdump_${prefix}_${ip}__${filter_}.pcap > /tmp/tcpdump_start.log 2>&1 &" + ${SSH} ${ip} "ps -ef | grep tcpdump" +} + +function tcpdump_stop() { + local -r ip=$1 + + printf "node %s: stopping tcpdump\n" "$ip" + ${SSH} ${ip} "ps -ef | grep tcpdump.sh" + ${SSH} ${ip} "sudo pkill -f tcpdump" + ${SSH} ${ip} "sudo xz -9ekvvf /tmp/*.pcap" + ${SSH} ${ip} "sudo ls -al /tmp/*.pcap" + # copy_logs will copy any *.xz files +} + # Collect the list of files on the hosts function collect_files() { local -r ip=$1 @@ -295,7 +439,7 @@ function collect_openstack_logs() { local -r node_type=${3} local oslogs="${folder}/oslogs" - printf "collect_openstack_logs for ${node_type} node: ${ip} into ${oslogs}\n" + printf "collect_openstack_logs for %s node: %s into %s\n" "${node_type}" "${ip}" "${oslogs}" rm -rf ${oslogs} mkdir -p ${oslogs} # There are always some logs in /opt/stack/logs and this also covers the @@ -335,7 +479,7 @@ fi ls -al /tmp/oslogs EOF # cat > ${WORKSPACE}/collect_openstack_logs.sh << EOF - printf "collect_openstack_logs for ${node_type} node: ${ip} into ${oslogs}, executing script\n" + printf "collect_openstack_logs for %s node: %s into %s, executing script\n" "${node_type}" "${ip}" "${oslogs}" cat ${WORKSPACE}/collect_openstack_logs.sh scp ${WORKSPACE}/collect_openstack_logs.sh ${ip}:/tmp ${SSH} ${ip} "bash /tmp/collect_openstack_logs.sh > /tmp/collect_openstack_logs.log 2>&1" @@ -381,11 +525,15 @@ EOF mkdir -p ${WORKSPACE}/archives mv /tmp/changes.txt ${WORKSPACE}/archives + mv /tmp/validations.txt ${WORKSPACE}/archives mv ${WORKSPACE}/rabbit.txt ${WORKSPACE}/archives + mv ${WORKSPACE}/haproxy.cfg ${WORKSPACE}/archives + ssh ${OPENSTACK_HAPROXY_1_IP} "sudo journalctl -u haproxy > /tmp/haproxy.log" + scp ${OPENSTACK_HAPROXY_1_IP}:/tmp/haproxy.log ${WORKSPACE}/archives/ sleep 5 # FIXME: Do not create .tar and gzip before copying. - for i in `seq 1 ${NUM_ODL_SYSTEM}`; do + for i in $(seq 1 "${NUM_ODL_SYSTEM}"); do CONTROLLERIP=ODL_SYSTEM_${i}_IP echo "collect_logs: for opendaylight controller ip: ${!CONTROLLERIP}" NODE_FOLDER="odl_${i}" @@ -424,7 +572,7 @@ EOF print_job_parameters > ${WORKSPACE}/archives/params.txt # Control Node - for i in `seq 1 ${NUM_OPENSTACK_CONTROL_NODES}`; do + for i in $(seq 1 "${NUM_OPENSTACK_CONTROL_NODES}"); do OSIP=OPENSTACK_CONTROL_NODE_${i}_IP if [ "$(is_openstack_feature_enabled n-cpu)" == "1" ]; then echo "collect_logs: for openstack combo node ip: ${!OSIP}" @@ -434,6 +582,7 @@ EOF NODE_FOLDER="control_${i}" fi mkdir -p ${NODE_FOLDER} + tcpdump_stop "${!OSIP}" scp extra_debug.sh ${!OSIP}:/tmp # Capture compute logs if this is a combo node if [ "$(is_openstack_feature_enabled n-cpu)" == "1" ]; then @@ -469,8 +618,10 @@ EOF scp ${!OSIP}:/tmp/dmesg.log ${NODE_FOLDER} scp ${!OSIP}:/tmp/extra_debug.log ${NODE_FOLDER} scp ${!OSIP}:/tmp/get_devstack.sh.txt ${NODE_FOLDER} + scp ${!OSIP}:/tmp/install_ovs.txt ${NODE_FOLDER} scp ${!OSIP}:/tmp/journalctl.log ${NODE_FOLDER} scp ${!OSIP}:/tmp/ovsdb-tool.log ${NODE_FOLDER} + scp ${!OSIP}:/tmp/tcpdump_start.log ${NODE_FOLDER} collect_files "${!OSIP}" "${NODE_FOLDER}" ${SSH} ${!OSIP} "sudo tar -cf - -C /var/log rabbitmq | xz -T 0 > /tmp/rabbitmq.tar.xz " scp ${!OSIP}:/tmp/rabbitmq.tar.xz ${NODE_FOLDER} @@ -491,11 +642,12 @@ EOF done # Compute Nodes - for i in `seq 1 ${NUM_OPENSTACK_COMPUTE_NODES}`; do + for i in $(seq 1 "${NUM_OPENSTACK_COMPUTE_NODES}"); do OSIP=OPENSTACK_COMPUTE_NODE_${i}_IP echo "collect_logs: for openstack compute node ip: ${!OSIP}" NODE_FOLDER="compute_${i}" mkdir -p ${NODE_FOLDER} + tcpdump_stop "${!OSIP}" scp extra_debug.sh ${!OSIP}:/tmp ${SSH} ${!OSIP} "bash /tmp/extra_debug.sh > /tmp/extra_debug.log 2>&1" scp ${!OSIP}:/etc/nova/nova.conf ${NODE_FOLDER} @@ -509,8 +661,10 @@ EOF scp ${!OSIP}:/tmp/dmesg.log ${NODE_FOLDER} scp ${!OSIP}:/tmp/extra_debug.log ${NODE_FOLDER} scp ${!OSIP}:/tmp/get_devstack.sh.txt ${NODE_FOLDER} + scp ${!OSIP}:/tmp/install_ovs.txt ${NODE_FOLDER} scp ${!OSIP}:/tmp/journalctl.log ${NODE_FOLDER} scp ${!OSIP}:/tmp/ovsdb-tool.log ${NODE_FOLDER} + scp ${!OSIP}:/tmp/tcpdump_start.log ${NODE_FOLDER} collect_files "${!OSIP}" "${NODE_FOLDER}" ${SSH} ${!OSIP} "sudo tar -cf - -C /var/log libvirt | xz -T 0 > /tmp/libvirt.tar.xz " scp ${!OSIP}:/tmp/libvirt.tar.xz ${NODE_FOLDER} @@ -540,3 +694,396 @@ EOF echo "tempest results not found in ${DEVSTACK_TEMPEST_DIR}/${TESTREPO}/0" fi } # collect_netvirt_logs() + +# Utility function for joining strings. +function join() { + delim=' ' + final=$1; shift + + for str in "$@" ; do + final=${final}${delim}${str} + done + + echo ${final} +} + +function get_nodes_list() { + # Create the string for nodes + for i in $(seq 1 "${NUM_ODL_SYSTEM}") ; do + CONTROLLERIP=ODL_SYSTEM_${i}_IP + nodes[$i]=${!CONTROLLERIP} + done + + nodes_list=$(join "${nodes[@]}") + echo ${nodes_list} +} + +function get_features() { + if [ ${CONTROLLERSCOPE} == 'all' ]; then + ACTUALFEATURES="odl-integration-compatible-with-all,${CONTROLLERFEATURES}" + export CONTROLLERMEM="3072m" + else + ACTUALFEATURES="odl-infrautils-ready,${CONTROLLERFEATURES}" + fi + + # Some versions of jenkins job builder result in feature list containing spaces + # and ending in newline. Remove all that. + ACTUALFEATURES=$(echo "${ACTUALFEATURES}" | tr -d '\n \r') + echo "ACTUALFEATURES: ${ACTUALFEATURES}" + + # In the case that we want to install features via karaf shell, a space separated list of + # ACTUALFEATURES IS NEEDED + SPACE_SEPARATED_FEATURES=$(echo "${ACTUALFEATURES}" | tr ',' ' ') + echo "SPACE_SEPARATED_FEATURES: ${SPACE_SEPARATED_FEATURES}" + + export ACTUALFEATURES + export SPACE_SEPARATED_FEATURES +} + +# Create the configuration script to be run on controllers. +function create_configuration_script() { + cat > ${WORKSPACE}/configuration-script.sh < ${WORKSPACE}/configuration-script.sh < ${WORKSPACE}/startup-script.sh < ${WORKSPACE}/startup-script.sh < ${WORKSPACE}/post-startup-script.sh < ${WORKSPACE}/post-startup-script.sh < ${WORKSPACE}/ps_before.log + pid=$(grep org.apache.karaf.main.Main ${WORKSPACE}/ps_before.log | grep -v grep | tr -s ' ' | cut -f2 -d' ') + echo "karaf main: org.apache.karaf.main.Main, pid:${pid}" + ssh ${!CONTROLLERIP} "${JAVA_HOME}/bin/jstack -l ${pid}" > ${WORKSPACE}/karaf_${i}_${pid}_threads_before.log || true + done +} + +# Function to build OVS from git repo +function build_ovs() { + local -r ip=$1 + local -r version=$2 + local -r rpm_path="$3" + + echo "Building OVS ${version} on ${ip} ..." + cat > "${WORKSPACE}"/build_ovs.sh << EOF +set -ex -o pipefail + +echo '---> Building openvswitch version ${version}' + +# Install running kernel devel packages +K_VERSION=\$(uname -r) +YUM_OPTS="-y --disablerepo=* --enablerepo=base,updates,extra,C*-base,C*-updates,C*-extras" +# Install centos-release to update vault repos from which to fetch +# kernel devel packages +sudo yum \${YUM_OPTS} install centos-release yum-utils @'Development Tools' rpm-build +sudo yum \${YUM_OPTS} install kernel-{devel,headers}-\${K_VERSION} + +TMP=\$(mktemp -d) +pushd \${TMP} + +git clone https://github.com/openvswitch/ovs.git +cd ovs + +if [ "${version}" = "v2.6.1-nsh" ]; then + git checkout v2.6.1 + echo "Will apply nsh patches for OVS version 2.6.1" + git clone https://github.com/yyang13/ovs_nsh_patches.git ../ovs_nsh_patches + git apply ../ovs_nsh_patches/v2.6.1_centos7/*.patch +else + git checkout ${version} +fi + +# On early versions of OVS, flake warnings would fail the build. +# Remove it. +sudo pip uninstall -y flake8 + +# Get rid of sphinx dep as it conflicts with the already +# installed one (via pip). Docs wont be built. +sed -i "/BuildRequires:.*sphinx.*/d" rhel/openvswitch-fedora.spec.in + +sed -e 's/@VERSION@/0.0.1/' rhel/openvswitch-fedora.spec.in > /tmp/ovs.spec +sed -e 's/@VERSION@/0.0.1/' rhel/openvswitch-kmod-fedora.spec.in > /tmp/ovs-kmod.spec +sed -e 's/@VERSION@/0.0.1/' rhel/openvswitch-dkms.spec.in > /tmp/ovs-dkms.spec +sudo yum-builddep \${YUM_OPTS} /tmp/ovs.spec /tmp/ovs-kmod.spec /tmp/ovs-dkms.spec +rm /tmp/ovs.spec /tmp/ovs-kmod.spec /tmp/ovs-dkms.spec +./boot.sh +./configure --build=x86_64-redhat-linux-gnu --host=x86_64-redhat-linux-gnu --with-linux=/lib/modules/\${K_VERSION}/build --program-prefix= --disable-dependency-tracking --prefix=/usr --exec-prefix=/usr --bindir=/usr/bin --sbindir=/usr/sbin --sysconfdir=/etc --datadir=/usr/share --includedir=/usr/include --libdir=/usr/lib64 --libexecdir=/usr/libexec --localstatedir=/var --sharedstatedir=/var/lib --mandir=/usr/share/man --infodir=/usr/share/info --enable-libcapng --enable-ssl --with-pkidir=/var/lib/openvswitch/pki PYTHON=/usr/bin/python2 +make rpm-fedora RPMBUILD_OPT="--without check" +# Build dkms only for now +# make rpm-fedora-kmod RPMBUILD_OPT='-D "kversion \${K_VERSION}"' +rpmbuild -D "_topdir \$(pwd)/rpm/rpmbuild" -bb --without check rhel/openvswitch-dkms.spec + +mkdir -p /tmp/ovs_rpms +cp -r rpm/rpmbuild/RPMS/* /tmp/ovs_rpms/ + +popd +rm -rf \${TMP} +EOF + + scp "${WORKSPACE}"/build_ovs.sh "${ip}":/tmp + ${SSH} "${ip}" " bash /tmp/build_ovs.sh >> /tmp/install_ovs.txt 2>&1" + scp -r "${ip}":/tmp/ovs_rpms/* "${rpm_path}/" + ${SSH} "${ip}" "rm -rf /tmp/ovs_rpms" +} + +# Install OVS RPMs from yum repo +function install_ovs_from_repo() { + local -r ip=$1 + local -r rpm_repo="$2" + + echo "Installing OVS from repo ${rpm_repo} on ${ip} ..." + cat > "${WORKSPACE}"/install_ovs.sh << EOF +set -ex -o pipefail + +echo '---> Installing openvswitch from ${rpm_repo}' + +# We need repoquery from yum-utils. +sudo yum -y install yum-utils + +# Get openvswitch packages offered by custom repo. +# dkms package will have priority over kmod. +OVS_REPO_OPTS="--repofrompath=ovs-repo,${rpm_repo} --disablerepo=* --enablerepo=ovs-repo" +OVS_PKGS=\$(repoquery \${OVS_REPO_OPTS} openvswitch) +OVS_SEL_PKG=\$(repoquery \${OVS_REPO_OPTS} openvswitch-selinux-policy) +OVS_DKMS_PKG=\$(repoquery \${OVS_REPO_OPTS} openvswitch-dkms) +OVS_KMOD_PKG=\$(repoquery \${OVS_REPO_OPTS} openvswitch-kmod) +[ -n "\${OVS_SEL_PKG}" ] && OVS_PKGS="\${OVS_PKGS} \${OVS_SEL_PKG}" +[ -n "\${OVS_DKMS_PKG}" ] && OVS_PKGS="\${OVS_PKGS} \${OVS_DKMS_PKG}" +[ -z "\${OVS_DKMS_PKG}" ] && [ -n "\${OVS_KMOD_PKG}" ] && OVS_PKGS="\${OVS_PKGS} \${OVS_KMOD_PKG}" + +# Bail with error if custom repo was provided but we could not +# find suitable packages there. +[ -z "\${OVS_PKGS}" ] && echo "No OVS packages found in custom repo." && exit 1 + +# Install kernel & devel packages for the openvswitch dkms package. +if [ -n "\${OVS_DKMS_PKG}" ]; then + # install centos-release to update vault repos from which to fetch + # kernel devel packages + sudo yum -y install centos-release + K_VERSION=\$(uname -r) + YUM_OPTS="-y --disablerepo=* --enablerepo=base,updates,extra,C*-base,C*-updates,C*-extras" + sudo yum \${YUM_OPTS} install kernel-{headers,devel}-\${K_VERSION} @'Development Tools' python-six +fi + +PREV_MOD=\$(sudo modinfo -n openvswitch || echo '') + +# Install OVS offered by custom repo. +sudo yum-config-manager --add-repo "${rpm_repo}" +sudo yum -y versionlock delete openvswitch-* +sudo yum -y remove openvswitch-* +sudo yum -y --nogpgcheck install \${OVS_PKGS} +sudo yum -y versionlock add \${OVS_PKGS} + +# Most recent OVS versions have some incompatibility with certain versions of iptables +# This below line will overcome that problem. +sudo modprobe openvswitch + +# Start OVS and print details +sudo systemctl start openvswitch +sudo systemctl enable openvswitch +sudo ovs-vsctl --retry -t 5 show +sudo modinfo openvswitch + +# dkms rpm install can fail silently (probably because the OVS version is +# incompatible with the running kernel), verify module was updated. +NEW_MOD=\$(sudo modinfo -n openvswitch || echo '') +[ "\${PREV_MOD}" != "\${NEW_MOD}" ] || (echo "Kernel module was not updated" && exit 1) +EOF + + scp "${WORKSPACE}"/install_ovs.sh "${ip}":/tmp + ${SSH} "${ip}" "bash /tmp/install_ovs.sh >> /tmp/install_ovs.txt 2>&1" +} + +# Install OVS RPMS from path +function install_ovs_from_path() { + local -r ip=$1 + local -r rpm_path="$2" + + echo "Creating OVS RPM repo on ${ip} ..." + ${SSH} "${ip}" "mkdir -p /tmp/ovs_rpms" + scp -r "${rpm_path}"/* "${ip}":/tmp/ovs_rpms + ${SSH} "${ip}" "sudo yum -y install createrepo && createrepo --database /tmp/ovs_rpms" + install_ovs_from_repo "${ip}" file:/tmp/ovs_rpms +} + +