X-Git-Url: https://git.opendaylight.org/gerrit/gitweb?a=blobdiff_plain;f=jjb%2Fintegration%2Fintegration-deploy-openstack-run-test.sh;h=f8d0910800e0daa71f053e20d3950a8f932b6464;hb=f63efd2b7c5e453af225a64b158e03e5a9bdbc28;hp=0ae2bd2522bdb3823836170e833cd7dc880947bc;hpb=bc4c42944a7d331f69dc95a91946bcb5ba6b20bf;p=releng%2Fbuilder.git diff --git a/jjb/integration/integration-deploy-openstack-run-test.sh b/jjb/integration/integration-deploy-openstack-run-test.sh index 0ae2bd252..3beb208db 100644 --- a/jjb/integration/integration-deploy-openstack-run-test.sh +++ b/jjb/integration/integration-deploy-openstack-run-test.sh @@ -1,10 +1,12 @@ -#@IgnoreInspection BashAddShebang +#!/bin/bash # Activate robotframework virtualenv # ${ROBOT_VENV} comes from the integration-install-robotframework.sh # script. # shellcheck source=${ROBOT_VENV}/bin/activate disable=SC1091 source ${ROBOT_VENV}/bin/activate PYTHON="${ROBOT_VENV}/bin/python" +SSH="ssh -t -t" +ADMIN_PASSWORD=admin # TODO: remove this work to run changes.py if/when it's moved higher up to be visible at the Robot level echo "showing recent changes that made it in to the distribution used by this job" @@ -12,91 +14,153 @@ $PYTHON -m pip install --upgrade urllib3 python ${WORKSPACE}/test/tools/distchanges/changes.py -d /tmp/distribution_folder \ -u ${ACTUAL_BUNDLE_URL} -b ${DISTROBRANCH} \ -r ssh://jenkins-${SILO}@git.opendaylight.org:29418 || true +cat << EOF +################################################# +## Deploy Openstack 3-node ## +################################################# +EOF -echo "#################################################" -echo "## Deploy Openstack 3-node ##" -echo "#################################################" +# Catch command errors and collect logs. +# This ensures logs are collected when script commands fail rather than simply exiting. +function trap_handler() { + local prog="$0" + local lastline="$1" + local lasterr="$2" + echo "${prog}: line ${lastline}: exit status of last command: ${lasterr}" + echo "command: ${BASH_COMMAND}" + collect_logs + exit 1 +} # trap_handler() + +trap 'trap_handler ${LINENO} ${$?}' ERR + +function print_job_parameters() { + cat << EOF + +Job parameters: +DISTROBRANCH: ${DISTROBRANCH} +DISTROSTREAM: ${DISTROSTREAM} +BUNDLE_URL: ${BUNDLE_URL} +CONTROLLERFEATURES: ${CONTROLLERFEATURES} +CONTROLLERDEBUGMAP: ${CONTROLLERDEBUGMAP} +TESTPLAN: ${TESTPLAN} +SUITES: ${SUITES} +PATCHREFSPEC: ${PATCHREFSPEC} +OPENSTACK_BRANCH: ${OPENSTACK_BRANCH} +DEVSTACK_HASH: ${DEVSTACK_HASH} +ODL_ML2_DRIVER_REPO: ${ODL_ML2_DRIVER_REPO} +ODL_ML2_BRANCH: ${ODL_ML2_BRANCH} +ODL_ML2_DRIVER_VERSION: ${ODL_ML2_DRIVER_VERSION} +ODL_ML2_PORT_BINDING: ${ODL_ML2_PORT_BINDING} +DEVSTACK_KUBERNETES_PLUGIN_REPO: ${DEVSTACK_KUBERNETES_PLUGIN_REPO} +DEVSTACK_LBAAS_PLUGIN_REPO: ${DEVSTACK_LBAAS_PLUGIN_REPO} +DEVSTACK_NETWORKING_SFC_PLUGIN_REPO: ${DEVSTACK_NETWORKING_SFC_PLUGIN_REPO} +ODL_ENABLE_L3_FWD: ${ODL_ENABLE_L3_FWD} +IPSEC_VXLAN_TUNNELS_ENABLED: ${IPSEC_VXLAN_TUNNELS_ENABLED} +PUBLIC_BRIDGE: ${PUBLIC_BRIDGE} +ENABLE_HAPROXY_FOR_NEUTRON: ${ENABLE_HAPROXY_FOR_NEUTRON} +ENABLE_OS_SERVICES: ${ENABLE_OS_SERVICES} +ENABLE_OS_COMPUTE_SERVICES: ${ENABLE_OS_COMPUTE_SERVICES} +ENABLE_OS_PLUGINS: ${ENABLE_OS_PLUGINS} +DISABLE_OS_SERVICES: ${DISABLE_OS_SERVICES} +TENANT_NETWORK_TYPE: ${TENANT_NETWORK_TYPE} +SECURITY_GROUP_MODE: ${SECURITY_GROUP_MODE} +PUBLIC_PHYSICAL_NETWORK: ${PUBLIC_PHYSICAL_NETWORK} +ENABLE_NETWORKING_L2GW: ${ENABLE_NETWORKING_L2GW} +CREATE_INITIAL_NETWORKS: ${CREATE_INITIAL_NETWORKS} +LBAAS_SERVICE_PROVIDER: ${LBAAS_SERVICE_PROVIDER} +NUM_OPENSTACK_SITES: ${NUM_OPENSTACK_SITES} +ODL_SFC_DRIVER: ${ODL_SFC_DRIVER} +ODL_SNAT_MODE: ${ODL_SNAT_MODE} +EOF +} -SSH="ssh -t -t" +print_job_parameters -function create_control_node_local_conf { -HOSTIP=$1 -MGRIP=$2 -ODL_OVS_MANAGERS="$3" -#Needs to be removed -if [ "${ODL_ML2_BRANCH}" != "stable/ocata" ]; then - RECLONE=no -else - RECLONE=yes -fi -local_conf_file_name=${WORKSPACE}/local.conf_control_${HOSTIP} -cat > ${local_conf_file_name} << EOF +function create_etc_hosts() { + NODE_IP=$1 + CTRL_IP=$2 + : > ${WORKSPACE}/hosts_file + for iter in `seq 1 ${NUM_OPENSTACK_COMPUTE_NODES}` + do + COMPUTE_IP=OPENSTACK_COMPUTE_NODE_${iter}_IP + if [ "${!COMPUTE_IP}" == "${NODE_IP}" ]; then + CONTROL_HNAME=$(${SSH} ${CTRL_IP} "hostname") + echo "${CTRL_IP} ${CONTROL_HNAME}" >> ${WORKSPACE}/hosts_file + else + COMPUTE_HNAME=$(${SSH} ${!COMPUTE_IP} "hostname") + echo "${!COMPUTE_IP} ${COMPUTE_HNAME}" >> ${WORKSPACE}/hosts_file + fi + done + + echo "Created the hosts file for ${NODE_IP}:" + cat ${WORKSPACE}/hosts_file +} # create_etc_hosts() + +# convert commas in csv strings to spaces (ssv) +function csv2ssv() { + local csv=$1 + if [ -n "${csv}" ]; then + ssv=$(echo ${csv} | sed 's/,/ /g' | sed 's/\ \ */\ /g') + fi + + echo "${ssv}" +} # csv2ssv + +# Add enable_services and disable_services to the local.conf +function add_os_services() { + local core_services=$1 + local enable_services=$2 + local disable_services=$3 + local local_conf_file_name=$4 + + cat >> ${local_conf_file_name} << EOF +enable_service $(csv2ssv "${core_services}") +EOF + if [ -n "${enable_services}" ]; then + cat >> ${local_conf_file_name} << EOF +enable_service $(csv2ssv "${enable_services}") +EOF + fi + if [ -n "${disable_services}" ]; then + cat >> ${local_conf_file_name} << EOF +disable_service $(csv2ssv "${disable_services}") +EOF + fi +} + +function create_control_node_local_conf() { + HOSTIP=$1 + MGRIP=$2 + ODL_OVS_MANAGERS="$3" + + local_conf_file_name=${WORKSPACE}/local.conf_control_${HOSTIP} + cat > ${local_conf_file_name} << EOF [[local|localrc]] LOGFILE=stack.sh.log USE_SCREEN=True SCREEN_LOGDIR=/opt/stack/data/log LOG_COLOR=False RECLONE=${RECLONE} -ETCD_PORT=2379 -EOF -IFS=, -for service_name in ${DISABLE_OS_SERVICES} -do -cat >> ${local_conf_file_name} << EOF -disable_service ${service_name} +disable_all_services EOF -done -for service_name in ${ENABLE_OS_SERVICES} -do -cat >> ${local_conf_file_name} << EOF -enable_service ${service_name} -EOF -done -for plugin_name in ${ENABLE_OS_PLUGINS} -do -if [ "$plugin_name" == "networking-odl" ]; then - ENABLE_PLUGIN_ARGS="${ODL_ML2_DRIVER_REPO} ${ODL_ML2_BRANCH}" -elif [ "$plugin_name" == "kuryr-kubernetes" ]; then - ENABLE_PLUGIN_ARGS="${DEVSTACK_KUBERNETES_PLUGIN_REPO} master" # note: kuryr-kubernetes only exists in master at the moment - IS_KUBERNETES_PLUGIN_ENABLED="yes" -elif [ "$plugin_name" == "neutron-lbaas" ]; then - ENABLE_PLUGIN_ARGS="${DEVSTACK_LBAAS_PLUGIN_REPO} ${OPENSTACK_BRANCH}" - IS_LBAAS_PLUGIN_ENABLED="yes" -elif [ "$plugin_name" == "networking-sfc" ]; then - ENABLE_PLUGIN_ARGS="${DEVSTACK_NETWORKING_SFC_PLUGIN_REPO} ${OPENSTACK_BRANCH}" -else - echo "Error: Invalid plugin $plugin_name, unsupported" - continue -fi -cat >> ${local_conf_file_name} << EOF -enable_plugin ${plugin_name} ${ENABLE_PLUGIN_ARGS} -EOF -done -unset IFS -if [ "${OPENSTACK_BRANCH}" == "master" ] || [ "${OPENSTACK_BRANCH}" == "stable/ocata" ]; then # Ocata+ - # placement is mandatory for nova since Ocata, note that this requires computes to enable placement-client - # this should be moved into enabled_services for each job (but only for Ocata) - echo "enable_service placement-api" >> ${local_conf_file_name} -fi -if [ "${OPENSTACK_BRANCH}" == "stable/ocata" ]; then # Ocata - # running kubernetes master against devstack ocata has some issues with etcd3 and this - # workaround is needed for things to work - if [ "$IS_KUBERNETES_PLUGIN_ENABLED" == "yes" ]; then - echo "disable_service etcd3" >> ${local_conf_file_name} - fi -fi -cat >> ${local_conf_file_name} << EOF + + add_os_services "${CORE_OS_CONTROL_SERVICES}" "${ENABLE_OS_SERVICES}" "${DISABLE_OS_SERVICES}" "${local_conf_file_name}" + + cat >> ${local_conf_file_name} << EOF + HOST_IP=${HOSTIP} SERVICE_HOST=\$HOST_IP - -NEUTRON_CREATE_INITIAL_NETWORKS=${CREATE_INITIAL_NETWORKS} -Q_PLUGIN=ml2 Q_ML2_TENANT_NETWORK_TYPE=${TENANT_NETWORK_TYPE} -Q_OVS_USE_VETH=True +NEUTRON_CREATE_INITIAL_NETWORKS=${CREATE_INITIAL_NETWORKS} -ENABLE_TENANT_TUNNELS=True +ODL_MODE=manual +ODL_MGR_IP=${MGRIP} +ODL_PORT=8080 +ODL_PORT_BINDING_CONTROLLER=${ODL_ML2_PORT_BINDING} +ODL_OVS_MANAGERS=${ODL_OVS_MANAGERS} MYSQL_HOST=\$SERVICE_HOST RABBIT_HOST=\$SERVICE_HOST @@ -104,86 +168,80 @@ GLANCE_HOSTPORT=\$SERVICE_HOST:9292 KEYSTONE_AUTH_HOST=\$SERVICE_HOST KEYSTONE_SERVICE_HOST=\$SERVICE_HOST -MYSQL_PASSWORD=mysql -RABBIT_PASSWORD=rabbit -SERVICE_TOKEN=service -SERVICE_PASSWORD=admin -ADMIN_PASSWORD=admin - -ODL_PORT=8080 -ODL_MODE=externalodl -ODL_PORT_BINDING_CONTROLLER=${ODL_ML2_PORT_BINDING} - -LIBVIRT_TYPE=qemu -ODL_MGR_IP=${MGRIP} +ADMIN_PASSWORD=${ADMIN_PASSWORD} +DATABASE_PASSWORD=${ADMIN_PASSWORD} +RABBIT_PASSWORD=${ADMIN_PASSWORD} +SERVICE_TOKEN=${ADMIN_PASSWORD} +SERVICE_PASSWORD=${ADMIN_PASSWORD} NEUTRON_LBAAS_SERVICE_PROVIDERV2=${LBAAS_SERVICE_PROVIDER} # Only relevant if neutron-lbaas plugin is enabled NEUTRON_SFC_DRIVERS=${ODL_SFC_DRIVER} # Only relevant if networking-sfc plugin is enabled NEUTRON_FLOWCLASSIFIER_DRIVERS=${ODL_SFC_DRIVER} # Only relevant if networking-sfc plugin is enabled +ETCD_PORT=2379 EOF -if [ "${ENABLE_NETWORKING_L2GW}" == "yes" ]; then -cat >> ${local_conf_file_name} << EOF + if [ "${ODL_ML2_DRIVER_VERSION}" == "v2" ]; then + echo "ODL_V2DRIVER=True" >> ${local_conf_file_name} + fi -enable_plugin networking-l2gw ${NETWORKING_L2GW_DRIVER} ${ODL_ML2_BRANCH} -NETWORKING_L2GW_SERVICE_DRIVER=L2GW:OpenDaylight:networking_odl.l2gateway.driver.OpenDaylightL2gwDriver:default -ENABLED_SERVICES+=,neutron,q-svc,nova,q-meta + IFS=, + for plugin_name in ${ENABLE_OS_PLUGINS}; do + if [ "$plugin_name" == "networking-odl" ]; then + ENABLE_PLUGIN_ARGS="${ODL_ML2_DRIVER_REPO} ${ODL_ML2_BRANCH}" + elif [ "$plugin_name" == "kuryr-kubernetes" ]; then + ENABLE_PLUGIN_ARGS="${DEVSTACK_KUBERNETES_PLUGIN_REPO} master" # note: kuryr-kubernetes only exists in master at the moment + elif [ "$plugin_name" == "neutron-lbaas" ]; then + ENABLE_PLUGIN_ARGS="${DEVSTACK_LBAAS_PLUGIN_REPO} ${OPENSTACK_BRANCH}" + IS_LBAAS_PLUGIN_ENABLED="yes" + elif [ "$plugin_name" == "networking-sfc" ]; then + ENABLE_PLUGIN_ARGS="${DEVSTACK_NETWORKING_SFC_PLUGIN_REPO} ${OPENSTACK_BRANCH}" + else + echo "Error: Invalid plugin $plugin_name, unsupported" + continue + fi + cat >> ${local_conf_file_name} << EOF +enable_plugin ${plugin_name} ${ENABLE_PLUGIN_ARGS} EOF -fi - -if [ "${ODL_ML2_DRIVER_VERSION}" == "v2" ]; then - echo "ODL_V2DRIVER=True" >> ${local_conf_file_name} -fi + done + unset IFS -echo "ODL_OVS_MANAGERS=${ODL_OVS_MANAGERS}" >> ${local_conf_file_name} + if [ "${ENABLE_NETWORKING_L2GW}" == "yes" ]; then + cat >> ${local_conf_file_name} << EOF -# if we are using the old netvirt impl, as determined by the feature name -# odl-ovsdb-openstack (note: new impl is odl-netvirt-openstack) then we -# want ODL_L3 to be True. New impl wants it False -if [[ ${CONTROLLERFEATURES} == *"odl-ovsdb-openstack"* ]]; then - ODL_L3=True -else - ODL_L3=False -fi +enable_plugin networking-l2gw ${NETWORKING_L2GW_DRIVER} ${ODL_ML2_BRANCH} +NETWORKING_L2GW_SERVICE_DRIVER=L2GW:OpenDaylight:networking_odl.l2gateway.driver.OpenDaylightL2gwDriver:default +EOF + fi -# if we are using the new netvirt impl, as determined by the feature name -# odl-netvirt-openstack (note: old impl is odl-ovsdb-openstack) then we -# want PROVIDER_MAPPINGS to be used -- this should be fixed if we want to support -# external networks in legacy netvirt -if [[ ${CONTROLLERFEATURES} == *"odl-netvirt-openstack"* ]]; then - ODL_PROVIDER_MAPPINGS="\${PUBLIC_PHYSICAL_NETWORK}:${PUBLIC_BRIDGE}" -else - ODL_PROVIDER_MAPPINGS= -fi + if [ "${ODL_ENABLE_L3_FWD}" == "yes" ]; then + cat >> ${local_conf_file_name} << EOF -if [ "${ODL_ENABLE_L3_FWD}" == "yes" ]; then -cat >> ${local_conf_file_name} << EOF PUBLIC_BRIDGE=${PUBLIC_BRIDGE} PUBLIC_PHYSICAL_NETWORK=${PUBLIC_PHYSICAL_NETWORK} ML2_VLAN_RANGES=${PUBLIC_PHYSICAL_NETWORK} ODL_PROVIDER_MAPPINGS=${ODL_PROVIDER_MAPPINGS} - -disable_service q-l3 -PUBLIC_INTERFACE=br100 EOF -SERVICE_PLUGINS="networking_odl.l3.l3_odl.OpenDaylightL3RouterPlugin" -if [ "${ENABLE_NETWORKING_L2GW}" == "yes" ]; then - SERVICE_PLUGINS+=", networking_l2gw.services.l2gateway.plugin.L2GatewayPlugin" -fi #check for ENABLE_NETWORKING_L2GW -if [ "${IS_LBAAS_PLUGIN_ENABLED}" == "yes" ]; then - SERVICE_PLUGINS+=", lbaasv2" -fi #check for ENABLE_LBAAS_PLUGIN -fi #check for ODL_ENABLE_L3_FWD + if [ "${ODL_ML2_DRIVER_VERSION}" == "v2" ]; then + SERVICE_PLUGINS="odl-router_v2" + else + SERVICE_PLUGINS="odl-router" + fi + if [ "${ENABLE_NETWORKING_L2GW}" == "yes" ]; then + SERVICE_PLUGINS+=", networking_l2gw.services.l2gateway.plugin.L2GatewayPlugin" + fi + if [ "${IS_LBAAS_PLUGIN_ENABLED}" == "yes" ]; then + SERVICE_PLUGINS+=", lbaasv2" + fi + fi #check for ODL_ENABLE_L3_FWD + + cat >> ${local_conf_file_name} << EOF -cat >> ${local_conf_file_name} << EOF [[post-config|\$NEUTRON_CONF]] [DEFAULT] service_plugins = ${SERVICE_PLUGINS} -EOF -cat >> ${local_conf_file_name} << EOF [[post-config|/etc/neutron/plugins/ml2/ml2_conf.ini]] [agent] minimize_polling=True @@ -194,6 +252,10 @@ minimize_polling=True physical_network_mtus = ${PUBLIC_PHYSICAL_NETWORK}:1440 path_mtu = 1490 +# workaround for port-status not working due to https://bugs.opendaylight.org/show_bug.cgi?id=9092 +[ml2_odl] +odl_features=nothing + [[post-config|/etc/neutron/dhcp_agent.ini]] [DEFAULT] force_metadata = True @@ -203,32 +265,22 @@ enable_isolated_metadata = True [DEFAULT] force_config_drive = False +[scheduler] +discover_hosts_in_cells_interval = 30 EOF -echo "local.conf Created...." -cat ${local_conf_file_name} -} + echo "Control local.conf created:" + cat ${local_conf_file_name} +} # create_control_node_local_conf() -function create_compute_node_local_conf { -HOSTIP=$1 -SERVICEHOST=$2 -MGRIP=$3 -ODL_OVS_MANAGERS="$4" -#Needs to be removed -if [ "${ODL_ML2_BRANCH}" != "stable/ocata" ]; then - RECLONE=no -else - RECLONE=yes -fi -if [ "${OPENSTACK_BRANCH}" == "master" ] || [ "${OPENSTACK_BRANCH}" == "stable/ocata" ]; then # Ocata+ - # placement is mandatory for nova since Ocata, note that this requires controller to enable placement-api - ENABLED_SERVICES="n-cpu,placement-client" -else - ENABLED_SERVICES="n-cpu" -fi +function create_compute_node_local_conf() { + HOSTIP=$1 + SERVICEHOST=$2 + MGRIP=$3 + ODL_OVS_MANAGERS="$4" -local_conf_file_name=${WORKSPACE}/local.conf_compute_${HOSTIP} -cat > ${local_conf_file_name} << EOF + local_conf_file_name=${WORKSPACE}/local.conf_compute_${HOSTIP} + cat > ${local_conf_file_name} << EOF [[local|localrc]] LOGFILE=stack.sh.log LOG_COLOR=False @@ -236,15 +288,22 @@ USE_SCREEN=True SCREEN_LOGDIR=/opt/stack/data/log RECLONE=${RECLONE} -NOVA_VNC_ENABLED=True -MULTI_HOST=1 -ENABLED_SERVICES=${ENABLED_SERVICES} +disable_all_services +EOF + + add_os_services "${CORE_OS_COMPUTE_SERVICES}" "${ENABLE_OS_COMPUTE_SERVICES}" "${DISABLE_OS_SERVICES}" "${local_conf_file_name}" + + cat >> ${local_conf_file_name} << EOF + HOST_IP=${HOSTIP} SERVICE_HOST=${SERVICEHOST} +Q_ML2_TENANT_NETWORK_TYPE=${TENANT_NETWORK_TYPE} -Q_PLUGIN=ml2 -ENABLE_TENANT_TUNNELS=True -Q_ML2_TENANT_NETWORK_TYPE=vxlan +ODL_MODE=manual +ODL_MGR_IP=${MGRIP} +ODL_PORT=8080 +ODL_PORT_BINDING_CONTROLLER=${ODL_ML2_PORT_BINDING} +ODL_OVS_MANAGERS=${ODL_OVS_MANAGERS} Q_HOST=\$SERVICE_HOST MYSQL_HOST=\$SERVICE_HOST @@ -253,56 +312,49 @@ GLANCE_HOSTPORT=\$SERVICE_HOST:9292 KEYSTONE_AUTH_HOST=\$SERVICE_HOST KEYSTONE_SERVICE_HOST=\$SERVICE_HOST -MYSQL_PASSWORD=mysql -RABBIT_PASSWORD=rabbit -SERVICE_TOKEN=service -SERVICE_PASSWORD=admin -ADMIN_PASSWORD=admin - -ODL_MODE=compute -ODL_PORT_BINDING_CONTROLLER=${ODL_ML2_PORT_BINDING} -LIBVIRT_TYPE=qemu -ODL_MGR_IP=${MGRIP} +ADMIN_PASSWORD=${ADMIN_PASSWORD} +DATABASE_PASSWORD=${ADMIN_PASSWORD} +RABBIT_PASSWORD=${ADMIN_PASSWORD} +SERVICE_TOKEN=${ADMIN_PASSWORD} +SERVICE_PASSWORD=${ADMIN_PASSWORD} EOF -if [[ "${ENABLE_OS_PLUGINS}" =~ networking-odl ]]; then -cat >> ${local_conf_file_name} << EOF + if [[ "${ENABLE_OS_PLUGINS}" =~ networking-odl ]]; then + cat >> ${local_conf_file_name} << EOF + enable_plugin networking-odl ${ODL_ML2_DRIVER_REPO} ${ODL_ML2_BRANCH} EOF -fi - -echo "ODL_OVS_MANAGERS=${ODL_OVS_MANAGERS}" >> ${local_conf_file_name} + fi -# if we are using the new netvirt impl, as determined by the feature name -# odl-netvirt-openstack (note: old impl is odl-ovsdb-openstack) then we -# want PROVIDER_MAPPINGS to be used -- this should be fixed if we want to support -# external networks in legacy netvirt -if [[ ${CONTROLLERFEATURES} == *"odl-netvirt-openstack"* ]]; then - ODL_PROVIDER_MAPPINGS="\${PUBLIC_PHYSICAL_NETWORK}:${PUBLIC_BRIDGE}" -else - ODL_PROVIDER_MAPPINGS= -fi + if [ "${ODL_ENABLE_L3_FWD}" == "yes" ]; then + cat >> ${local_conf_file_name} << EOF -if [ "${ODL_ENABLE_L3_FWD}" == "yes" ]; then -cat >> ${local_conf_file_name} << EOF -# Uncomment lines below if odl-compute is to be used for l3 forwarding -Q_L3_ENABLED=True -ODL_L3=${ODL_L3} -PUBLIC_INTERFACE=br100 # FIXME do we use br100 at all? PUBLIC_BRIDGE=${PUBLIC_BRIDGE} PUBLIC_PHYSICAL_NETWORK=${PUBLIC_PHYSICAL_NETWORK} ODL_PROVIDER_MAPPINGS=${ODL_PROVIDER_MAPPINGS} +Q_L3_ENABLED=True +ODL_L3=${ODL_L3} EOF -fi -echo "local.conf Created...." -cat ${local_conf_file_name} -} + fi -function configure_haproxy_for_neutron_requests () { -MGRIP=$1 -ODL_IPS=(${2//,/ }) + cat >> ${local_conf_file_name} << EOF -cat > ${WORKSPACE}/install_ha_proxy.sh<< EOF +[[post-config|/etc/nova/nova.conf]] +[api] +auth_strategy = keystone +[DEFAULT] +use_neutron = True +EOF + + echo "Compute local.conf created:" + cat ${local_conf_file_name} +} # create_compute_node_local_conf() + +function configure_haproxy_for_neutron_requests() { + MGRIP=$1 + ODL_IPS=(${2//,/ }) + + cat > ${WORKSPACE}/install_ha_proxy.sh<< EOF sudo systemctl stop firewalld sudo yum -y install policycoreutils-python haproxy EOF @@ -333,34 +385,32 @@ listen opendaylight balance source EOF -odlindex=1 -for odlip in ${ODL_IPS[*]} -do -cat >> ${WORKSPACE}/haproxy.cfg << EOF + odlindex=1 + for odlip in ${ODL_IPS[*]}; do + cat >> ${WORKSPACE}/haproxy.cfg << EOF server controller-${odlindex} ${odlip}:8080 check fall 5 inter 2000 rise 2 EOF -odlindex=$((odlindex+1)) -done + odlindex=$((odlindex+1)) + done -cat >> ${WORKSPACE}/haproxy.cfg << EOF + cat >> ${WORKSPACE}/haproxy.cfg << EOF listen opendaylight_rest bind ${MGRIP}:8181 balance source EOF -odlindex=1 -for odlip in ${ODL_IPS[*]} -do -cat >> ${WORKSPACE}/haproxy.cfg << EOF + odlindex=1 + for odlip in ${ODL_IPS[*]}; do + cat >> ${WORKSPACE}/haproxy.cfg << EOF server controller-rest-${odlindex} ${odlip}:8181 check fall 5 inter 2000 rise 2 EOF -odlindex=$((odlindex+1)) -done + odlindex=$((odlindex+1)) + done -echo "Dump haproxy.cfg" -cat ${WORKSPACE}/haproxy.cfg + echo "Dump haproxy.cfg" + cat ${WORKSPACE}/haproxy.cfg -cat > ${WORKSPACE}/deploy_ha_proxy.sh<< EOF + cat > ${WORKSPACE}/deploy_ha_proxy.sh<< EOF sudo chown haproxy:haproxy /tmp/haproxy.cfg sudo sed -i 's/\\/etc\\/haproxy\\/haproxy.cfg/\\/tmp\\/haproxy.cfg/g' /usr/lib/systemd/system/haproxy.service sudo /usr/sbin/semanage permissive -a haproxy_t @@ -370,26 +420,35 @@ sudo netstat -tunpl sudo systemctl status haproxy true EOF -scp ${WORKSPACE}/install_ha_proxy.sh ${MGRIP}:/tmp -${SSH} ${MGRIP} "sudo bash /tmp/install_ha_proxy.sh" -scp ${WORKSPACE}/haproxy.cfg ${MGRIP}:/tmp -scp ${WORKSPACE}/deploy_ha_proxy.sh ${MGRIP}:/tmp -${SSH} ${MGRIP} "sudo bash /tmp/deploy_ha_proxy.sh" + + scp ${WORKSPACE}/install_ha_proxy.sh ${MGRIP}:/tmp + ${SSH} ${MGRIP} "sudo bash /tmp/install_ha_proxy.sh" + scp ${WORKSPACE}/haproxy.cfg ${MGRIP}:/tmp + scp ${WORKSPACE}/deploy_ha_proxy.sh ${MGRIP}:/tmp + ${SSH} ${MGRIP} "sudo bash /tmp/deploy_ha_proxy.sh" +} # configure_haproxy_for_neutron_requests() + +# Collect the list of files on the hosts +function list_files () { + local ip=$1 + local folder=$2 + ${SSH} ${ip} "sudo find /etc > /tmp/find.etc.txt" + ${SSH} ${ip} "sudo find /opt/stack > /tmp/find.opt.stack.txt" + ${SSH} ${ip} "sudo find /var > /tmp/find2.txt" + ${SSH} ${ip} "sudo find /var > /tmp/find.var.txt" + rsync --rsync-path="sudo rsync" --list-only -arvhe ssh ${ip}:/etc/ > ${folder}/rsync.etc.txt + rsync --rsync-path="sudo rsync" --list-only -arvhe ssh ${ip}:/opt/stack/ > ${folder}/rsync.opt.stack.txt + rsync --rsync-path="sudo rsync" --list-only -arvhe ssh ${ip}:/var/ > ${folder}/rsync.var.txt + scp ${ip}:/tmp/find.etc.txt ${folder} + scp ${ip}:/tmp/find.opt.stack.txt ${folder} + scp ${ip}:/tmp/find2.txt ${folder} + scp ${ip}:/tmp/find.var.txt ${folder} } -function collect_logs_and_exit () { -set +e # We do not want to create red dot just because something went wrong while fetching logs. -for i in `seq 1 ${NUM_ODL_SYSTEM}` -do - CONTROLLERIP=ODL_SYSTEM_${i}_IP - echo "Lets's take the karaf thread dump again..." - KARAF_PID=$(ssh ${!CONTROLLERIP} "ps aux | grep ${KARAF_ARTIFACT} | grep -v grep | tr -s ' ' | cut -f2 -d' '") - ssh ${!CONTROLLERIP} "jstack $KARAF_PID"> ${WORKSPACE}/karaf_${i}_threads_after.log || true - echo "killing karaf process..." - ${SSH} "${!CONTROLLERIP}" bash -c 'ps axf | grep karaf | grep -v grep | awk '"'"'{print "kill -9 " $1}'"'"' | sh' -done +function collect_logs () { + set +e # We do not want to create red dot just because something went wrong while fetching logs. -cat > extra_debug.sh << EOF + cat > extra_debug.sh << EOF echo -e "/usr/sbin/lsmod | /usr/bin/grep openvswitch\n" /usr/sbin/lsmod | /usr/bin/grep openvswitch echo -e "\ngrep ct_ /var/log/openvswitch/ovs-vswitchd.log\n" @@ -401,126 +460,351 @@ sudo netstat -punta echo -e "\nsudo getenforce\n" sudo getenforce echo -e "\njournalctl > /tmp/journalctl.log\n" -journalctl > /tmp/journalctl.log +sudo journalctl > /tmp/journalctl.log +echo -e "\nsudo systemctl status httpd\n" +sudo systemctl status httpd +echo -e "\nenv\n" +env +source /opt/stack/devstack/openrc admin admin +echo -e "\nenv after openrc\n" +env EOF -sleep 5 -# FIXME: Do not create .tar and gzip before copying. -for i in `seq 1 ${NUM_ODL_SYSTEM}` -do - CONTROLLERIP=ODL_SYSTEM_${i}_IP - ${SSH} "${!CONTROLLERIP}" "cp -r /tmp/${BUNDLEFOLDER}/data/log /tmp/odl_log" - ${SSH} "${!CONTROLLERIP}" "tar -cf /tmp/odl${i}_karaf.log.tar /tmp/odl_log/*" - scp "${!CONTROLLERIP}:/tmp/odl${i}_karaf.log.tar" "${WORKSPACE}/odl${i}_karaf.log.tar" - ${SSH} "${!CONTROLLERIP}" "tar -cf /tmp/odl${i}_zrpcd.log.tar /tmp/zrpcd.init.log" - scp "${!CONTROLLERIP}:/tmp/odl${i}_zrpcd.log.tar" "${WORKSPACE}/odl${i}_zrpcd.log.tar" - tar -xvf ${WORKSPACE}/odl${i}_karaf.log.tar -C . --strip-components 2 --transform s/karaf/odl${i}_karaf/g - grep "ROBOT MESSAGE\| ERROR " odl${i}_karaf.log > odl${i}_err.log - grep "ROBOT MESSAGE\|Exception" odl${i}_karaf.log > odl${i}_exception.log - grep "ROBOT MESSAGE\| ERROR \| WARN \|Exception" odl${i}_karaf.log > odl${i}_err_warn_exception.log - rm ${WORKSPACE}/odl${i}_karaf.log.tar -done - -# Since this log collection work is happening before the archive build macro which also -# creates the ${WORKSPACE}/archives dir, we have to do it here first. The mkdir in the -# archives build step will essentially be a noop. -mkdir -p ${WORKSPACE}/archives - -# Control Node -for i in `seq 1 ${NUM_OPENSTACK_CONTROL_NODES}` -do - OS_CTRL_IP=OPENSTACK_CONTROL_NODE_${i}_IP - OS_CTRL_FOLDER="control_${i}" - mkdir -p ${OS_CTRL_FOLDER} - scp ${!OS_CTRL_IP}:/opt/stack/devstack/nohup.out ${OS_CTRL_FOLDER}/stack.log - scp ${!OS_CTRL_IP}:/var/log/openvswitch/ovs-vswitchd.log ${OS_CTRL_FOLDER}/ovs-vswitchd.log - scp ${!OS_CTRL_IP}:/var/log/openvswitch/ovsdb-server.log ${OS_CTRL_FOLDER}/ovsdb-server.log - scp ${!OS_CTRL_IP}:/etc/neutron/neutron.conf ${OS_CTRL_FOLDER}/neutron.conf - scp ${!OS_CTRL_IP}:/etc/nova/nova.conf ${OS_CTRL_FOLDER}/nova.conf - scp ${!OS_CTRL_IP}:/etc/kuryr/kuryr.conf ${OS_CTRL_FOLDER}/kuryr.conf - scp ${!OS_CTRL_IP}:/etc/neutron/neutron_lbaas.conf ${OS_CTRL_FOLDER}/neutron-lbaas.conf - scp ${!OS_CTRL_IP}:/etc/neutron/services/loadbalancer/haproxy/lbaas_agent.ini ${OS_CTRL_FOLDER}/lbaas-agent.ini - rsync -avhe ssh ${!OS_CTRL_IP}:/opt/stack/logs/* ${OS_CTRL_FOLDER} # rsync to prevent copying of symbolic links - # Use rsync with sudo to get access to the log dir. - rsync --rsync-path="sudo rsync" -avhe ssh ${!OS_CTRL_IP}:/var/log/audit/audit.log ${OS_CTRL_FOLDER} - scp extra_debug.sh ${!OS_CTRL_IP}:/tmp - ${SSH} ${!OS_CTRL_IP} "bash /tmp/extra_debug.sh > /tmp/extra_debug.log" - scp ${!OS_CTRL_IP}:/tmp/extra_debug.log ${OS_CTRL_FOLDER}/extra_debug.log - scp ${!OS_CTRL_IP}:/tmp/journalctl.log ${OS_CTRL_FOLDER} - rsync --rsync-path="sudo rsync" -avhe ssh ${!OS_CTRL_IP}:/var/log/messages ${OS_CTRL_FOLDER} - scp ${!OS_CTRL_IP}:/tmp/*.xz ${OS_CTRL_FOLDER}/ - mv local.conf_control_${!OS_CTRL_IP} ${OS_CTRL_FOLDER}/local.conf - mv ${OS_CTRL_FOLDER} ${WORKSPACE}/archives/ -done + # Since this log collection work is happening before the archive build macro which also + # creates the ${WORKSPACE}/archives dir, we have to do it here first. The mkdir in the + # archives build step will essentially be a noop. + mkdir -p ${WORKSPACE}/archives + + sleep 5 + # FIXME: Do not create .tar and gzip before copying. + for i in `seq 1 ${NUM_ODL_SYSTEM}`; do + CONTROLLERIP=ODL_SYSTEM_${i}_IP + echo "collect_logs: for opendaylight controller ip: ${!CONTROLLERIP}" + NODE_FOLDER="odl_${i}" + mkdir -p ${NODE_FOLDER} + echo "Lets's take the karaf thread dump again..." + ${SSH} ${!CONTROLLERIP} "sudo ps aux > /tmp/ps.log" + KARAF_PID=$(ssh ${!CONTROLLERIP} "ps aux | grep ${KARAF_ARTIFACT} | grep -v grep | tr -s ' ' | cut -f2 -d' '") + ssh ${!CONTROLLERIP} "jstack ${KARAF_PID}"> ${WORKSPACE}/karaf_${i}_threads_after.log || true + echo "killing karaf process..." + ${SSH} "${!CONTROLLERIP}" bash -c 'ps axf | grep karaf | grep -v grep | awk '"'"'{print "kill -9 " $1}'"'"' | sh' + ${SSH} ${!CONTROLLERIP} "sudo journalctl > /tmp/journalctl.log" + scp ${!CONTROLLERIP}:/tmp/journalctl.log ${NODE_FOLDER} + ${SSH} ${!CONTROLLERIP} "dmesg -T > /tmp/dmesg.log" + scp ${!CONTROLLERIP}:/tmp/dmesg.log ${NODE_FOLDER} + ${SSH} ${!CONTROLLERIP} "cp -r /tmp/${BUNDLEFOLDER}/data/log /tmp/odl_log" + ${SSH} ${!CONTROLLERIP} "tar -cf /tmp/odl${i}_karaf.log.tar /tmp/odl_log/*" + scp ${!CONTROLLERIP}:/tmp/odl${i}_karaf.log.tar ${NODE_FOLDER} + ${SSH} ${!CONTROLLERIP} "tar -cf /tmp/odl${i}_zrpcd.log.tar /tmp/zrpcd.init.log" + scp ${!CONTROLLERIP}:/tmp/odl${i}_zrpcd.log.tar ${NODE_FOLDER} + tar -xvf ${NODE_FOLDER}/odl${i}_karaf.log.tar -C ${NODE_FOLDER} --strip-components 2 --transform s/karaf/odl${i}_karaf/g + grep "ROBOT MESSAGE\| ERROR " ${NODE_FOLDER}/odl${i}_karaf.log > ${NODE_FOLDER}/odl${i}_err.log + grep "ROBOT MESSAGE\| ERROR \| WARN \|Exception" \ + ${NODE_FOLDER}/odl${i}_karaf.log > ${NODE_FOLDER}/odl${i}_err_warn_exception.log + # Print ROBOT lines and print Exception lines. For exception lines also print the previous line for context + sed -n -e '/ROBOT MESSAGE/P' -e '$!N;/Exception/P;D' ${NODE_FOLDER}/odl${i}_karaf.log > ${NODE_FOLDER}/odl${i}_exception.log + rm ${NODE_FOLDER}/odl${i}_karaf.log.tar + mv karaf_${i}_threads* ${NODE_FOLDER} + mv ${NODE_FOLDER} ${WORKSPACE}/archives/ + done -# Compute Nodes -for i in `seq 1 ${NUM_OPENSTACK_COMPUTE_NODES}` -do - OSIP=OPENSTACK_COMPUTE_NODE_${i}_IP - OS_COMPUTE_FOLDER="compute_${i}" - mkdir -p ${OS_COMPUTE_FOLDER} - scp ${!OSIP}:/opt/stack/devstack/nohup.out ${OS_COMPUTE_FOLDER}/stack.log - scp ${!OSIP}:/var/log/openvswitch/ovs-vswitchd.log ${OS_COMPUTE_FOLDER}/ovs-vswitchd.log - scp ${!OSIP}:/var/log/openvswitch/ovsdb-server.log ${OS_COMPUTE_FOLDER}/ovsdb-server.log - scp ${!OSIP}:/var/log/libvirt/libvirtd.log* ${OS_COMPUTE_FOLDER} - scp ${!OSIP}:/var/log/libvirt/qeum/*.log ${OS_COMPUTE_FOLDER} - scp ${!OSIP}:/etc/nova/nova.conf ${OS_COMPUTE_FOLDER}/nova.conf - rsync -avhe ssh ${!OSIP}:/opt/stack/logs/* ${OS_COMPUTE_FOLDER} # rsync to prevent copying of symbolic links - # Use rsync with sudo to get access to the log dir. Also can't use wildcard because the dirs only have - # exec permissions which doesn't allow ls. - rsync --rsync-path="sudo rsync" -avhe ssh ${!OSIP}:/var/log/libvirt ${OS_COMPUTE_FOLDER} - rsync --rsync-path="sudo rsync" -avhe ssh ${!OSIP}:/var/log/audit/audit.log ${OS_COMPUTE_FOLDER} - scp extra_debug.sh ${!OSIP}:/tmp - ${SSH} ${!OSIP} "bash /tmp/extra_debug.sh > /tmp/extra_debug.log" - scp ${!OSIP}:/tmp/extra_debug.log ${OS_COMPUTE_FOLDER}/extra_debug.log - scp ${!OSIP}:/tmp/journalctl.log ${OS_COMPUTE_FOLDER} - rsync --rsync-path="sudo rsync" -avhe ssh ${!OSIP}:/var/log/messages ${OS_COMPUTE_FOLDER} - scp ${!OSIP}:/tmp/*.xz ${OS_COMPUTE_FOLDER}/ - mv local.conf_compute_${!OSIP} ${OS_COMPUTE_FOLDER}/local.conf - mv ${OS_COMPUTE_FOLDER} ${WORKSPACE}/archives/ -done + print_job_parameters > ${WORKSPACE}/archives/params.txt + + # Control Node + for i in `seq 1 ${NUM_OPENSTACK_CONTROL_NODES}`; do + OSIP=OPENSTACK_CONTROL_NODE_${i}_IP + echo "collect_logs: for openstack control node ip: ${!OSIP}" + NODE_FOLDER="control_${i}" + mkdir -p ${NODE_FOLDER} + scp ${!OSIP}:/etc/dnsmasq.conf ${NODE_FOLDER} + scp ${!OSIP}:/etc/keystone/keystone.conf ${NODE_FOLDER} + scp ${!OSIP}:/etc/keystone/keystone-uwsgi-admin.ini ${NODE_FOLDER} + scp ${!OSIP}:/etc/keystone/keystone-uwsgi-public.ini ${NODE_FOLDER} + scp ${!OSIP}:/etc/kuryr/kuryr.conf ${NODE_FOLDER} + scp ${!OSIP}:/etc/neutron/dhcp_agent.ini ${NODE_FOLDER} + scp ${!OSIP}:/etc/neutron/metadata_agent.ini ${NODE_FOLDER} + scp ${!OSIP}:/etc/neutron/neutron.conf ${NODE_FOLDER} + scp ${!OSIP}:/etc/neutron/neutron_lbaas.conf ${NODE_FOLDER} + scp ${!OSIP}:/etc/neutron/plugins/ml2/ml2_conf.ini ${NODE_FOLDER} + scp ${!OSIP}:/etc/neutron/services/loadbalancer/haproxy/lbaas_agent.ini ${NODE_FOLDER} + scp ${!OSIP}:/etc/nova/nova.conf ${NODE_FOLDER} + scp ${!OSIP}:/etc/nova/nova-api-uwsgi.ini ${NODE_FOLDER} + scp ${!OSIP}:/etc/nova/nova_cell1.conf ${NODE_FOLDER} + scp ${!OSIP}:/etc/nova/nova-cpu.conf ${NODE_FOLDER} + scp ${!OSIP}:/etc/nova/placement-uwsgi.ini ${NODE_FOLDER} + scp ${!OSIP}:/etc/openstack/clouds.yaml ${NODE_FOLDER} + scp ${!OSIP}:/opt/stack/devstack/.stackenv ${NODE_FOLDER} + scp ${!OSIP}:/opt/stack/devstack/nohup.out ${NODE_FOLDER}/stack.log + scp ${!OSIP}:/opt/stack/devstack/openrc ${NODE_FOLDER} + scp ${!OSIP}:/opt/stack/requirements/upper-constraints.txt ${NODE_FOLDER} + scp ${!OSIP}:/opt/stack/tempest/etc/tempest.conf ${NODE_FOLDER} + scp ${!OSIP}:/tmp/get_devstack.sh.txt ${NODE_FOLDER} + scp ${!OSIP}:/var/log/openvswitch/ovs-vswitchd.log ${NODE_FOLDER} + scp ${!OSIP}:/var/log/openvswitch/ovsdb-server.log ${NODE_FOLDER} + list_files "${!OSIP}" "${NODE_FOLDER}" + rsync --rsync-path="sudo rsync" -avhe ssh ${!OSIP}:/etc/hosts ${NODE_FOLDER} + rsync --rsync-path="sudo rsync" -avhe ssh ${!OSIP}:/usr/lib/systemd/system/haproxy.service ${NODE_FOLDER} + rsync --rsync-path="sudo rsync" -avhe ssh ${!OSIP}:/var/log/audit/audit.log ${NODE_FOLDER} + rsync --rsync-path="sudo rsync" -avhe ssh ${!OSIP}:/var/log/httpd/keystone_access.log ${NODE_FOLDER} + rsync --rsync-path="sudo rsync" -avhe ssh ${!OSIP}:/var/log/httpd/keystone.log ${NODE_FOLDER} + rsync --rsync-path="sudo rsync" -avhe ssh ${!OSIP}:/var/log/messages ${NODE_FOLDER} + rsync --rsync-path="sudo rsync" -avhe ssh ${!OSIP}:/var/log/rabbitmq ${NODE_FOLDER} + rsync -avhe ssh ${!OSIP}:/opt/stack/logs/* ${NODE_FOLDER} # rsync to prevent copying of symbolic links + scp extra_debug.sh ${!OSIP}:/tmp + ${SSH} ${!OSIP} "bash /tmp/extra_debug.sh > /tmp/extra_debug.log" + scp ${!OSIP}:/tmp/extra_debug.log ${NODE_FOLDER} + scp ${!OSIP}:/tmp/journalctl.log ${NODE_FOLDER} + scp ${!OSIP}:/tmp/*.xz ${NODE_FOLDER} + ${SSH} ${!CONTROLLERIP} "dmesg -T > /tmp/dmesg.log" + scp ${!CONTROLLERIP}:/tmp/dmesg.log ${NODE_FOLDER} + mv local.conf_control_${!OSIP} ${NODE_FOLDER}/local.conf + mv /tmp/qdhcp ${NODE_FOLDER} + mv ${NODE_FOLDER} ${WORKSPACE}/archives/ + done -find local.conf* -print0 | xargs -0 -I % mv % %.log + # Compute Nodes + for i in `seq 1 ${NUM_OPENSTACK_COMPUTE_NODES}`; do + OSIP=OPENSTACK_COMPUTE_NODE_${i}_IP + echo "collect_logs: for openstack compute node ip: ${!OSIP}" + NODE_FOLDER="compute_${i}" + mkdir -p ${NODE_FOLDER} + scp ${!OSIP}:/etc/nova/nova.conf ${NODE_FOLDER} + scp ${!OSIP}:/etc/nova/nova-cpu.conf ${NODE_FOLDER} + scp ${!OSIP}:/etc/openstack/clouds.yaml ${NODE_FOLDER} + scp ${!OSIP}:/opt/stack/devstack/.stackenv ${NODE_FOLDER} + scp ${!OSIP}:/opt/stack/devstack/nohup.out ${NODE_FOLDER}/stack.log + scp ${!OSIP}:/opt/stack/devstack/openrc ${NODE_FOLDER} + scp ${!OSIP}:/opt/stack/requirements/upper-constraints.txt ${NODE_FOLDER} + scp ${!OSIP}:/tmp/get_devstack.sh.txt ${NODE_FOLDER} + scp ${!OSIP}:/var/log/openvswitch/ovs-vswitchd.log ${NODE_FOLDER} + scp ${!OSIP}:/var/log/openvswitch/ovsdb-server.log ${NODE_FOLDER} + list_files "${!OSIP}" "${NODE_FOLDER}" + rsync --rsync-path="sudo rsync" -avhe ssh ${!OSIP}:/etc/hosts ${NODE_FOLDER} + rsync --rsync-path="sudo rsync" -avhe ssh ${!OSIP}:/var/log/audit/audit.log ${NODE_FOLDER} + rsync --rsync-path="sudo rsync" -avhe ssh ${!OSIP}:/var/log/libvirt ${NODE_FOLDER} + rsync --rsync-path="sudo rsync" -avhe ssh ${!OSIP}:/var/log/messages ${NODE_FOLDER} + rsync --rsync-path="sudo rsync" -avhe ssh ${!OSIP}:/var/log/nova-agent.log ${NODE_FOLDER} + rsync -avhe ssh ${!OSIP}:/opt/stack/logs/* ${NODE_FOLDER} # rsync to prevent copying of symbolic links + scp extra_debug.sh ${!OSIP}:/tmp + ${SSH} ${!OSIP} "bash /tmp/extra_debug.sh > /tmp/extra_debug.log" + scp ${!OSIP}:/tmp/extra_debug.log ${NODE_FOLDER} + scp ${!OSIP}:/tmp/journalctl.log ${NODE_FOLDER} + scp ${!OSIP}:/tmp/*.xz ${NODE_FOLDER}/ + ${SSH} ${!OSIP} "dmesg -T > /tmp/dmesg.log" + scp ${!OSIP}:/tmp/dmesg.log ${NODE_FOLDER} + mv local.conf_compute_${!OSIP} ${NODE_FOLDER}/local.conf + mv ${NODE_FOLDER} ${WORKSPACE}/archives/ + done -# Tempest -DEVSTACK_TEMPEST_DIR="/opt/stack/tempest" -if ssh ${OPENSTACK_CONTROL_NODE_1_IP} "sudo sh -c '[ -f ${DEVSTACK_TEMPEST_DIR}/.testrepository/0 ]'"; then # if Tempest results exist - ssh ${OPENSTACK_CONTROL_NODE_1_IP} "for I in \$(sudo ls ${DEVSTACK_TEMPEST_DIR}/.testrepository/ | grep -E '^[0-9]+$'); do sudo sh -c \"${DEVSTACK_TEMPEST_DIR}/.tox/tempest/bin/subunit-1to2 < ${DEVSTACK_TEMPEST_DIR}/.testrepository/\${I} >> ${DEVSTACK_TEMPEST_DIR}/subunit_log.txt\"; done" - ssh ${OPENSTACK_CONTROL_NODE_1_IP} "sudo sh -c '${DEVSTACK_TEMPEST_DIR}/.tox/tempest/bin/python ${DEVSTACK_TEMPEST_DIR}/.tox/tempest/lib/python2.7/site-packages/os_testr/subunit2html.py ${DEVSTACK_TEMPEST_DIR}/subunit_log.txt ${DEVSTACK_TEMPEST_DIR}/tempest_results.html'" + # Tempest + DEVSTACK_TEMPEST_DIR="/opt/stack/tempest" + TESTREPO=".stestr" TEMPEST_LOGS_DIR=${WORKSPACE}/archives/tempest - mkdir -p ${TEMPEST_LOGS_DIR} - scp ${OPENSTACK_CONTROL_NODE_1_IP}:${DEVSTACK_TEMPEST_DIR}/tempest_results.html ${TEMPEST_LOGS_DIR} - scp ${OPENSTACK_CONTROL_NODE_1_IP}:${DEVSTACK_TEMPEST_DIR}/tempest.log ${TEMPEST_LOGS_DIR} - mv ${WORKSPACE}/tempest_output* ${TEMPEST_LOGS_DIR} -fi + # Look for tempest test results in the $TESTREPO dir and copy if found + if ${SSH} ${OPENSTACK_CONTROL_NODE_1_IP} "sudo sh -c '[ -f ${DEVSTACK_TEMPEST_DIR}/${TESTREPO}/0 ]'"; then + ${SSH} ${OPENSTACK_CONTROL_NODE_1_IP} "for I in \$(sudo ls ${DEVSTACK_TEMPEST_DIR}/${TESTREPO}/ | grep -E '^[0-9]+$'); do sudo sh -c \"${DEVSTACK_TEMPEST_DIR}/.tox/tempest/bin/subunit-1to2 < ${DEVSTACK_TEMPEST_DIR}/${TESTREPO}/\${I} >> ${DEVSTACK_TEMPEST_DIR}/subunit_log.txt\"; done" + ${SSH} ${OPENSTACK_CONTROL_NODE_1_IP} "sudo sh -c '${DEVSTACK_TEMPEST_DIR}/.tox/tempest/bin/python ${DEVSTACK_TEMPEST_DIR}/.tox/tempest/lib/python2.7/site-packages/os_testr/subunit2html.py ${DEVSTACK_TEMPEST_DIR}/subunit_log.txt ${DEVSTACK_TEMPEST_DIR}/tempest_results.html'" + mkdir -p ${TEMPEST_LOGS_DIR} + scp ${OPENSTACK_CONTROL_NODE_1_IP}:${DEVSTACK_TEMPEST_DIR}/tempest_results.html ${TEMPEST_LOGS_DIR} + scp ${OPENSTACK_CONTROL_NODE_1_IP}:${DEVSTACK_TEMPEST_DIR}/tempest.log ${TEMPEST_LOGS_DIR} + mv ${WORKSPACE}/tempest_output* ${TEMPEST_LOGS_DIR} + else + echo "tempest results not found in ${DEVSTACK_TEMPEST_DIR}/${TESTREPO}/0" + fi +} # collect_logs() + +# Following three functions are debugging helpers when debugging devstack changes. +# Keeping them for now so we can simply call them when needed. +ctrlhn="" +comp1hn="" +comp2hn="" +function get_hostnames () { + set +e + local ctrlip=${OPENSTACK_CONTROL_NODE_1_IP} + local comp1ip=${OPENSTACK_COMPUTE_NODE_1_IP} + local comp2ip=${OPENSTACK_COMPUTE_NODE_2_IP} + ctrlhn=$(${SSH} ${ctrlip} "hostname") + comp1hn=$(${SSH} ${comp1ip} "hostname") + comp2hn=$(${SSH} ${comp2ip} "hostname") + echo "hostnames: ${ctrlhn}, ${comp1hn}, ${comp2hn}" + set -e +} + +function check_firewall() { + set +e + echo $- + local ctrlip=${OPENSTACK_CONTROL_NODE_1_IP} + local comp1ip=${OPENSTACK_COMPUTE_NODE_1_IP} + local comp2ip=${OPENSTACK_COMPUTE_NODE_2_IP} + + echo "check_firewall on control" + ${SSH} ${ctrlip} " + sudo systemctl status firewalld + sudo systemctl -l status iptables + sudo iptables --line-numbers -nvL + " || true + echo "check_firewall on compute 1" + ${SSH} ${comp1ip} " + sudo systemctl status firewalld + sudo systemctl -l status iptables + sudo iptables --line-numbers -nvL + " || true + echo "check_firewall on compute 2" + ${SSH} ${comp2ip} " + sudo systemctl status firewalld + sudo systemctl -l status iptables + sudo iptables --line-numbers -nvL + " || true +} + +function get_service () { + set +e + local iter=$1 + #local idx=$2 + local ctrlip=${OPENSTACK_CONTROL_NODE_1_IP} + local comp1ip=${OPENSTACK_COMPUTE_NODE_1_IP} + + #if [ ${idx} -eq 1 ]; then + if [ ${iter} -eq 1 ] || [ ${iter} -gt 16 ]; then + curl http://${ctrlip}:5000 + curl http://${ctrlip}:35357 + curl http://${ctrlip}/identity + ${SSH} ${ctrlip} " + source /opt/stack/devstack/openrc admin admin; + env + openstack configuration show --unmask; + openstack service list + openstack --os-cloud devstack-admin --os-region RegionOne compute service list + openstack hypervisor list; + " || true + check_firewall + fi + #fi + set -e +} + +# Check if rabbitmq is ready by looking for a pid in it's status. +# The function returns the status of the grep command which callers can check. +function is_rabbitmq_ready() { + local -r ip=$1 + rm -f rabbit.txt + ${SSH} ${ip} "sudo rabbitmqctl status" > rabbit.txt + grep pid rabbit.txt } +# retry the given command ($3) until success for a number of iterations ($1) +# sleeping ($2) between tries. +function retry() { + set +e + local -r -i max_tries=${1} + local -r -i sleep_time=${2} + local -r cmd=${3} + local -i retries=1 + local -i rc=1 + while true; do + echo "retry ${cmd}: attempt: ${retries}" + ${cmd} + rc=$? + if ((${rc} == 0)); then + break; + else + if ((${retries} == ${max_tries})); then + break + else + ((retries++)) + sleep ${sleep_time} + fi + fi + done + set -e + return ${rc} +} + +# if we are using the new netvirt impl, as determined by the feature name +# odl-netvirt-openstack (note: old impl is odl-ovsdb-openstack) then we +# want PROVIDER_MAPPINGS to be used -- this should be fixed if we want to support +# external networks in legacy netvirt +if [[ ${CONTROLLERFEATURES} == *"odl-netvirt-openstack"* ]]; then + ODL_PROVIDER_MAPPINGS="\${PUBLIC_PHYSICAL_NETWORK}:${PUBLIC_BRIDGE}" +else + ODL_PROVIDER_MAPPINGS= +fi + +# if we are using the old netvirt impl, as determined by the feature name +# odl-ovsdb-openstack (note: new impl is odl-netvirt-openstack) then we +# want ODL_L3 to be True. New impl wants it False +if [[ ${CONTROLLERFEATURES} == *"odl-ovsdb-openstack"* ]]; then + ODL_L3=True +else + ODL_L3=False +fi + +RECLONE=False + +# Always compare the lists below against the devstack upstream ENABLED_SERVICES in +# https://github.com/openstack-dev/devstack/blob/master/stackrc#L52 +# ODL CSIT does not use vnc, cinder, q-agt, q-l3 or horizon so they are not included below. +# collect performance stats +CORE_OS_CONTROL_SERVICES="dstat" +# Glance +CORE_OS_CONTROL_SERVICES+=",g-api,g-reg" +# Keystone +CORE_OS_CONTROL_SERVICES+=",key" +# Nova - services to support libvirt +CORE_OS_CONTROL_SERVICES+=",n-api,n-api-meta,n-cauth,n-cond,n-crt,n-obj,n-sch" +# ODL - services to connect to ODL +CORE_OS_CONTROL_SERVICES+=",odl-compute,odl-neutron" +# Neutron +CORE_OS_CONTROL_SERVICES+=",q-dhcp,q-meta,q-svc" +# Additional services +CORE_OS_CONTROL_SERVICES+=",mysql,rabbit" + +# computes only need nova and odl +CORE_OS_COMPUTE_SERVICES="n-cpu,odl-compute" + cat > ${WORKSPACE}/disable_firewall.sh << EOF sudo systemctl stop firewalld -sudo systemctl stop iptables +# Open these ports to match the tutorial vms +# http/https (80/443), samba (445), netbios (137,138,139) +sudo iptables -I INPUT -p tcp -m multiport --dports 80,443,139,445 -j ACCEPT +sudo iptables -I INPUT -p udp -m multiport --dports 137,138 -j ACCEPT +# OpenStack services as well as vxlan tunnel ports 4789 and 9876 +# identity public/admin (5000/35357), ampq (5672), vnc (6080), nova (8774), glance (9292), neutron (9696) +sudo sudo iptables -I INPUT -p tcp -m multiport --dports 5000,5672,6080,8774,9292,9696,35357 -j ACCEPT +sudo sudo iptables -I INPUT -p udp -m multiport --dports 4789,9876 -j ACCEPT +sudo iptables-save > /etc/sysconfig/iptables +sudo systemctl restart iptables +sudo iptables --line-numbers -nvL true EOF cat > ${WORKSPACE}/get_devstack.sh << EOF sudo systemctl stop firewalld -sudo yum install bridge-utils -y -sudo systemctl stop NetworkManager +sudo yum install bridge-utils python-pip -y +#sudo systemctl stop NetworkManager #Disable NetworkManager and kill dhclient and dnsmasq sudo systemctl stop NetworkManager sudo killall dhclient sudo killall dnsmasq #Workaround for mysql failure -echo "127.0.0.1 localhost \${HOSTNAME}" > /tmp/hosts -echo "::1 localhost \${HOSTNAME}" >> /tmp/hosts +echo "127.0.0.1 localhost \${HOSTNAME}" >> /tmp/hosts +echo "::1 localhost \${HOSTNAME}" >> /tmp/hosts sudo mv /tmp/hosts /etc/hosts -sudo /usr/sbin/brctl addbr br100 -#sudo ifconfig eth0 mtu 2000 sudo mkdir /opt/stack sudo chmod 777 /opt/stack cd /opt/stack -git clone https://git.openstack.org/openstack-dev/devstack +echo "git clone https://git.openstack.org/openstack-dev/devstack --branch ${OPENSTACK_BRANCH}" +git clone https://git.openstack.org/openstack-dev/devstack --branch ${OPENSTACK_BRANCH} cd devstack -git checkout $OPENSTACK_BRANCH +if [ -n "${DEVSTACK_HASH}" ]; then + echo "git checkout ${DEVSTACK_HASH}" + git checkout ${DEVSTACK_HASH} +fi +git --no-pager log --pretty=format:'%h %<(13)%ar%<(13)%cr %<(20,trunc)%an%d %s\n%b' -n20 +echo "workaround: adjust wait from 60s to 1800s (30m)" +sed -i 's/wait_for_compute 60/wait_for_compute 1800/g' /opt/stack/devstack/lib/nova +# TODO: modify sleep 1 to sleep 60, search wait_for_compute, then first sleep 1 +# that would just reduce the number of logs in the compute stack.log EOF cat > "${WORKSPACE}/setup_host_cell_mapping.sh" << EOF @@ -530,23 +814,14 @@ sudo nova-manage db sync sudo nova-manage cell_v2 discover_hosts EOF -[ "$NUM_OPENSTACK_SITES" ] || NUM_OPENSTACK_SITES=1 +NUM_OPENSTACK_SITES=${NUM_OPENSTACK_SITES:-1} compute_index=1 odl_index=1 os_node_list=() os_interval=$(( ${NUM_OPENSTACK_SYSTEM} / ${NUM_OPENSTACK_SITES} )) ha_proxy_index=${os_interval} -cat > "${WORKSPACE}/manual_install_package.sh" << EOF -cd /opt/stack -git clone "\$1" -cd "\$2" -git checkout "\$3" -sudo python setup.py install -EOF - -for i in `seq 1 ${NUM_OPENSTACK_SITES}` -do +for i in `seq 1 ${NUM_OPENSTACK_SITES}`; do if [ "${ENABLE_HAPROXY_FOR_NEUTRON}" == "yes" ]; then echo "Configure HAProxy" ODL_HAPROXYIP_PARAM=OPENSTACK_HAPROXY_${i}_IP @@ -557,7 +832,6 @@ do ODL_IP_PARAM3=ODL_SYSTEM_$((odl_index++))_IP ODLMGRIP[$i]=${!ODL_HAPROXYIP_PARAM} # ODL Northbound uses HAProxy VIP ODL_OVS_MGRS[$i]="${!ODL_IP_PARAM1},${!ODL_IP_PARAM2},${!ODL_IP_PARAM3}" # OVSDB connects to all ODL IPs - configure_haproxy_for_neutron_requests ${!ODL_HAPROXYIP_PARAM} "${ODL_OVS_MGRS[$i]}" else ODL_IP_PARAM=ODL_SYSTEM_${i}_IP @@ -566,106 +840,169 @@ do fi done -for i in `seq 1 ${NUM_OPENSTACK_CONTROL_NODES}` -do - echo "Stack the Control Node" +# Begin stacking the nodes, starting with the controller(s) and then the compute(s) + +for i in `seq 1 ${NUM_OPENSTACK_CONTROL_NODES}`; do CONTROLIP=OPENSTACK_CONTROL_NODE_${i}_IP + echo "Configure the stack of the control node ${i} of ${NUM_OPENSTACK_CONTROL_NODES}: ${CONTROLIP}" + scp ${WORKSPACE}/disable_firewall.sh ${!CONTROLIP}:/tmp + ${SSH} ${!CONTROLIP} "sudo bash /tmp/disable_firewall.sh" + create_etc_hosts ${!CONTROLIP} + scp ${WORKSPACE}/hosts_file ${!CONTROLIP}:/tmp/hosts scp ${WORKSPACE}/get_devstack.sh ${!CONTROLIP}:/tmp - ${SSH} ${!CONTROLIP} "bash /tmp/get_devstack.sh" + ${SSH} ${!CONTROLIP} "bash /tmp/get_devstack.sh > /tmp/get_devstack.sh.txt 2>&1" create_control_node_local_conf ${!CONTROLIP} ${ODLMGRIP[$i]} "${ODL_OVS_MGRS[$i]}" scp ${WORKSPACE}/local.conf_control_${!CONTROLIP} ${!CONTROLIP}:/opt/stack/devstack/local.conf + echo "Stack the control node ${i} of ${NUM_OPENSTACK_CONTROL_NODES}: ${CONTROLIP}" ssh ${!CONTROLIP} "cd /opt/stack/devstack; nohup ./stack.sh > /opt/stack/devstack/nohup.out 2>&1 &" ssh ${!CONTROLIP} "ps -ef | grep stack.sh" ssh ${!CONTROLIP} "ls -lrt /opt/stack/devstack/nohup.out" os_node_list+=(${!CONTROLIP}) - - #Workaround for stable/newton jobs + # Workaround for stable/newton jobs + # TODO: can this be removed now? if [ "${ODL_ML2_BRANCH}" == "stable/newton" ]; then ssh ${!CONTROLIP} "cd /opt/stack; git clone https://git.openstack.org/openstack/requirements; cd requirements; git checkout stable/newton; sed -i /appdirs/d upper-constraints.txt" fi done -for i in `seq 1 ${NUM_OPENSTACK_COMPUTE_NODES}` -do - echo "Stack the Compute Node" +# This is a backup to the CELLSV2_SETUP=singleconductor workaround. Keeping it here as an easy lookup +# if needed. +# Let the control node get started to avoid a race condition where the computes start and try to access +# the nova_cell1 on the control node before it is created. If that happens, the nova-compute service on the +# compute exits and does not attempt to restart. +# 180s is chosen because in test runs the control node usually finished in 17-20 minutes and the computes finished +# in 17 minutes, so take the max difference of 3 minutes and the jobs should still finish around the same time. +# one of the following errors is seen in the compute n-cpu.log: +# Unhandled error: NotAllowed: Connection.open: (530) NOT_ALLOWED - access to vhost 'nova_cell1' refused for user 'stackrabbit' +# AccessRefused: (0, 0): (403) ACCESS_REFUSED - Login was refused using authentication mechanism AMQPLAIN. For details see the broker logfile. +# Compare that timestamp to this log in the control stack.log: sudo rabbitmqctl set_permissions -p nova_cell1 stackrabbit +# If the n-cpu.log is earlier than the control stack.log timestamp then the failure condition is likely hit. +WAIT_FOR_RABBITMQ_MINUTES=60 +echo "Wait a maximum of ${WAIT_FOR_RABBITMQ_MINUTES}m until rabbitmq is ready to allow the controller to create nova_cell1 before the computes need it" +retry ${WAIT_FOR_RABBITMQ_MINUTES} 60 "is_rabbitmq_ready ${OPENSTACK_CONTROL_NODE_1_IP}" +rc=$? +if ((${rc} == 0)); then + echo "rabbitmq is ready, starting ${NUM_OPENSTACK_COMPUTE_NODES} compute(s)" +else + echo "rabbitmq was not ready in ${WAIT_FOR_RABBITMQ_MINUTES}m" + collect_logs + exit 1 +fi + +for i in `seq 1 ${NUM_OPENSTACK_COMPUTE_NODES}`; do NUM_COMPUTES_PER_SITE=$((NUM_OPENSTACK_COMPUTE_NODES / NUM_OPENSTACK_SITES)) SITE_INDEX=$((((i - 1) / NUM_COMPUTES_PER_SITE) + 1)) # We need the site index to infer the control node IP for this compute COMPUTEIP=OPENSTACK_COMPUTE_NODE_${i}_IP CONTROLIP=OPENSTACK_CONTROL_NODE_${SITE_INDEX}_IP + echo "Configure the stack of the compute node ${i} of ${NUM_OPENSTACK_COMPUTE_NODES}: ${COMPUTEIP}" + scp ${WORKSPACE}/disable_firewall.sh "${!COMPUTEIP}:/tmp" + ${SSH} "${!COMPUTEIP}" "sudo bash /tmp/disable_firewall.sh" + create_etc_hosts ${!COMPUTEIP} ${!CONTROLIP} + scp ${WORKSPACE}/hosts_file ${!COMPUTEIP}:/tmp/hosts scp ${WORKSPACE}/get_devstack.sh ${!COMPUTEIP}:/tmp - ${SSH} ${!COMPUTEIP} "bash /tmp/get_devstack.sh" + ${SSH} ${!COMPUTEIP} "bash /tmp/get_devstack.sh > /tmp/get_devstack.sh.txt 2>&1" + if [ "${ODL_ML2_BRANCH}" == "stable/ocata" ]; then + echo "Updating requirements for ${ODL_ML2_BRANCH}" + echo "Workaround for https://review.openstack.org/#/c/491032/" + echo "Modify upper-constraints to use libvirt-python 3.2.0" + ${SSH} ${!COMPUTEIP} " + cd /opt/stack; + git clone https://git.openstack.org/openstack/requirements; + cd requirements; + git checkout stable/ocata; + sed -i s/libvirt-python===2.5.0/libvirt-python===3.2.0/ upper-constraints.txt + " + fi create_compute_node_local_conf ${!COMPUTEIP} ${!CONTROLIP} ${ODLMGRIP[$SITE_INDEX]} "${ODL_OVS_MGRS[$SITE_INDEX]}" scp ${WORKSPACE}/local.conf_compute_${!COMPUTEIP} ${!COMPUTEIP}:/opt/stack/devstack/local.conf + echo "Stack the compute node ${i} of ${NUM_OPENSTACK_COMPUTE_NODES}: ${COMPUTEIP}" ssh ${!COMPUTEIP} "cd /opt/stack/devstack; nohup ./stack.sh > /opt/stack/devstack/nohup.out 2>&1 &" ssh ${!COMPUTEIP} "ps -ef | grep stack.sh" os_node_list+=(${!COMPUTEIP}) done -echo ${os_node_list[*]} +echo "nodelist: ${os_node_list[*]}" +# This script runs on the openstack nodes. It greps for a string that devstack writes when stacking is complete. +# The script then writes a status depending on the grep output that is later scraped by the robot vm to control +# the status polling. cat > ${WORKSPACE}/check_stacking.sh << EOF > /tmp/stack_progress ps -ef | grep "stack.sh" | grep -v grep ret=\$? if [ \${ret} -eq 1 ]; then - grep "This is your host IP address:" /opt/stack/devstack/nohup.out - if [ \$? -eq 0 ]; then - echo "Stacking Complete" > /tmp/stack_progress - else - echo "Stacking Failed" > /tmp/stack_progress - fi + grep "This is your host IP address:" /opt/stack/devstack/nohup.out + if [ \$? -eq 0 ]; then + echo "Stacking Complete" > /tmp/stack_progress + else + echo "Stacking Failed" > /tmp/stack_progress + fi elif [ \${ret} -eq 0 ]; then - echo "Still Stacking" > /tmp/stack_progress + echo "Still Stacking" > /tmp/stack_progress fi EOF -#the checking is repeated for an hour +# devstack debugging +# get_hostnames + +# Check if the stacking is finished. Poll all nodes every 60s for one hour. iteration=0 in_progress=1 while [ ${in_progress} -eq 1 ]; do -iteration=$(($iteration + 1)) -for index in "${!os_node_list[@]}" -do -echo "Check the status of stacking in ${os_node_list[index]}" -scp ${WORKSPACE}/check_stacking.sh ${os_node_list[index]}:/tmp -${SSH} ${os_node_list[index]} "bash /tmp/check_stacking.sh" -scp ${os_node_list[index]}:/tmp/stack_progress . -#debug -cat stack_progress -stacking_status=`cat stack_progress` -if [ "$stacking_status" == "Still Stacking" ]; then - continue -elif [ "$stacking_status" == "Stacking Failed" ]; then - collect_logs_and_exit - exit 1 -elif [ "$stacking_status" == "Stacking Complete" ]; then - unset 'os_node_list[index]' - if [ ${#os_node_list[@]} -eq 0 ]; then - in_progress=0 - fi -fi -done - echo "sleep for a minute before the next check" - sleep 60 - if [ ${iteration} -eq 60 ]; then - collect_logs_and_exit - exit 1 - fi + iteration=$(($iteration + 1)) + for index in "${!os_node_list[@]}"; do + echo "node $index ${os_node_list[index]}: checking stacking status attempt ${iteration} of 60" + scp ${WORKSPACE}/check_stacking.sh ${os_node_list[index]}:/tmp + ${SSH} ${os_node_list[index]} "bash /tmp/check_stacking.sh" + scp ${os_node_list[index]}:/tmp/stack_progress . + cat stack_progress + stacking_status=`cat stack_progress` + # devstack debugging + # get_service "${iteration}" "${index}" + if [ "$stacking_status" == "Still Stacking" ]; then + continue + elif [ "$stacking_status" == "Stacking Failed" ]; then + echo "node $index ${os_node_list[index]}: stacking has failed" + collect_logs + exit 1 + elif [ "$stacking_status" == "Stacking Complete" ]; then + echo "node $index ${os_node_list[index]}: stacking complete" + unset 'os_node_list[index]' + if [ ${#os_node_list[@]} -eq 0 ]; then + in_progress=0 + fi + fi + done + echo "sleep for a minute before the next check" + sleep 60 + if [ ${iteration} -eq 60 ]; then + echo "stacking has failed - took longer than 60m" + collect_logs + exit 1 + fi done +# Further configuration now that stacking is complete. NUM_COMPUTES_PER_SITE=$((NUM_OPENSTACK_COMPUTE_NODES / NUM_OPENSTACK_SITES)) -for i in `seq 1 ${NUM_OPENSTACK_SITES}` -do +for i in `seq 1 ${NUM_OPENSTACK_SITES}`; do echo "Configure the Control Node" CONTROLIP=OPENSTACK_CONTROL_NODE_${i}_IP - - #Need to disable firewalld and iptables in control node - echo "Stop Firewall in Control Node for compute nodes to be able to reach the ports and add to hypervisor-list" - scp ${WORKSPACE}/disable_firewall.sh ${!CONTROLIP}:/tmp - ${SSH} ${!CONTROLIP} "sudo bash /tmp/disable_firewall.sh" + # Gather Compute IPs for the site + for j in `seq 1 ${NUM_COMPUTES_PER_SITE}`; do + COMPUTE_INDEX=$(((i-1) * NUM_COMPUTES_PER_SITE + j)) + IP_VAR=OPENSTACK_COMPUTE_NODE_${COMPUTE_INDEX}_IP + COMPUTE_IPS[$((j-1))]=${!IP_VAR} + done echo "sleep for 60s and print hypervisor-list" sleep 60 + # In Ocata if we do not enable the n-cpu in control node then + # we need to discover hosts manually and ensure that they are mapped to cells. + # reference: https://ask.openstack.org/en/question/102256/how-to-configure-placement-service-for-compute-node-on-ocata/ + if [ "${OPENSTACK_BRANCH}" == "stable/ocata" ]; then + scp ${WORKSPACE}/setup_host_cell_mapping.sh ${!CONTROLIP}:/tmp + ${SSH} ${!CONTROLIP} "sudo bash /tmp/setup_host_cell_mapping.sh" + fi ${SSH} ${!CONTROLIP} "cd /opt/stack/devstack; source openrc admin admin; nova hypervisor-list" # in the case that we are doing openstack (control + compute) all in one node, then the number of hypervisors # will be the same as the number of openstack systems. However, if we are doing multinode openstack then the @@ -679,18 +1016,10 @@ do num_hypervisors=$(${SSH} ${!CONTROLIP} "cd /opt/stack/devstack; source openrc admin admin; openstack hypervisor list -f value | wc -l" | tail -1 | tr -d "\r") if ! [ "${num_hypervisors}" ] || ! [ ${num_hypervisors} -eq ${expected_num_hypervisors} ]; then echo "Error: Only $num_hypervisors hypervisors detected, expected $expected_num_hypervisors" - collect_logs_and_exit + collect_logs exit 1 fi - # For Ocata, if we do not enable the n-cpu in control node - # We need to discover hosts manually and ensure that they are mapped to cells. - # reference: https://ask.openstack.org/en/question/102256/how-to-configure-placement-service-for-compute-node-on-ocata/ - if [ "${OPENSTACK_BRANCH}" == "stable/ocata" ]; then - scp ${WORKSPACE}/setup_host_cell_mapping.sh ${!CONTROLIP}:/tmp - ${SSH} ${!CONTROLIP} "sudo bash /tmp/setup_host_cell_mapping.sh" - fi - # upgrading pip, urllib3 and httplib2 so that tempest tests can be run on openstack control node # this needs to happen after devstack runs because it seems devstack is pulling in specific versions # of these libs that are not working for tempest. @@ -699,26 +1028,17 @@ do ${SSH} ${!CONTROLIP} "sudo pip install httplib2 --upgrade" # Gather Compute IPs for the site - for j in `seq 1 ${NUM_COMPUTES_PER_SITE}` - do + for j in `seq 1 ${NUM_COMPUTES_PER_SITE}`; do COMPUTE_INDEX=$(((i-1) * NUM_COMPUTES_PER_SITE + j)) IP_VAR=OPENSTACK_COMPUTE_NODE_${COMPUTE_INDEX}_IP COMPUTE_IPS[$((j-1))]=${!IP_VAR} done - # Need to disable firewalld and iptables in compute nodes as well - for ip in ${COMPUTE_IPS[*]} - do - scp ${WORKSPACE}/disable_firewall.sh "${ip}:/tmp" - ${SSH} "${ip}" "sudo bash /tmp/disable_firewall.sh" - done - # External Network echo "prepare external networks by adding vxlan tunnels between all nodes on a separate bridge..." # FIXME Should there be a unique gateway IP and devstack index for each site? devstack_index=1 - for ip in ${!CONTROLIP} ${COMPUTE_IPS[*]} - do + for ip in ${!CONTROLIP} ${COMPUTE_IPS[*]}; do # FIXME - Workaround, ODL (new netvirt) currently adds PUBLIC_BRIDGE as a port in br-int since it doesn't see such a bridge existing when we stack ${SSH} $ip "sudo ovs-vsctl --if-exists del-port br-int $PUBLIC_BRIDGE" ${SSH} $ip "sudo ovs-vsctl --may-exist add-br $PUBLIC_BRIDGE -- set bridge $PUBLIC_BRIDGE other-config:disable-in-band=true other_config:hwaddr=f6:00:00:ff:01:0$((devstack_index++))" @@ -727,10 +1047,8 @@ do # ipsec support if [ "${IPSEC_VXLAN_TUNNELS_ENABLED}" == "yes" ]; then ALL_NODES=(${!CONTROLIP} ${COMPUTE_IPS[*]}) - for ((inx_ip1=0; inx_ip1<$((${#ALL_NODES[@]} - 1)); inx_ip1++)) - do - for ((inx_ip2=$((inx_ip1 + 1)); inx_ip2<${#ALL_NODES[@]}; inx_ip2++)) - do + for ((inx_ip1=0; inx_ip1<$((${#ALL_NODES[@]} - 1)); inx_ip1++)); do + for ((inx_ip2=$((inx_ip1 + 1)); inx_ip2<${#ALL_NODES[@]}; inx_ip2++)); do KEY1=0x$(dd if=/dev/urandom count=32 bs=1 2> /dev/null| xxd -p -c 64) KEY2=0x$(dd if=/dev/urandom count=32 bs=1 2> /dev/null| xxd -p -c 64) ID=0x$(dd if=/dev/urandom count=4 bs=1 2> /dev/null| xxd -p -c 8) @@ -748,8 +1066,7 @@ do done done - for ip in ${!CONTROLIP} ${COMPUTE_IPS[*]} - do + for ip in ${!CONTROLIP} ${COMPUTE_IPS[*]}; do echo "ip xfrm configuration for node $ip:" ${SSH} $ip "sudo ip xfrm policy list" ${SSH} $ip "sudo ip xfrm state list" @@ -772,6 +1089,10 @@ do sudo ip netns exec pnf_ns ifconfig pnf_veth1 up ${EXTNET_PNF_IP}/24; sudo ovs-vsctl add-port ${PUBLIC_BRIDGE} pnf_veth0; " + # Control Node - set VXLAN TEP IP for Genius Auto TZ + ${SSH} ${!CONTROLIP} " + sudo ovs-vsctl set O . external_ids:tep-ip=${!CONTROLIP}; + " # Control Node - external net internet address simulation ${SSH} ${!CONTROLIP} " @@ -781,8 +1102,7 @@ do # Computes compute_index=1 - for compute_ip in ${COMPUTE_IPS[*]} - do + for compute_ip in ${COMPUTE_IPS[*]}; do # Tunnel from controller to compute COMPUTEPORT=compute$(( compute_index++ ))_vxlan ${SSH} ${!CONTROLIP} " @@ -793,6 +1113,10 @@ do ${SSH} $compute_ip " sudo ovs-vsctl add-port $PUBLIC_BRIDGE $CONTROLPORT -- set interface $CONTROLPORT type=vxlan options:local_ip=$compute_ip options:remote_ip=${!CONTROLIP} options:dst_port=9876 options:key=flow " + #Compute Node - set VXLAN TEP IP for Genius Auto TZ + ${SSH} $compute_ip " + sudo ovs-vsctl set O . external_ids:tep-ip=${compute_ip}; + " done done @@ -821,18 +1145,32 @@ echo "Changing the testplan path..." cat "${testplan_filepath}" | sed "s:integration:${WORKSPACE}:" > testplan.txt cat testplan.txt +# Use the testplan if specific SUITES are not defined. if [ -z "${SUITES}" ]; then SUITES=`egrep -v '(^[[:space:]]*#|^[[:space:]]*$)' testplan.txt | tr '\012' ' '` +else + newsuites="" + workpath="${WORKSPACE}/test/csit/suites" + for suite in ${SUITES}; do + fullsuite="${workpath}/${suite}" + if [ -z ${newsuites} ]; then + newsuites+=${fullsuite} + else + newsuites+=" "${fullsuite} + fi + done + SUITES=${newsuites} fi -#Environment Variables Needed to execute Openstack Client for NEtvirt Jobs +# TODO: run openrc on control node and then scrape the vars from it +# Environment Variables Needed to execute Openstack Client for NetVirt Jobs cat > /tmp/os_netvirt_client_rc << EOF export OS_USERNAME=admin export OS_PASSWORD=admin export OS_PROJECT_NAME=admin export OS_USER_DOMAIN_NAME=default export OS_PROJECT_DOMAIN_NAME=default -export OS_AUTH_URL="http://${!CONTROLIP}:35357/v3" +export OS_AUTH_URL="http://${!CONTROLIP}/identity" export OS_IDENTITY_API_VERSION=3 export OS_IMAGE_API_VERSION=2 export OS_TENANT_NAME=admin @@ -841,10 +1179,18 @@ EOF source /tmp/os_netvirt_client_rc -#FIXME currently support only 1 site echo "Starting Robot test suites ${SUITES} ..." # please add pybot -v arguments on a single line and alphabetized -pybot -N ${TESTPLAN} --removekeywords wuks -c critical -e exclude -e skip_if_${DISTROSTREAM} \ +suite_num=0 +for suite in ${SUITES}; do + # prepend a incrmental counter to the suite name so that the full robot log combining all the suites as is done + # in the rebot step below will list all the suites in chronological order as rebot seems to alphabatize them + let "suite_num = suite_num + 1" + suite_index="$(printf %02d ${suite_num})" + suite_name="$(basename ${suite} | cut -d. -f1)" + log_name="${suite_index}_${suite_name}" + pybot -N ${log_name} --removekeywords wuks -c critical -e exclude -e skip_if_${DISTROSTREAM} \ + --log log_${log_name}.html --report None --output output_${log_name}.xml \ -v BUNDLEFOLDER:${BUNDLEFOLDER} \ -v BUNDLE_URL:${ACTUAL_BUNDLE_URL} \ -v CONTROLLER_USER:${USER} \ @@ -890,14 +1236,17 @@ pybot -N ${TESTPLAN} --removekeywords wuks -c critical -e exclude -e skip_if_${D -v TOOLS_SYSTEM_2_IP:${TOOLS_SYSTEM_2_IP} \ -v USER_HOME:${HOME} \ -v WORKSPACE:/tmp \ - ${TESTOPTIONS} ${SUITES} || true + ${TESTOPTIONS} ${suite} || true +done +#rebot exit codes seem to be different +rebot --output ${WORKSPACE}/output.xml --log log_full.html --report None -N openstack output_*.xml || true -echo "Examining the files in data/log and checking filesize" +echo "Examining the files in data/log and checking file size" ssh ${ODL_SYSTEM_IP} "ls -altr /tmp/${BUNDLEFOLDER}/data/log/" ssh ${ODL_SYSTEM_IP} "du -hs /tmp/${BUNDLEFOLDER}/data/log/*" echo "Tests Executed" -collect_logs_and_exit +collect_logs true # perhaps Jenkins is testing last exit code # vim: ts=4 sw=4 sts=4 et ft=sh :