Merge "Find major version from stream, avoid mismatches"
[releng/builder.git] / jjb / integration / integration-deploy-controller-run-test.sh
1 #@IgnoreInspection BashAddShebang
2 # Activate robotframework virtualenv
3 # ${ROBOT_VENV} comes from the integration-install-robotframework.sh
4 # script.
5 # shellcheck source=${ROBOT_VENV}/bin/activate disable=SC1091
6 source ${ROBOT_VENV}/bin/activate
7
8 CONTROLLERMEM="2048m"
9
10 if [ "${ENABLE_HAPROXY_FOR_NEUTRON}" == "yes" ]; then
11     echo "Configure cluster"
12     AKKACONF=/tmp/${BUNDLEFOLDER}/configuration/initial/akka.conf
13     MODULESCONF=/tmp/${BUNDLEFOLDER}/configuration/initial/modules.conf
14     MODULESHARDSCONF=/tmp/${BUNDLEFOLDER}/configuration/initial/module-shards.conf
15 fi
16
17 if [ ${CONTROLLERSCOPE} == 'all' ]; then
18     ACTUALFEATURES="odl-integration-compatible-with-all,${CONTROLLERFEATURES}"
19     CONTROLLERMEM="3072m"
20     COOLDOWN_PERIOD="180"
21 else
22     ACTUALFEATURES="${CONTROLLERFEATURES}"
23     COOLDOWN_PERIOD="60"
24 fi
25
26 # Some versions of jenkins job builder result in feature list containing spaces
27 # and ending in newline. Remove all that.
28 ACTUALFEATURES=`echo "${ACTUALFEATURES}" | tr -d '\n \r'`
29
30 if [ -f "${WORKSPACE}/test/csit/scriptplans/${TESTPLAN}" ]; then
31     echo "scriptplan exists!!!"
32     echo "Changing the scriptplan path..."
33     cat ${WORKSPACE}/test/csit/scriptplans/${TESTPLAN} | sed "s:integration:${WORKSPACE}:" > scriptplan.txt
34     cat scriptplan.txt
35     for line in $( egrep -v '(^[[:space:]]*#|^[[:space:]]*$)' scriptplan.txt ); do
36         echo "Executing ${line}..."
37         # shellcheck source=${line} disable=SC1091
38         source ${line}
39     done
40 fi
41
42 cat > ${WORKSPACE}/configuration-script.sh <<EOF
43
44 echo "Changing to /tmp"
45 cd /tmp
46
47 echo "Downloading the distribution..."
48 wget --progress=dot:mega '${ACTUAL_BUNDLE_URL}'
49
50 echo "Extracting the new controller..."
51 unzip -q ${BUNDLE}
52
53 echo "Configuring the startup features..."
54 FEATURESCONF=/tmp/${BUNDLEFOLDER}/etc/org.apache.karaf.features.cfg
55 CUSTOMPROP=/tmp/${BUNDLEFOLDER}/etc/custom.properties
56 sed -ie "s/\(featuresBoot=\|featuresBoot =\)/featuresBoot = ${ACTUALFEATURES},/g" \${FEATURESCONF}
57 sed -ie "s%mvn:org.opendaylight.integration/features-integration-index/${BUNDLEVERSION}/xml/features%mvn:org.opendaylight.integration/features-integration-index/${BUNDLEVERSION}/xml/features,mvn:org.opendaylight.integration/features-integration-test/${BUNDLEVERSION}/xml/features,mvn:org.apache.karaf.decanter/apache-karaf-decanter/1.0.0/xml/features%g" \${FEATURESCONF}
58 cat \${FEATURESCONF}
59
60 if [ "${ODL_ENABLE_L3_FWD}" == "yes" ]; then
61     echo "Enable the l3.fwd in custom.properties..."
62     echo "ovsdb.l3.fwd.enabled=yes" >> \${CUSTOMPROP}
63 fi
64 cat \${CUSTOMPROP}
65
66 echo "Configuring the log..."
67 LOGCONF=/tmp/${BUNDLEFOLDER}/etc/org.ops4j.pax.logging.cfg
68 sed -ie 's/log4j.appender.out.maxBackupIndex=10/log4j.appender.out.maxBackupIndex=1/g' \${LOGCONF}
69 # FIXME: Make log size limit configurable from build parameter.
70 sed -ie 's/log4j.appender.out.maxFileSize=1MB/log4j.appender.out.maxFileSize=30GB/g' \${LOGCONF}
71 echo "log4j.logger.org.opendaylight.yangtools.yang.parser.repo.YangTextSchemaContextResolver = WARN" >> \${LOGCONF}
72 # Add custom logging levels
73 # CONTROLLERDEBUGMAP is expected to be a key:value map of space separated values like "module:level module2:level2"
74 # where module is abbreviated and does not include org.opendaylight
75 unset IFS
76 if [ -n "${CONTROLLERDEBUGMAP}" ]; then
77     for kv in ${CONTROLLERDEBUGMAP}; do
78         module=\${kv%%:*}
79         level=\${kv#*:}
80         if [ -n \${module} ] && [ -n \${level} ]; then
81             echo "log4j.logger.org.opendaylight.\${module} = \${level}" >> \${LOGCONF}
82         fi
83     done
84 fi
85 cat \${LOGCONF}
86
87 echo "Configure java home and max memory..."
88 MEMCONF=/tmp/${BUNDLEFOLDER}/bin/setenv
89 sed -ie 's%^# export JAVA_HOME%export JAVA_HOME="\${JAVA_HOME:-${JAVA_HOME}}"%g' \${MEMCONF}
90 sed -ie 's/JAVA_MAX_MEM="2048m"/JAVA_MAX_MEM="${CONTROLLERMEM}"/g' \${MEMCONF}
91 cat \${MEMCONF}
92
93 echo "Listing all open ports on controller system..."
94 netstat -pnatu
95
96 echo "Set Java version"
97 sudo /usr/sbin/alternatives --install /usr/bin/java java ${JAVA_HOME}/bin/java 1
98 sudo /usr/sbin/alternatives --set java ${JAVA_HOME}/bin/java
99 echo "JDK default version..."
100 java -version
101
102 echo "Set JAVA_HOME"
103 export JAVA_HOME="${JAVA_HOME}"
104 # Did you know that in HERE documents, single quote is an ordinary character, but backticks are still executing?
105 JAVA_RESOLVED=\`readlink -e "\${JAVA_HOME}/bin/java"\`
106 echo "Java binary pointed at by JAVA_HOME: \${JAVA_RESOLVED}"
107
108 if [ "${ENABLE_HAPROXY_FOR_NEUTRON}" == "yes" ]; then
109
110     # Copy shard file if exists
111     if [ -f /tmp/custom_shard_config.txt ]; then
112         echo "Custom shard config exists!!!"
113         echo "Copying the shard config..."
114         cp /tmp/custom_shard_config.txt /tmp/${BUNDLEFOLDER}/bin/
115     fi
116
117     echo "Configuring cluster"
118     /tmp/${BUNDLEFOLDER}/bin/configure_cluster.sh \$1 \$2
119
120     echo "Dump akka.conf"
121     cat ${AKKACONF}
122
123     echo "Dump modules.conf"
124     cat ${MODULESCONF}
125
126      echo "Dump module-shards.conf"
127      cat ${MODULESHARDSCONF}
128 fi
129
130 EOF
131
132 # Create the startup script to be run on controller.
133 cat > ${WORKSPACE}/startup-script.sh <<EOF
134
135 echo "Redirecting karaf console output to karaf_console.log"
136 export KARAF_REDIRECT="/tmp/${BUNDLEFOLDER}/data/log/karaf_console.log"
137
138 echo "Starting controller..."
139 /tmp/${BUNDLEFOLDER}/bin/start
140
141 EOF
142
143 cat > ${WORKSPACE}/post-startup-script.sh <<EOF
144
145 echo "Waiting for controller to come up..."
146 COUNT="0"
147 while true; do
148     RESP="\$( curl --user admin:admin -sL -w "%{http_code} %{url_effective}\\n" http://localhost:8181/restconf/modules -o /dev/null )"
149     echo \$RESP
150     if [ "${ENABLE_HAPROXY_FOR_NEUTRON}" == "yes" ]; then
151         SHARD="\$( curl --user admin:admin -sL -w "%{http_code} %{url_effective}\\n" http://localhost:8181/jolokia/read/org.opendaylight.controller:Category=Shards,name=\member-\$1-shard-inventory-config,type=DistributedConfigDatastore)"
152         echo \$SHARD
153     fi
154     if ([[ \$RESP == *"200"* ]] && ([[ "${ENABLE_HAPROXY_FOR_NEUTRON}" != "yes" ]] || [[ \$SHARD  == *'"status":200'* ]])); then
155         echo Controller is UP
156         break
157     elif (( "\$COUNT" > "600" )); then
158         echo Timeout Controller DOWN
159         echo "Dumping first 500K bytes of karaf log..."
160         head --bytes=500K "/tmp/${BUNDLEFOLDER}/data/log/karaf.log"
161         echo "Dumping last 500K bytes of karaf log..."
162         tail --bytes=500K "/tmp/${BUNDLEFOLDER}/data/log/karaf.log"
163         echo "Listing all open ports on controller system"
164         netstat -pnatu
165         exit 1
166     else
167         COUNT=\$(( \${COUNT} + 1 ))
168         sleep 1
169         if [[ \$((\$COUNT % 5)) == 0 ]]; then
170             echo already waited \${COUNT} seconds...
171         fi
172     fi
173 done
174
175 echo "Listing all open ports on controller system..."
176 netstat -pnatu
177
178 function exit_on_log_file_message {
179     echo "looking for \"\$1\" in log file"
180     if grep --quiet "\$1" "/tmp/${BUNDLEFOLDER}/data/log/karaf.log"; then
181         echo ABORTING: found "\$1"
182         echo "Dumping first 500K bytes of karaf log..."
183         head --bytes=500K "/tmp/${BUNDLEFOLDER}/data/log/karaf.log"
184         echo "Dumping last 500K bytes of karaf log..."
185         tail --bytes=500K "/tmp/${BUNDLEFOLDER}/data/log/karaf.log"
186         exit 1
187     fi
188 }
189
190 exit_on_log_file_message 'BindException: Address already in use'
191 exit_on_log_file_message 'server is unhealthy'
192
193 EOF
194
195 [ "$NUM_OPENSTACK_SITES" ] || NUM_OPENSTACK_SITES=1
196 NUM_ODLS_PER_SITE=$((NUM_ODL_SYSTEM / NUM_OPENSTACK_SITES))
197 for i in `seq 1 ${NUM_OPENSTACK_SITES}`
198 do
199     # Get full list of ODL nodes for this site
200     odl_node_list=
201     for j in `seq 1 ${NUM_ODLS_PER_SITE}`
202     do
203         odl_ip=ODL_SYSTEM_$(((i - 1) * NUM_ODLS_PER_SITE + j))_IP
204         odl_node_list="${odl_node_list} ${!odl_ip}"
205     done
206
207     for j in `seq 1 ${NUM_ODLS_PER_SITE}`
208     do
209         odl_ip=ODL_SYSTEM_$(((i - 1) * NUM_ODLS_PER_SITE + j))_IP
210         # Copy over the config script to controller and execute it (parameters are used only for cluster)
211         echo "Execute the configuration script on controller ${!odl_ip} for index $j with node list ${odl_node_list}"
212         scp ${WORKSPACE}/configuration-script.sh ${!odl_ip}:/tmp
213         ssh ${!odl_ip} "bash /tmp/configuration-script.sh ${j} '${odl_node_list}'"
214     done
215 done
216
217 echo "Locating config plan to use..."
218 configplan_filepath="${WORKSPACE}/test/csit/configplans/${STREAMTESTPLAN}"
219 if [ ! -f "${configplan_filepath}" ]; then
220     configplan_filepath="${WORKSPACE}/test/csit/configplans/${TESTPLAN}"
221 fi
222
223 if [ -f "${configplan_filepath}" ]; then
224     echo "configplan exists!!!"
225     echo "Changing the configplan path..."
226     cat ${configplan_filepath} | sed "s:integration:${WORKSPACE}:" > configplan.txt
227     cat configplan.txt
228     for line in $( egrep -v '(^[[:space:]]*#|^[[:space:]]*$)' configplan.txt ); do
229         echo "Executing ${line}..."
230         # shellcheck source=${line} disable=SC1091
231         source ${line}
232     done
233 fi
234
235 # Copy over the startup script to controller and execute it.
236 for i in `seq 1 ${NUM_ODL_SYSTEM}`
237 do
238     CONTROLLERIP=ODL_SYSTEM_${i}_IP
239     echo "Execute the startup script on controller ${!CONTROLLERIP}"
240     scp ${WORKSPACE}/startup-script.sh ${!CONTROLLERIP}:/tmp
241     ssh ${!CONTROLLERIP} "bash /tmp/startup-script.sh"
242 done
243
244 seed_index=1
245 for i in `seq 1 ${NUM_ODL_SYSTEM}`
246 do
247     CONTROLLERIP=ODL_SYSTEM_${i}_IP
248     echo "Execute the post startup script on controller ${!CONTROLLERIP}"
249     scp ${WORKSPACE}/post-startup-script.sh ${!CONTROLLERIP}:/tmp
250     ssh ${!CONTROLLERIP} "bash /tmp/post-startup-script.sh $(( seed_index++ ))"
251     if [ $(( $i % (${NUM_ODL_SYSTEM} / ${NUM_OPENSTACK_SITES}) )) == 0 ]; then
252         seed_index=1
253     fi
254 done
255
256 echo "Cool down for ${COOLDOWN_PERIOD} seconds :)..."
257 sleep ${COOLDOWN_PERIOD}
258
259 echo "Generating controller variables..."
260 for i in `seq 1 ${NUM_ODL_SYSTEM}`
261 do
262     CONTROLLERIP=ODL_SYSTEM_${i}_IP
263     odl_variables=${odl_variables}" -v ${CONTROLLERIP}:${!CONTROLLERIP}"
264     echo "Lets's take the karaf thread dump"
265     KARAF_PID=$(ssh ${!CONTROLLERIP} "ps aux | grep ${KARAF_ARTIFACT} | grep -v grep | tr -s ' ' | cut -f2 -d' '")
266     ssh ${!CONTROLLERIP} "jstack $KARAF_PID"> ${WORKSPACE}/karaf_${i}_threads_before.log || true
267 done
268
269 if [ ${NUM_OPENSTACK_SYSTEM} -gt 0 ]; then
270    echo "Exiting without running tests to deploy openstack for testing"
271    exit
272 fi
273
274 echo "Generating mininet variables..."
275 for i in `seq 1 ${NUM_TOOLS_SYSTEM}`
276 do
277     MININETIP=TOOLS_SYSTEM_${i}_IP
278     tools_variables=${tools_variables}" -v ${MININETIP}:${!MININETIP}"
279 done
280
281 echo "Locating test plan to use..."
282 testplan_filepath="${WORKSPACE}/test/csit/testplans/${STREAMTESTPLAN}"
283 if [ ! -f "${testplan_filepath}" ]; then
284     testplan_filepath="${WORKSPACE}/test/csit/testplans/${TESTPLAN}"
285 fi
286
287 echo "Changing the testplan path..."
288 cat "${testplan_filepath}" | sed "s:integration:${WORKSPACE}:" > testplan.txt
289 cat testplan.txt
290 SUITES=$( egrep -v '(^[[:space:]]*#|^[[:space:]]*$)' testplan.txt | tr '\012' ' ' )
291
292 echo "Starting Robot test suites ${SUITES} ..."
293 pybot -N ${TESTPLAN} --removekeywords wuks -c critical -e exclude -e skip_if_${DISTROSTREAM} -v BUNDLEFOLDER:${BUNDLEFOLDER} -v WORKSPACE:/tmp \
294 -v JAVA_HOME:${JAVA_HOME} -v BUNDLE_URL:${ACTUAL_BUNDLE_URL} -v NEXUSURL_PREFIX:${NEXUSURL_PREFIX} \
295 -v CONTROLLER:${ODL_SYSTEM_IP} -v ODL_SYSTEM_IP:${ODL_SYSTEM_IP} -v ODL_SYSTEM_1_IP:${ODL_SYSTEM_IP} \
296 -v CONTROLLER_USER:${USER} -v ODL_SYSTEM_USER:${USER} \
297 -v TOOLS_SYSTEM_IP:${TOOLS_SYSTEM_IP} -v TOOLS_SYSTEM_2_IP:${TOOLS_SYSTEM_2_IP} -v TOOLS_SYSTEM_3_IP:${TOOLS_SYSTEM_3_IP} \
298 -v TOOLS_SYSTEM_4_IP:${TOOLS_SYSTEM_4_IP} -v TOOLS_SYSTEM_5_IP:${TOOLS_SYSTEM_5_IP} -v TOOLS_SYSTEM_6_IP:${TOOLS_SYSTEM_6_IP} \
299 -v TOOLS_SYSTEM_USER:${USER} -v JDKVERSION:${JDKVERSION} -v ODL_STREAM:${DISTROSTREAM} -v NUM_ODL_SYSTEM:${NUM_ODL_SYSTEM} \
300 -v MININET:${TOOLS_SYSTEM_IP} -v MININET1:${TOOLS_SYSTEM_2_IP} -v MININET2:${TOOLS_SYSTEM_3_IP} \
301 -v MININET3:${TOOLS_SYSTEM_4_IP} -v MININET4:${TOOLS_SYSTEM_5_IP} -v MININET5:${TOOLS_SYSTEM_6_IP} \
302 -v MININET_USER:${USER} -v USER_HOME:${HOME} ${TESTOPTIONS} ${SUITES} || true
303 # FIXME: Sort (at least -v) options alphabetically.
304
305 echo "Examining the files in data/log and checking filesize"
306 ssh ${ODL_SYSTEM_IP} "ls -altr /tmp/${BUNDLEFOLDER}/data/log/"
307 ssh ${ODL_SYSTEM_IP} "du -hs /tmp/${BUNDLEFOLDER}/data/log/*"
308
309 for i in `seq 1 ${NUM_ODL_SYSTEM}`
310 do
311     CONTROLLERIP=ODL_SYSTEM_${i}_IP
312     echo "Lets's take the karaf thread dump again..."
313     KARAF_PID=$(ssh ${!CONTROLLERIP} "ps aux | grep ${KARAF_ARTIFACT} | grep -v grep | tr -s ' ' | cut -f2 -d' '")
314     ssh ${!CONTROLLERIP} "jstack $KARAF_PID"> ${WORKSPACE}/karaf_${i}_threads_after.log || true
315     echo "Killing ODL"
316     set +e  # We do not want to create red dot just because something went wrong while fetching logs.
317     ssh "${!CONTROLLERIP}" bash -c 'ps axf | grep karaf | grep -v grep | awk '"'"'{print "kill -9 " $1}'"'"' | sh'
318 done
319
320 sleep 5
321 # FIXME: Unify the copy process between various scripts.
322 # TODO: Use rsync.
323 for i in `seq 1 ${NUM_ODL_SYSTEM}`
324 do
325     CONTROLLERIP=ODL_SYSTEM_${i}_IP
326     echo "Compressing karaf.log ${i}"
327     ssh ${!CONTROLLERIP} gzip --best /tmp/${BUNDLEFOLDER}/data/log/karaf.log
328     echo "Fetching compressed karaf.log ${i}"
329     scp "${!CONTROLLERIP}:/tmp/${BUNDLEFOLDER}/data/log/karaf.log.gz" "odl${i}_karaf.log.gz" && ssh ${!CONTROLLERIP} rm -f "/tmp/${BUNDLEFOLDER}/data/log/karaf.log.gz"
330     # TODO: Should we compress the output log file as well?
331     scp "${!CONTROLLERIP}:/tmp/${BUNDLEFOLDER}/data/log/karaf_console.log" "odl${i}_karaf_console.log" && ssh ${!CONTROLLERIP} rm -f "/tmp/${BUNDLEFOLDER}/data/log/karaf_console.log"
332     echo "Fetch GC logs"
333     # FIXME: Put member index in filename, instead of directory name.
334     mkdir -p "gclogs-${i}"
335     scp "${!CONTROLLERIP}:/tmp/${BUNDLEFOLDER}/data/log/*.log" "gclogs-${i}/" && ssh ${!CONTROLLERIP} rm -f "/tmp/${BUNDLEFOLDER}/data/log/*.log"
336 done
337
338 echo "Examine copied files"
339 ls -lt
340
341 true  # perhaps Jenkins is testing last exit code
342
343 # vim: ts=4 sw=4 sts=4 et ft=sh :