Merge "Update global-jjb to v0.10.4"
[releng/builder.git] / jjb / integration / integration-deploy-controller-run-test.sh
1 #@IgnoreInspection BashAddShebang
2 # Activate robotframework virtualenv
3 # ${ROBOT_VENV} comes from the integration-install-robotframework.sh
4 # script.
5 # shellcheck source=${ROBOT_VENV}/bin/activate disable=SC1091
6 source ${ROBOT_VENV}/bin/activate
7
8 FEATURESCONF=/tmp/${BUNDLEFOLDER}/etc/org.apache.karaf.features.cfg
9 CUSTOMPROP=/tmp/${BUNDLEFOLDER}/etc/custom.properties
10 LOGCONF=/tmp/${BUNDLEFOLDER}/etc/org.ops4j.pax.logging.cfg
11 MEMCONF=/tmp/${BUNDLEFOLDER}/bin/setenv
12 CONTROLLERMEM="2048m"
13
14 if [ "${ENABLE_HAPROXY_FOR_NEUTRON}" == "yes" ]; then
15     echo "Configure cluster"
16     AKKACONF=/tmp/${BUNDLEFOLDER}/configuration/initial/akka.conf
17     MODULESCONF=/tmp/${BUNDLEFOLDER}/configuration/initial/modules.conf
18     MODULESHARDSCONF=/tmp/${BUNDLEFOLDER}/configuration/initial/module-shards.conf
19 fi
20
21 if [ ${CONTROLLERSCOPE} == 'all' ]; then
22     ACTUALFEATURES="odl-integration-compatible-with-all,${CONTROLLERFEATURES}"
23     CONTROLLERMEM="3072m"
24     COOLDOWN_PERIOD="180"
25 else
26     ACTUALFEATURES="${CONTROLLERFEATURES}"
27     COOLDOWN_PERIOD="60"
28 fi
29
30 # Some versions of jenkins job builder result in feature list containing spaces
31 # and ending in newline. Remove all that.
32 ACTUALFEATURES=`echo "${ACTUALFEATURES}" | tr -d '\n \r'`
33
34 if [ -f "${WORKSPACE}/test/csit/scriptplans/${TESTPLAN}" ]; then
35     echo "scriptplan exists!!!"
36     echo "Changing the scriptplan path..."
37     cat ${WORKSPACE}/test/csit/scriptplans/${TESTPLAN} | sed "s:integration:${WORKSPACE}:" > scriptplan.txt
38     cat scriptplan.txt
39     for line in $( egrep -v '(^[[:space:]]*#|^[[:space:]]*$)' scriptplan.txt ); do
40         echo "Executing ${line}..."
41         # shellcheck source=${line} disable=SC1091
42         source ${line}
43     done
44 fi
45
46 cat > ${WORKSPACE}/configuration-script.sh <<EOF
47
48 echo "Changing to /tmp"
49 cd /tmp
50
51 echo "Downloading the distribution..."
52 wget --progress=dot:mega '${ACTUAL_BUNDLE_URL}'
53
54 echo "Extracting the new controller..."
55 unzip -q ${BUNDLE}
56
57 echo "Configuring the startup features..."
58 sed -ie "s/\(featuresBoot=\|featuresBoot =\)/featuresBoot = ${ACTUALFEATURES},/g" ${FEATURESCONF}
59
60 FEATURE_INDEX_STRING="features-integration-index"
61 FEATURE_TEST_STRING="features-integration-test"
62 if [[ "$KARAF_VERSION" == "karaf4" ]]; then
63     FEATURE_INDEX_STRING="features-index"
64     FEATURE_TEST_STRING="features-test"
65 fi
66
67 sed -ie "s%mvn:org.opendaylight.integration/\${FEATURE_INDEX_STRING}/${BUNDLEVERSION}/xml/features%mvn:org.opendaylight.integration/\${FEATURE_INDEX_STRING}/${BUNDLEVERSION}/xml/features,mvn:org.opendaylight.integration/\${FEATURE_TEST_STRING}/${BUNDLEVERSION}/xml/features,mvn:org.apache.karaf.decanter/apache-karaf-decanter/1.0.0/xml/features%g" ${FEATURESCONF}
68 cat ${FEATURESCONF}
69
70 if [ "${ODL_ENABLE_L3_FWD}" == "yes" ]; then
71     echo "Enable the l3.fwd in custom.properties..."
72     echo "ovsdb.l3.fwd.enabled=yes" >> ${CUSTOMPROP}
73 fi
74 cat ${CUSTOMPROP}
75
76 echo "Configuring the log..."
77 sed -ie 's/log4j.appender.out.maxBackupIndex=10/log4j.appender.out.maxBackupIndex=1/g' ${LOGCONF}
78 # FIXME: Make log size limit configurable from build parameter.
79 sed -ie 's/log4j.appender.out.maxFileSize=1MB/log4j.appender.out.maxFileSize=30GB/g' ${LOGCONF}
80 echo "log4j.logger.org.opendaylight.yangtools.yang.parser.repo.YangTextSchemaContextResolver = WARN" >> ${LOGCONF}
81 # Add custom logging levels
82 # CONTROLLERDEBUGMAP is expected to be a key:value map of space separated values like "module:level module2:level2"
83 # where module is abbreviated and does not include org.opendaylight
84 unset IFS
85 if [ -n "${CONTROLLERDEBUGMAP}" ]; then
86     for kv in ${CONTROLLERDEBUGMAP}; do
87         module=\${kv%%:*}
88         level=\${kv#*:}
89         if [ -n \${module} ] && [ -n \${level} ]; then
90             echo "log4j.logger.org.opendaylight.\${module} = \${level}" >> ${LOGCONF}
91         fi
92     done
93 fi
94 cat ${LOGCONF}
95
96 echo "Configure java home and max memory..."
97 sed -ie 's%^# export JAVA_HOME%export JAVA_HOME="\${JAVA_HOME:-${JAVA_HOME}}"%g' ${MEMCONF}
98 sed -ie 's/JAVA_MAX_MEM="2048m"/JAVA_MAX_MEM="${CONTROLLERMEM}"/g' ${MEMCONF}
99 cat ${MEMCONF}
100
101 echo "Listing all open ports on controller system..."
102 netstat -pnatu
103
104 echo "Set Java version"
105 sudo /usr/sbin/alternatives --install /usr/bin/java java ${JAVA_HOME}/bin/java 1
106 sudo /usr/sbin/alternatives --set java ${JAVA_HOME}/bin/java
107 echo "JDK default version..."
108 java -version
109
110 echo "Set JAVA_HOME"
111 export JAVA_HOME="${JAVA_HOME}"
112 # Did you know that in HERE documents, single quote is an ordinary character, but backticks are still executing?
113 JAVA_RESOLVED=\`readlink -e "\${JAVA_HOME}/bin/java"\`
114 echo "Java binary pointed at by JAVA_HOME: \${JAVA_RESOLVED}"
115
116 if [ "${ENABLE_HAPROXY_FOR_NEUTRON}" == "yes" ]; then
117
118     # Copy shard file if exists
119     if [ -f /tmp/custom_shard_config.txt ]; then
120         echo "Custom shard config exists!!!"
121         echo "Copying the shard config..."
122         cp /tmp/custom_shard_config.txt /tmp/${BUNDLEFOLDER}/bin/
123     fi
124
125     echo "Configuring cluster"
126     /tmp/${BUNDLEFOLDER}/bin/configure_cluster.sh \$1 \$2
127
128     echo "Dump akka.conf"
129     cat ${AKKACONF}
130
131     echo "Dump modules.conf"
132     cat ${MODULESCONF}
133
134      echo "Dump module-shards.conf"
135      cat ${MODULESHARDSCONF}
136 fi
137
138 EOF
139
140 # Create the startup script to be run on controller.
141 cat > ${WORKSPACE}/startup-script.sh <<EOF
142
143 echo "Redirecting karaf console output to karaf_console.log"
144 export KARAF_REDIRECT="/tmp/${BUNDLEFOLDER}/data/log/karaf_console.log"
145
146 echo "Starting controller..."
147 /tmp/${BUNDLEFOLDER}/bin/start
148
149 EOF
150
151 cat > ${WORKSPACE}/post-startup-script.sh <<EOF
152
153 echo "Waiting for controller to come up..."
154 COUNT="0"
155 while true; do
156     RESP="\$( curl --user admin:admin -sL -w "%{http_code} %{url_effective}\\n" http://localhost:8181/restconf/modules -o /dev/null )"
157     echo \$RESP
158     if [ "${ENABLE_HAPROXY_FOR_NEUTRON}" == "yes" ]; then
159         SHARD="\$( curl --user admin:admin -sL -w "%{http_code} %{url_effective}\\n" http://localhost:8181/jolokia/read/org.opendaylight.controller:Category=Shards,name=\member-\$1-shard-inventory-config,type=DistributedConfigDatastore)"
160         echo \$SHARD
161     fi
162     if ([[ \$RESP == *"200"* ]] && ([[ "${ENABLE_HAPROXY_FOR_NEUTRON}" != "yes" ]] || [[ \$SHARD  == *'"status":200'* ]])); then
163         echo Controller is UP
164         break
165     elif (( "\$COUNT" > "600" )); then
166         echo Timeout Controller DOWN
167         echo "Dumping first 500K bytes of karaf log..."
168         head --bytes=500K "/tmp/${BUNDLEFOLDER}/data/log/karaf.log"
169         echo "Dumping last 500K bytes of karaf log..."
170         tail --bytes=500K "/tmp/${BUNDLEFOLDER}/data/log/karaf.log"
171         echo "Listing all open ports on controller system"
172         netstat -pnatu
173         exit 1
174     else
175         COUNT=\$(( \${COUNT} + 1 ))
176         sleep 1
177         if [[ \$((\$COUNT % 5)) == 0 ]]; then
178             echo already waited \${COUNT} seconds...
179         fi
180     fi
181 done
182
183 echo "Listing all open ports on controller system..."
184 netstat -pnatu
185
186 function exit_on_log_file_message {
187     echo "looking for \"\$1\" in log file"
188     if grep --quiet "\$1" "/tmp/${BUNDLEFOLDER}/data/log/karaf.log"; then
189         echo ABORTING: found "\$1"
190         echo "Dumping first 500K bytes of karaf log..."
191         head --bytes=500K "/tmp/${BUNDLEFOLDER}/data/log/karaf.log"
192         echo "Dumping last 500K bytes of karaf log..."
193         tail --bytes=500K "/tmp/${BUNDLEFOLDER}/data/log/karaf.log"
194         exit 1
195     fi
196 }
197
198 exit_on_log_file_message 'BindException: Address already in use'
199 exit_on_log_file_message 'server is unhealthy'
200
201 EOF
202
203 [ "$NUM_OPENSTACK_SITES" ] || NUM_OPENSTACK_SITES=1
204 NUM_ODLS_PER_SITE=$((NUM_ODL_SYSTEM / NUM_OPENSTACK_SITES))
205 for i in `seq 1 ${NUM_OPENSTACK_SITES}`
206 do
207     # Get full list of ODL nodes for this site
208     odl_node_list=
209     for j in `seq 1 ${NUM_ODLS_PER_SITE}`
210     do
211         odl_ip=ODL_SYSTEM_$(((i - 1) * NUM_ODLS_PER_SITE + j))_IP
212         odl_node_list="${odl_node_list} ${!odl_ip}"
213     done
214
215     for j in `seq 1 ${NUM_ODLS_PER_SITE}`
216     do
217         odl_ip=ODL_SYSTEM_$(((i - 1) * NUM_ODLS_PER_SITE + j))_IP
218         # Copy over the config script to controller and execute it (parameters are used only for cluster)
219         echo "Execute the configuration script on controller ${!odl_ip} for index $j with node list ${odl_node_list}"
220         scp ${WORKSPACE}/configuration-script.sh ${!odl_ip}:/tmp
221         ssh ${!odl_ip} "bash /tmp/configuration-script.sh ${j} '${odl_node_list}'"
222     done
223 done
224
225 echo "Locating config plan to use..."
226 configplan_filepath="${WORKSPACE}/test/csit/configplans/${STREAMTESTPLAN}"
227 if [ ! -f "${configplan_filepath}" ]; then
228     configplan_filepath="${WORKSPACE}/test/csit/configplans/${TESTPLAN}"
229 fi
230
231 if [ -f "${configplan_filepath}" ]; then
232     echo "configplan exists!!!"
233     echo "Changing the configplan path..."
234     cat ${configplan_filepath} | sed "s:integration:${WORKSPACE}:" > configplan.txt
235     cat configplan.txt
236     for line in $( egrep -v '(^[[:space:]]*#|^[[:space:]]*$)' configplan.txt ); do
237         echo "Executing ${line}..."
238         # shellcheck source=${line} disable=SC1091
239         source ${line}
240     done
241 fi
242
243 # Copy over the startup script to controller and execute it.
244 for i in `seq 1 ${NUM_ODL_SYSTEM}`
245 do
246     CONTROLLERIP=ODL_SYSTEM_${i}_IP
247     echo "Execute the startup script on controller ${!CONTROLLERIP}"
248     scp ${WORKSPACE}/startup-script.sh ${!CONTROLLERIP}:/tmp
249     ssh ${!CONTROLLERIP} "bash /tmp/startup-script.sh"
250 done
251
252 seed_index=1
253 for i in `seq 1 ${NUM_ODL_SYSTEM}`
254 do
255     CONTROLLERIP=ODL_SYSTEM_${i}_IP
256     echo "Execute the post startup script on controller ${!CONTROLLERIP}"
257     scp ${WORKSPACE}/post-startup-script.sh ${!CONTROLLERIP}:/tmp
258     ssh ${!CONTROLLERIP} "bash /tmp/post-startup-script.sh $(( seed_index++ ))"
259     if [ $(( $i % (${NUM_ODL_SYSTEM} / ${NUM_OPENSTACK_SITES}) )) == 0 ]; then
260         seed_index=1
261     fi
262 done
263
264 echo "Cool down for ${COOLDOWN_PERIOD} seconds :)..."
265 sleep ${COOLDOWN_PERIOD}
266
267 echo "Generating controller variables..."
268 for i in `seq 1 ${NUM_ODL_SYSTEM}`
269 do
270     CONTROLLERIP=ODL_SYSTEM_${i}_IP
271     odl_variables=${odl_variables}" -v ${CONTROLLERIP}:${!CONTROLLERIP}"
272     echo "Lets's take the karaf thread dump"
273     KARAF_PID=$(ssh ${!CONTROLLERIP} "ps aux | grep ${KARAF_ARTIFACT} | grep -v grep | tr -s ' ' | cut -f2 -d' '")
274     ssh ${!CONTROLLERIP} "jstack $KARAF_PID"> ${WORKSPACE}/karaf_${i}_threads_before.log || true
275 done
276
277 if [ ${NUM_OPENSTACK_SYSTEM} -gt 0 ]; then
278    echo "Exiting without running tests to deploy openstack for testing"
279    exit
280 fi
281
282 echo "Generating mininet variables..."
283 for i in `seq 1 ${NUM_TOOLS_SYSTEM}`
284 do
285     MININETIP=TOOLS_SYSTEM_${i}_IP
286     tools_variables=${tools_variables}" -v ${MININETIP}:${!MININETIP}"
287 done
288
289 echo "Locating test plan to use..."
290 testplan_filepath="${WORKSPACE}/test/csit/testplans/${STREAMTESTPLAN}"
291 if [ ! -f "${testplan_filepath}" ]; then
292     testplan_filepath="${WORKSPACE}/test/csit/testplans/${TESTPLAN}"
293 fi
294
295 echo "Changing the testplan path..."
296 cat "${testplan_filepath}" | sed "s:integration:${WORKSPACE}:" > testplan.txt
297 cat testplan.txt
298 SUITES=$( egrep -v '(^[[:space:]]*#|^[[:space:]]*$)' testplan.txt | tr '\012' ' ' )
299
300 echo "Starting Robot test suites ${SUITES} ..."
301 pybot -N ${TESTPLAN} --removekeywords wuks -c critical -e exclude -e skip_if_${DISTROSTREAM} -v BUNDLEFOLDER:${BUNDLEFOLDER} -v WORKSPACE:/tmp \
302 -v JAVA_HOME:${JAVA_HOME} -v BUNDLE_URL:${ACTUAL_BUNDLE_URL} -v NEXUSURL_PREFIX:${NEXUSURL_PREFIX} \
303 -v CONTROLLER:${ODL_SYSTEM_IP} -v ODL_SYSTEM_IP:${ODL_SYSTEM_IP} -v ODL_SYSTEM_1_IP:${ODL_SYSTEM_IP} \
304 -v CONTROLLER_USER:${USER} -v ODL_SYSTEM_USER:${USER} \
305 -v TOOLS_SYSTEM_IP:${TOOLS_SYSTEM_IP} -v TOOLS_SYSTEM_2_IP:${TOOLS_SYSTEM_2_IP} -v TOOLS_SYSTEM_3_IP:${TOOLS_SYSTEM_3_IP} \
306 -v TOOLS_SYSTEM_4_IP:${TOOLS_SYSTEM_4_IP} -v TOOLS_SYSTEM_5_IP:${TOOLS_SYSTEM_5_IP} -v TOOLS_SYSTEM_6_IP:${TOOLS_SYSTEM_6_IP} \
307 -v TOOLS_SYSTEM_USER:${USER} -v JDKVERSION:${JDKVERSION} -v ODL_STREAM:${DISTROSTREAM} -v NUM_ODL_SYSTEM:${NUM_ODL_SYSTEM} \
308 -v MININET:${TOOLS_SYSTEM_IP} -v MININET1:${TOOLS_SYSTEM_2_IP} -v MININET2:${TOOLS_SYSTEM_3_IP} \
309 -v MININET3:${TOOLS_SYSTEM_4_IP} -v MININET4:${TOOLS_SYSTEM_5_IP} -v MININET5:${TOOLS_SYSTEM_6_IP} \
310 -v MININET_USER:${USER} -v USER_HOME:${HOME} ${TESTOPTIONS} ${SUITES} || true
311 # FIXME: Sort (at least -v) options alphabetically.
312
313 echo "Examining the files in data/log and checking filesize"
314 ssh ${ODL_SYSTEM_IP} "ls -altr /tmp/${BUNDLEFOLDER}/data/log/"
315 ssh ${ODL_SYSTEM_IP} "du -hs /tmp/${BUNDLEFOLDER}/data/log/*"
316
317 for i in `seq 1 ${NUM_ODL_SYSTEM}`
318 do
319     CONTROLLERIP=ODL_SYSTEM_${i}_IP
320     echo "Lets's take the karaf thread dump again..."
321     KARAF_PID=$(ssh ${!CONTROLLERIP} "ps aux | grep ${KARAF_ARTIFACT} | grep -v grep | tr -s ' ' | cut -f2 -d' '")
322     ssh ${!CONTROLLERIP} "jstack $KARAF_PID"> ${WORKSPACE}/karaf_${i}_threads_after.log || true
323     echo "Killing ODL"
324     set +e  # We do not want to create red dot just because something went wrong while fetching logs.
325     ssh "${!CONTROLLERIP}" bash -c 'ps axf | grep karaf | grep -v grep | awk '"'"'{print "kill -9 " $1}'"'"' | sh'
326 done
327
328 sleep 5
329 # FIXME: Unify the copy process between various scripts.
330 # TODO: Use rsync.
331 for i in `seq 1 ${NUM_ODL_SYSTEM}`
332 do
333     CONTROLLERIP=ODL_SYSTEM_${i}_IP
334     echo "Compressing karaf.log ${i}"
335     ssh ${!CONTROLLERIP} gzip --best /tmp/${BUNDLEFOLDER}/data/log/karaf.log
336     echo "Fetching compressed karaf.log ${i}"
337     scp "${!CONTROLLERIP}:/tmp/${BUNDLEFOLDER}/data/log/karaf.log.gz" "odl${i}_karaf.log.gz" && ssh ${!CONTROLLERIP} rm -f "/tmp/${BUNDLEFOLDER}/data/log/karaf.log.gz"
338     # TODO: Should we compress the output log file as well?
339     scp "${!CONTROLLERIP}:/tmp/${BUNDLEFOLDER}/data/log/karaf_console.log" "odl${i}_karaf_console.log" && ssh ${!CONTROLLERIP} rm -f "/tmp/${BUNDLEFOLDER}/data/log/karaf_console.log"
340     echo "Fetch GC logs"
341     # FIXME: Put member index in filename, instead of directory name.
342     mkdir -p "gclogs-${i}"
343     scp "${!CONTROLLERIP}:/tmp/${BUNDLEFOLDER}/data/log/*.log" "gclogs-${i}/" && ssh ${!CONTROLLERIP} rm -f "/tmp/${BUNDLEFOLDER}/data/log/*.log"
344 done
345
346 echo "Examine copied files"
347 ls -lt
348
349 true  # perhaps Jenkins is testing last exit code
350
351 # vim: ts=4 sw=4 sts=4 et ft=sh :