# Project branches
stream:
- - beryllium:
+ - boron:
branch: 'master'
+ - beryllium:
+ branch: 'stable/beryllium'
- stable-lithium:
branch: 'stable/lithium'
- boron:
branch: master
jdks: openjdk7,openjdk8
+ autorelease: true
- beryllium:
branch: beryllium
jdks: openjdk7,openjdk8
+ autorelease: true
- stable-lithium:
branch: stable/lithium
jdks: openjdk7
jdks:
- openjdk7
- openjdk8
- disable_autorelease: True
+ disable_autorelease: False
- beryllium:
branch: 'beryllium'
jdk: openjdk7
jdks:
- openjdk7
- openjdk8
- disable_autorelease: True
+ disable_autorelease: False
- stable-lithium:
branch: 'stable/lithium'
jdk: openjdk7
--- /dev/null
+- project:
+ name: didm-csit-discovery
+ jobs:
+ - '{project}-csit-1node-{functionality}-{install}-{stream}'
+ - '{project}-csit-verify-1node-{functionality}'
+
+ # The project name
+ project: 'didm'
+
+ # The functionality under test
+ functionality: 'discovery'
+
+ # Project branches
+ stream:
+ - boron:
+ branch: 'master'
+ - beryllium:
+ branch: 'stable/beryllium'
+ - stable-lithium:
+ branch: 'stable/lithium'
+
+ install:
+ - only:
+ scope: 'only'
+
+ # Features to install
+ install-features: 'odl-didm-all'
+
+ # Robot custom options
+ robot-options: ''
+
+ # Trigger jobs (upstream dependencies)
+ trigger-jobs: 'didm-distribution-{stream},openflowplugin-distribution-{stream}'
--- /dev/null
+- project:
+ name: groupbasedpolicy-csit-netvirt
+ jobs:
+ - '{project}-csit-verify-1node-{functionality}'
+
+ # The project name
+ project: 'groupbasedpolicy'
+
+ # The functionality under test
+ functionality: 'devstack'
+
+ # Mininet configuration
+ mininet-image: 'rk-c7-devstack'
+ mininet-cloud-name: 'Rackspace DFW - Devstack'
+ mininet-vms: 1
+
+ # Features to install
+ install-features: 'odl-groupbasedpolicy-neutronmapper,odl-restconf'
+
+ # Robot custom options
+ robot-options: ''
+
echo "Installing all features..."
sshpass -p karaf ${WORKSPACE}/${BUNDLEFOLDER}/bin/client -u karaf "feature:install ${ACTUALFEATURES}" || echo $? > ${WORKSPACE}/error.txt
-echo "Killing ODL and fetching Karaf log..."
-tail --bytes=1M "${WORKSPACE}/${BUNDLEFOLDER}/data/log/karaf.log" > "karaf.log"
+echo "dumping first 500K bytes of karaf log..."
+head --bytes=500K "${WORKSPACE}/${BUNDLEFOLDER}/data/log/karaf.log" > "karaf.log"
+echo "dumping last 500K bytes of karaf log..."
+tail --bytes=500K "${WORKSPACE}/${BUNDLEFOLDER}/data/log/karaf.log" > "karaf.log"
# TODO: Do we want different name for karaf.log chunk to signal it may be not complete?
+echo "killing karaf process..."
ps axf | grep karaf | grep -v grep | awk '{print "kill -9 " $1}' | sh
sleep 5
xz -9ekvv "${WORKSPACE}/${BUNDLEFOLDER}/data/log/karaf.log"
break
elif (( "\$COUNT" > "600" )); then
echo Timeout Controller DOWN
- echo "Dumping Karaf log..."
- tail --bytes=1M "/tmp/${BUNDLEFOLDER}/data/log/karaf.log"
+ echo "Dumping first 500K bytes of karaf log..."
+ head --bytes=500K "/tmp/${BUNDLEFOLDER}/data/log/karaf.log"
+ echo "Dumping last 500K bytes of karaf log..."
+ tail --bytes=500K "/tmp/${BUNDLEFOLDER}/data/log/karaf.log"
echo "Listing all open ports on controller system"
netstat -natu
exit 1
echo "looking for \"\$1\" in log file"
if grep --quiet "\$1" /tmp/${BUNDLEFOLDER}/data/log/karaf.log; then
echo ABORTING: found "\$1"
- echo "Dumping Karaf log..."
- tail --bytes=1M "/tmp/${BUNDLEFOLDER}/data/log/karaf.log"
+ echo "Dumping first 500K bytes of karaf log..."
+ head --bytes=500K "/tmp/${BUNDLEFOLDER}/data/log/karaf.log"
+ echo "Dumping last 500K bytes of karaf log..."
+ tail --bytes=500K "/tmp/${BUNDLEFOLDER}/data/log/karaf.log"
exit 1
fi
}
echo "looking for \"$1\" in karaf.log file"
if grep --quiet "$1" ${WORKSPACE}/${BUNDLEFOLDER}/data/log/karaf.log; then
echo ABORTING: found "$1"
- echo "Dumping first 1M of karaf.log..."
- tail --bytes=1M ${WORKSPACE}/${BUNDLEFOLDER}/data/log/karaf.log
+ echo "Dumping first 500K bytes of karaf log..."
+ head --bytes=500K ${WORKSPACE}/${BUNDLEFOLDER}/data/log/karaf.log
+ echo "Dumping last 500K bytes of karaf log..."
+ tail --bytes=500K ${WORKSPACE}/${BUNDLEFOLDER}/data/log/karaf.log
cp ${WORKSPACE}/${BUNDLEFOLDER}/data/log/karaf.log .
cp ${WORKSPACE}/${BUNDLEFOLDER}/data/log/karaf_console.log .
exit 1
echo "looking for \"$1\" in karaf_console.log file"
if grep --quiet "$1" ${WORKSPACE}/${BUNDLEFOLDER}/data/log/karaf_console.log; then
echo ABORTING: found "$1"
- echo "Dumping first 1M of karaf_console.log..."
- tail --bytes=1M ${WORKSPACE}/${BUNDLEFOLDER}/data/log/karaf_console.log
+ echo "Dumping first 500K bytes of karaf log..."
+ head --bytes=500K ${WORKSPACE}/${BUNDLEFOLDER}/data/log/karaf_console.log
+ echo "Dumping last 500K bytes of karaf log..."
+ tail --bytes=500K ${WORKSPACE}/${BUNDLEFOLDER}/data/log/karaf_console.log
cp ${WORKSPACE}/${BUNDLEFOLDER}/data/log/karaf.log .
cp ${WORKSPACE}/${BUNDLEFOLDER}/data/log/karaf_console.log .
exit 1
break
elif (( "\$COUNT" > "600" )); then
echo Timeout Controller DOWN
- echo "Dumping Karaf log..."
- tail --bytes=1M "/tmp/${BUNDLEFOLDER}/data/log/karaf.log"
+ echo "Dumping first 500K bytes of karaf log..."
+ head --bytes=500K "/tmp/${BUNDLEFOLDER}/data/log/karaf.log"
+ echo "Dumping last 500K bytes of karaf log..."
+ tail --bytes=500K "/tmp/${BUNDLEFOLDER}/data/log/karaf.log"
echo "Listing all open ports on controller system"
netstat -natu
exit 1
echo "looking for \"\$1\" in log file"
if grep --quiet "\$1" "/tmp/${BUNDLEFOLDER}/data/log/karaf.log"; then
echo ABORTING: found "\$1"
- echo "Dumping Karaf log..."
- tail --bytes=1M "/tmp/${BUNDLEFOLDER}/data/log/karaf.log"
+ echo "Dumping first 500K bytes of karaf log..."
+ head --bytes=500K "/tmp/${BUNDLEFOLDER}/data/log/karaf.log"
+ echo "Dumping last 500K bytes of karaf log..."
+ tail --bytes=500K "/tmp/${BUNDLEFOLDER}/data/log/karaf.log"
exit 1
fi
}
-v MININET:${TOOLS_SYSTEM_IP} -v MININET1:${TOOLS_SYSTEM_2_IP} -v MININET2:${TOOLS_SYSTEM_3_IP} -v MININET_USER:${USER} \
-v USER_HOME:${HOME} ${TESTOPTIONS} ${SUITES} || true
-echo "Killing ODL and fetching Karaf log..."
set +e # We do not want to create red dot just because something went wrong while fetching logs.
for i in `seq 1 ${NUM_ODL_SYSTEM}`
do
CONTROLLERIP=ODL_SYSTEM_${i}_IP
- ssh "${!CONTROLLERIP}" tail --bytes=1M "/tmp/${BUNDLEFOLDER}/data/log/karaf.log" > "odl${i}_karaf.log"
+ echo "dumping first 500K bytes of karaf log..." > "odl${i}_karaf.log"
+ ssh "${!CONTROLLERIP}" head --bytes=500K "/tmp/${BUNDLEFOLDER}/data/log/karaf.log" >> "odl${i}_karaf.log"
+ echo "dumping last 500K bytes of karaf log..." >> "odl${i}_karaf.log"
+ ssh "${!CONTROLLERIP}" tail --bytes=500K "/tmp/${BUNDLEFOLDER}/data/log/karaf.log" >> "odl${i}_karaf.log"
+ echo "killing karaf process..."
ssh "${!CONTROLLERIP}" bash -c 'ps axf | grep karaf | grep -v grep | awk '"'"'{print "kill -9 " $1}'"'"' | sh'
done
sleep 5
STREAMS:
- boron:
branch: master
- jdks: openjdk7,openjdk8
+ jdks: openjdk8
autorelease: true
- beryllium:
branch: stable/beryllium
stream:
- boron:
branch: 'master'
- jdk: openjdk7
+ jdk: openjdk8
jdks:
- - openjdk7
- openjdk8
disable_autorelease: False
- beryllium:
--- /dev/null
+- project:
+ name: netconf-csit-clustering
+ jobs:
+ - '{project}-csit-3node-{functionality}-{install}-{stream}'
+
+ # The project name
+ project: 'netconf'
+
+ # The functionality under test
+ functionality: 'clustering'
+
+ # Project branches
+ stream:
+ - boron:
+ branch: 'master'
+ - beryllium:
+ branch: 'stable/beryllium'
+
+ install:
+ - only:
+ scope: 'only'
+ - all:
+ scope: 'all'
+
+ # Features to install
+ install-features: 'odl-restconf,odl-netconf-clustered-topology'
+
+ # Robot custom options
+ robot-options: ''
+
+ # Trigger jobs (upstream dependencies)
+ trigger-jobs: 'odlparent-distribution-{stream},yangtools-distribution-{stream},controller-distribution-{stream},netconf-distribution-{stream},mdsal-distribution-{stream}'
- project:
name: neutron-daily-openstack
jobs:
+ # TODO: Do we want boron job too?
- 'neutron-daily-openstack-beryllium'
project: 'neutron'
- boron:
branch: master
jdks: openjdk7,openjdk8
+ autorelease: true
- beryllium:
branch: stable/beryllium
jdks: openjdk7,openjdk8
+ autorelease: true
DEPENDENCIES: yangtools,controller,openflowplugin
jdks:
- openjdk7
- openjdk8
- disable_autorelease: True
+ disable_autorelease: False
- beryllium:
branch: 'stable/beryllium'
jdk: openjdk7
jdks:
- openjdk7
- openjdk8
- disable_autorelease: True
+ disable_autorelease: False
project: 'ofextensions/circuitsw'
- project:
- name: openflowplugin-csit-flow-services-beryllium
+ name: openflowplugin-csit-flow-services
jobs:
- '{project}-csit-1node-{functionality}-{install}-{stream}'
- '{project}-csit-verify-1node-{functionality}'
STREAMS:
- boron:
branch: master
- jdks: openjdk7,openjdk8
+ jdks: openjdk8
autorelease: true
- beryllium:
branch: stable/beryllium
stream:
- boron:
branch: 'master'
- jdk: openjdk7
+ jdk: openjdk8
jdks:
- - openjdk7
- openjdk8
disable_autorelease: False
- beryllium:
+- project:
+ name: yangtools-verify-boron
+ jobs:
+ - 'yangtools-verify-{stream}-{maven}-{jdk}'
+
+ jdk:
+ - openjdk8
+
+ # {mvn*} values are defined in releng/builder jjb/releng-defaults.yaml
+ maven:
+ - mvn32:
+ mvnver: '{mvn32}'
+ - mvn33:
+ mvnver: '{mvn33}'
+
+ stream:
+ - boron:
+ branch: master
+
+ project: 'yangtools'
+
- project:
name: yangtools-verify-beryllium
jobs:
mvnver: '{mvn33}'
stream:
- - boron:
- branch: master
- beryllium:
branch: stable/beryllium
for stream_dict in cfg.get("STREAMS"):
streams.update(stream_dict)
else:
- streams = {"beryllium": jjblib.STREAM_DEFAULTS["beryllium"]}
+ streams = {"boron": jjblib.STREAM_DEFAULTS["boron"]}
first_stream = streams.iterkeys().next() # Keep master branch at top.
sonar_branch = streams[first_stream]["branch"]
STREAM_DEFAULTS = collections.OrderedDict([
- ("beryllium", {"branch": "master", "jdks": "openjdk8"}),
+ ("boron", {"branch": "master", "jdks": "openjdk8"}),
+ ("beryllium", {"branch": "stable/beryllium", "jdks": "openjdk8"}),
("stable-lithium", {"branch": "stable/lithium", "jdks": "openjdk7"}),
("stable-helium", {"branch": "stable/helium", "jdks": "openjdk7"}),
])