--- /dev/null
+#!/bin/bash
+
+# make sure we don't require tty for sudo operations
+cat <<EOF >/etc/sudoers.d/89-jenkins-user-defaults
+Defaults:jenkins !requiretty
+jenkins ALL = NOPASSWD: ALL
+EOF
+
+# make sure the firewall is stopped
+service iptables stop
+
+# vim: sw=2 ts=2 sts=2 et :
# built, they are now. Make sure they're install
yum install -q -y netopeer-server-sl CPqD-ofsoftswitch13
+# netaddr and ipaddress libraries can be useful on this system as
+# some tests are starting to push pyhon scripts/tools to this VM
+# during CI tests
+yum install -q -y python-{ipaddress,netaddr}
+
# the vagrant configuration for netopeer doesn't configure SSH correctly
# as it uses and here document via echo and not cat fix that
cat << EOSSH >> /etc/ssh/sshd_config
- project:
- name: bgpcep-csit-userfeatures-master
+ name: bgpcep-csit-userfeatures
jobs:
- '{project}-csit-1node-cds-{functionality}-{install}-{stream}'
+ - '{project}-csit-verify-{functionality}-{stream}'
# The project name
project: 'bgpcep'
stream:
- master:
branch: 'master'
+ - stable-lithium:
+ branch: 'stable/lithium'
install:
- only:
stream:
- master:
branch: 'master'
+ - stable-lithium:
+ branch: 'stable/lithium'
project: 'bgpcep'
jdk: 'openjdk7'
- integration-bundleurl:
bundleurl: 'last'
+ triggers:
+ - timed: 'H H * * *'
+
builders:
- trigger-builds:
- project: '{csit-list}'
--- /dev/null
+- project:
+ name: integration-job-patch-test-stable-lithium
+ jobs:
+ - 'integration-patch-test-stable-lithium'
+
+ jdk: 'openjdk7'
+
+- job-template:
+ name: 'integration-patch-test-stable-lithium'
+
+ project-type: maven
+ node: dynamic_verify
+ jdk: '{jdk}'
+
+ logrotate:
+ daysToKeep: '{build-days-to-keep}'
+ numToKeep: '{build-num-to-keep}'
+ artifactDaysToKeep: '{build-artifact-days-to-keep}'
+ artifactNumToKeep: '{build-artifact-num-to-keep}'
+
+ parameters:
+ - project-parameter:
+ project: 'integration'
+ - integration-patch-git-url
+
+ scm:
+ - integration-gerrit-scm:
+ credentials-id: '{ssh-credentials}'
+ branch: 'stable/lithium'
+ - integration-git-scm:
+ credentials-id: '{ssh-credentials}'
+ branch: 'master'
+
+ wrappers:
+ - build-timeout
+ - ssh-agent-credentials:
+ users:
+ - '{ssh-credentials}'
+
+ triggers:
+ - gerrit:
+ server-name: 'OpenDaylight'
+ trigger-on:
+ - comment-added-contains-event:
+ comment-contains-value: 'test-integration'
+ projects:
+ - project-compare-type: 'ANT'
+ project-pattern: '**'
+ branches:
+ - branch-compare-type: 'ANT'
+ branch-pattern: '**/stable/lithium'
+ skip-vote:
+ successful: true
+ failed: true
+ unstable: true
+ notbuilt: true
+
+ prebuilders:
+ - integration-get-bundle-url
+ - inject:
+ properties-file: 'bundle.txt'
+ - wipe-org-opendaylight-repo
+ - maven-target:
+ maven-version: '{mvn32}'
+ pom: '$GERRIT_PROJECT/pom.xml'
+ goals: 'clean install -Dmaven.repo.local=/tmp/r -Dorg.ops4j.pax.url.mvn.localRepository=/tmp/r source:jar javadoc:jar'
+ java-opts:
+ - '-Xmx4096m -XX:MaxPermSize=1024m -Dmaven.compile.fork=true'
+ settings: '{integration-settings}'
+ global-settings: '{odl-global-settings}'
+
+ maven:
+ maven-name: '{mvn32}'
+ root-pom: 'integration/pom.xml'
+ goals: 'clean install -Dmaven.repo.local=/tmp/r -Dorg.ops4j.pax.url.mvn.localRepository=/tmp/r source:jar javadoc:jar'
+ maven-opts: '-Xmx1024m -XX:MaxPermSize=256m -Dmaven.compile.fork=true'
+ settings: '{integration-settings}'
+ global-settings: '{odl-global-settings}'
+ ignore-upstream-changes: true
+
+ postbuilders:
+ - trigger-builds:
+ - project: 'integration-distribution-test-master'
+ block: true
+ predefined-parameters:
+ BUNDLEURL=$BUNDLEURL
+
+ publishers:
+ - email-notification:
+ email-prefix: '[integration]'
+
--- /dev/null
+- project:
+ name: openflowplugin-csit-flow-services-more-lithium-redesign
+ jobs:
+ - '{project}-csit-1node-cds-{functionality}-{install}-{stream}'
+
+ # The project name
+ project: 'openflowplugin'
+
+ # The functionality under test
+ functionality: 'flow-services-more-lithium-redesign'
+
+ # Project branches
+ stream:
+ - master:
+ branch: 'master'
+
+ install:
+ - only:
+ scope: 'only'
+
+ # Features to install
+ install-features: 'odl-openflowplugin-flow-services-ui-li'
+
+ # Robot custom options
+ robot-options: '-e adsal'
+
+ # Trigger jobs (upstream dependencies)
+ trigger-jobs: 'yangtools-distribution-{stream},controller-distribution-{stream},openflowjava-distribution-{stream},openflowplugin-distribution-{stream}'
--- /dev/null
+- project:
+ name: openflowplugin-csit-flow-services-more-master
+ jobs:
+ - '{project}-csit-1node-cds-{functionality}-{install}-{stream}'
+
+ # The project name
+ project: 'openflowplugin'
+
+ # The functionality under test
+ functionality: 'flow-services-more'
+
+ # Project branches
+ stream:
+ - master:
+ branch: 'master'
+
+ install:
+ - only:
+ scope: 'only'
+
+ # Features to install
+ install-features: 'odl-openflowplugin-flow-services-ui,odl-openflowplugin-adsal-compatibility,odl-nsf-all'
+
+ # Robot custom options
+ robot-options: ''
+
+ # Trigger jobs (upstream dependencies)
+ trigger-jobs: 'yangtools-distribution-{stream},controller-distribution-{stream},openflowjava-distribution-{stream},openflowplugin-distribution-{stream}'
+
robot-options: ''
# Trigger jobs (upstream dependencies)
- trigger-jobs: 'yangtools-distribution-{stream},controller-distribution-{stream},openflowjava-distribution-{stream},openflowplugin-distribution-{stream}'
+ trigger-jobs: ''
+BRANCHES:
+- master:
+ jdks: openjdk7
+- stable/helium:
+ jdks: openjdk7
+- stable/lithium:
+ jdks: openjdk7
DEPENDENCIES: odlparent,controller,yangtools
jdk: openjdk7
jdks:
- openjdk7
+ - stable-lithium:
+ branch: 'stable/lithium'
+ jdk: openjdk7
+ jdks:
+ - openjdk7
project: 'snmp4sdn'
--- /dev/null
+basic-mininet-fedora-node can be used to take an already converted
+Rackspace native base image to a basic system with mininet and other
+test tools for use in the OpenDaylight build and testing environment
+
+Please see the rackspace-convert-base vagrant setup for creation of the
+needed base image.
+
+This vagrant expects (by default) a personal Rackspace image named
+
+'Fedora 21 - Vagrant ready'
+
+To spin up and utilize.
+
+$ RSIMAGE='${baseimagename}' vagrant up --provider=rackspace
+
+Will execute this vagrant against a differently named base image
+
+$ RSRESEAL=true vagrant up --provider=rackspace
+
+NOTE: resealing will cause the vagrant to run the resealing operation.
+This operation will intentionally destroy current SSH pubkeys installed
+on the system as well as reset log files and network configurations. You
+have been warned.
--- /dev/null
+# -*- mode: ruby -*-
+# vi: set ft=ruby sw=2 ts=2 sts=2 et :
+
+# Vagrantfile API/syntax version. Don't touch unless you know what you're doing!
+VAGRANTFILE_API_VERSION = "2"
+
+Vagrant.configure(VAGRANTFILE_API_VERSION) do |config|
+ # root off of the rackspace provider dummy box
+ config.vm.box = "dummy"
+
+ # rackspace systems, even with cloud-init
+ # don't seem to have the cloud int user ${osname} (or similar)
+ # getting the ssh key for some reason, root does for sure
+ # so use that
+ config.ssh.username = 'root'
+
+ # Fedora and EL systems default to requiring tty for sudo
+ # This should have been disabled with the Vagrant ready
+ # base box conversion (see rackspace-convert-base vagrant)
+ # but just to be safe
+ config.ssh.pty = true
+
+ # make sure to set the following in your
+ # ~/.vagrant.d/boxes/dummy/0/rackspace/Vagrantfile
+ # rs.username
+ # rs.api_key
+ # rs.rackspace_region
+ #
+ # If you are not using an SSH token / smartcard also set this
+ # rs.key_name
+ # config.ssh.private_key_path -- set this outside the rackspace block
+ # in your base box
+ config.vm.provider :rackspace do |rs|
+ # create these base builds always on the smallest system possible
+ rs.flavor = 'general1-1'
+
+ # allow for switching to ORD cloud but default to DFW
+ if (ENV['RSREGION'] == 'ord')
+ rs.rackspace_region = :ord
+ else
+ rs.rackspace_region = :dfw
+ end
+
+ # Default the Fedora 21 - Vagrant ready image unless overriden by a RSIMAGE
+ # environment variable
+ if ENV['RSIMAGE']
+ rs.image = ENV['RSIMAGE']
+ else
+ rs.image = 'Fedora 21 - Vagrant ready'
+ end
+ end
+
+ # run our bootstrapping for the ovsdb-devstack system
+ config.vm.provision 'shell', path: 'bootstrap.sh'
+
+ # set RSRESEAL to... anything if you want to snap an image of this box
+ # not setting the environment variable will cause the system to come
+ # up fully and not be in a resealable state
+ if ENV['RSRESEAL']
+ config.vm.provision 'shell', path: 'system_reseal.sh'
+ end
+end
--- /dev/null
+#!/bin/bash
+
+# vim: sw=4 ts=4 sts=4 et tw=72 :
+
+# update os
+yum clean all
+yum update -q -y
+
+# install mininet
+git clone git://github.com/mininet/mininet
+cd mininet
+git checkout -b 2.2.1 2.2.1
+cd ..
+mininet/util/install.sh -nf
+
--- /dev/null
+#!/bin/bash
+
+# clean-up from any prior cloud-init networking
+rm -rf /etc/sysconfig/network-scripts/{ifcfg,route}-eth*
+
+rm -rf /etc/Pegasus/*.cnf /etc/Pegasus/*.crt /etc/Pegasus/*.csr /etc/Pegasus/*.pem /etc/Pegasus/*.srl /root/anaconda-ks.cfg /root/anaconda-post.log /root/initial-setup-ks.cfg /root/install.log /root/install.log.syslog /var/cache/fontconfig/* /var/cache/gdm/* /var/cache/man/* /var/lib/AccountService/users/* /var/lib/fprint/* /var/lib/logrotate.status /var/log/*.log* /var/log/BackupPC/LOG /var/log/ConsoleKit/* /var/log/anaconda.syslog /var/log/anaconda/* /var/log/apache2/*_log /var/log/apache2/*_log-* /var/log/apt/* /var/log/aptitude* /var/log/audit/* /var/log/btmp* /var/log/ceph/*.log /var/log/chrony/*.log /var/log/cron* /var/log/cups/*_log /var/log/debug* /var/log/dmesg* /var/log/exim4/* /var/log/faillog* /var/log/gdm/* /var/log/glusterfs/*glusterd.vol.log /var/log/glusterfs/glusterfs.log /var/log/httpd/*log /var/log/installer/* /var/log/jetty/jetty-console.log /var/log/journal/* /var/log/lastlog* /var/log/libvirt/libvirtd.log /var/log/libvirt/lxc/*.log /var/log/libvirt/qemu/*.log /var/log/libvirt/uml/*.log /var/log/lightdm/* /var/log/mail/* /var/log/maillog* /var/log/messages* /var/log/ntp /var/log/ntpstats/* /var/log/ppp/connect-errors /var/log/rhsm/* /var/log/sa/* /var/log/secure* /var/log/setroubleshoot/*.log /var/log/spooler* /var/log/squid/*.log /var/log/syslog* /var/log/tallylog* /var/log/tuned/tuned.log /var/log/wtmp* /var/named/data/named.run
+
+rm -rf ~/.viminfo /etc/ssh/ssh*key*
+
+echo "********************************************"
+echo "* PLEASE SNAPSHOT IMAGE AT THIS TIME *"
+echo "********************************************"