+++ /dev/null
-IMAGE_NAME=ZZCI - CentOS 7 - builder - x86_64 - 20200811-062543.500
-VOLUME_SIZE=100
-HARDWARE_ID=odl-highcpu-2
- odl-maven-jobs-jdk11
- odl-maven-verify-jobs
- csit-gate-list: "netvirt-csit-1node-0cmb-1ctl-2cmp-apex-queens-gate-snat-conntrack-{stream}"
- block_on_csit: false
-
stream: silicon
project: "genius"
project-name: "genius"
- odl-maven-jobs-jdk11
- odl-maven-verify-jobs
- csit-gate-list: "netvirt-csit-1node-0cmb-1ctl-2cmp-apex-queens-gate-snat-conntrack-{stream}"
- block_on_csit: false
-
stream: aluminium
project: "genius"
project-name: "genius"
- odl-maven-jobs-jdk11
- odl-maven-verify-jobs
- csit-gate-list: "netvirt-csit-1node-0cmb-1ctl-2cmp-apex-queens-gate-snat-conntrack-{stream}"
- block_on_csit: false
-
stream: magnesium
project: "genius"
project-name: "genius"
- odl-maven-jobs
- odl-maven-verify-jobs
- csit-gate-list: "netvirt-csit-1node-0cmb-1ctl-2cmp-apex-queens-gate-snat-conntrack-{stream}"
- block_on_csit: false
-
stream: sodium
project: "genius"
project-name: "genius"
# Shell Scripts #
#################
-# Macro: integration-get-apex-addresses
-# Operation: this macro gets the IP addresses of the dynamic vms from apex snapshots
-# Used by: {project}-csit-* apex related job templates
-
-- builder:
- name: integration-get-apex-addresses
- builders:
- - shell: !include-raw: integration-get-apex-addresses.sh
-
-# Macro: integration-apex-run-tests
-# Operation: this macro just runs the robot suites
-# Used by: {project}-csit-* apex related job templates
-
-- builder:
- name: integration-apex-run-tests
- builders:
- - shell: !include-raw: integration-apex-run-tests.sh
-
# Macro: integration-get-slave-addresses
# Operation: this macro gets the IP addresses of the dynamic vms
# Used by: {project}-csit-* job templates
- shell: !include-raw:
- integration-deploy-openstack-run-test.sh
-- builder:
- name: integration-apex-image-manager
- builders:
- - shell: !include-raw:
- - ../opendaylight-infra-apex-image-management.sh
-
- builder:
name: integration-compare-distributions
# Operation: will compare an already created distribtion (probably last
- email-notification:
email-recipients: "{email-recipients}"
email-prefix: "[{project}]"
-
-#####################
-# inttest-csit-apex #
-#####################
-
-- inttest_csit_apex: &inttest_csit_apex
- name: inttest-csit-apex
-
- #####################
- # Default variables #
- #####################
-
- openstack-cloud: vex
- openstack-heat-template: csit-apex-3-instance-type.yaml
- openstack-heat-template-dir: openstack-hot
- controller-max-mem: "2048m"
-
- #####################
- # Job configuration #
- #####################
-
- parameters:
- - integration-csit-parameters:
- archive-artifacts: "{archive-artifacts}"
- branch: "{branch}"
- os-cloud: "{os-cloud}"
- # CSIT
- bundle-url: "{bundle-url}"
- repo-url: "{repo-url}"
- robot-options: "{robot-options}"
- # Controller
- controller-scope: ""
- controller-max-mem: "{controller-max-mem}"
- debug-map: "{debug-map}"
- install-features: "{install-features}"
- use-features-boot: "{use-features-boot}"
- elasticsearch-attribute: "{elasticsearch-attribute}"
- - integration-stack-3-type:
- vm_0_count: "{apex_controller_count}"
- vm_0_flavor: "{apex_controller_flavor}"
- vm_0_image: "{apex_controller_image}"
- vm_1_count: "1"
- vm_1_flavor: "{apex_compute_0_flavor}"
- vm_1_image: "{apex_compute_0_image}"
- vm_2_count: "1"
- vm_2_flavor: "{apex_compute_1_flavor}"
- vm_2_image: "{apex_compute_1_image}"
- - integration-distribution-branch:
- branch: "{branch}"
- - integration-distribution-stream:
- stream: "{stream}"
- - integration-test-plan:
- test-plan: "{project}-{functionality}.txt"
- - integration-test-suites:
- test-suites: "{test-suites}"
- - integration-patch-refspec:
- branch: "master"
- - integration-jdk-version:
- jdkversion: "{jre}"
- - distribution-karaf-version:
- karaf-version: "{karaf-version}"
- - string:
- name: OPENSTACK_BRANCH
- default: "{openstack-branch}"
- description: "Openstack version to use with devstack"
- - string:
- name: IPSEC_VXLAN_TUNNELS_ENABLED
- default: "{ipsec-vxlan-tunnels-enabled}"
- description: "Enable ipsec over vxlan support for all controllers and computes"
- - string:
- name: PUBLIC_BRIDGE
- default: "{public-bridge}"
- description: "Specifies public bridge for provider and external networking"
- - string:
- name: SECURITY_GROUP_MODE
- default: "{security-group-mode}"
- description: "Security Group Mode to be used in netvirt aclservice config"
- - string:
- name: ENABLE_ITM_DIRECT_TUNNELS
- default: "{enable-itm-direct-tunnels}"
- description: "Tunnel Configuration mode to be used in genius ifm config"
- - string:
- name: PUBLIC_PHYSICAL_NETWORK
- default: "{public-physical-network}"
- description: "provider physical network used for flat/vlan networks"
- - string:
- name: ODL_SNAT_MODE
- default: "{odl-snat-mode}"
- description: "The SNAT mode to be configured - options are conntrack/controller."
- - string:
- name: GROUP_ADD_MOD_ENABLED
- default: "{group-add-mod-enabled}"
- description: "The Group add mod enabled mode to be configured - options are true/false."
-
- scm:
- - integration-gerrit-scm:
- basedir: "test"
- refspec: "$PATCHREFSPEC"
- branch: "master"
-
- triggers:
- - timed: "{schedule}"
-
- builders:
- - lf-infra-pre-build
- - integration-stack-3-type:
- openstack-cloud: "{openstack-cloud}"
- openstack-heat-template: "{openstack-heat-template}"
- openstack-heat-template-dir: "{openstack-heat-template-dir}"
- - integration-install-robotframework
- - integration-detect-variables
- - integration-install-common-functions
- - inject:
- properties-file: "env.properties"
- - integration-get-apex-addresses
- - inject:
- properties-file: "slave_addresses.txt"
- - integration-apex-run-tests
- - integration-cleanup-tmp
-
- publishers:
- - email-notification:
- email-recipients: "{email-recipients}"
- email-prefix: "[{project}]"
- - integration-csit-publish:
- robot_other_files:
- - tempest_results.html
- robot-pass-threshold: 100.0
- robot-unstable-threshold: 0.0
- - lf-stack-delete:
- openstack-cloud: "{openstack-cloud}"
- - lf-infra-publish
-
-- job-template:
- # yamllint disable-line rule:line-length
- name: "{prefix}{project}-csit-{topology}-{os-cmb-cnt}cmb-{os-ctl-cnt}ctl-{os-cmp-cnt}cmp-apex-{openstack}-{functionality}-{stream}"
- id: inttest-csit-apex
- <<: *inttest_csit_common
- # yamllint disable-line rule:key-duplicates
- <<: *inttest_csit_apex
-
-- job-template:
- # yamllint disable-line rule:line-length
- name: "{prefix}{project}-csit-{topology}-{os-cmb-cnt}cmb-{os-ctl-cnt}ctl-{os-cmp-cnt}cmp-apex-{openstack}-{functionality}-{stream}"
- id: inttest-csit-apex-custom
- <<: *inttest_csit_common
- # yamllint disable-line rule:key-duplicates
- <<: *inttest_csit_apex
-
- parameters:
- - integration-csit-parameters:
- archive-artifacts: "{archive-artifacts}"
- branch: "{branch}"
- os-cloud: "{os-cloud}"
- # CSIT
- bundle-url: "{bundle-url}"
- repo-url: "{repo-url}"
- robot-options: "{robot-options}"
- # Controller
- controller-scope: ""
- controller-max-mem: "{controller-max-mem}"
- debug-map: "{debug-map}"
- install-features: "{install-features}"
- use-features-boot: "{use-features-boot}"
- elasticsearch-attribute: "{elasticsearch-attribute}"
- - integration-stack-3-type:
- vm_0_count: "{apex_controller_count}"
- vm_0_flavor: "{apex_controller_flavor}"
- vm_0_image: "{apex_controller_image}"
- vm_1_count: "1"
- vm_1_flavor: "{apex_compute_0_flavor}"
- vm_1_image: "{apex_compute_0_image}"
- vm_2_count: "1"
- vm_2_flavor: "{apex_compute_1_flavor}"
- vm_2_image: "{apex_compute_1_image}"
- - integration-distribution-branch:
- branch: "{branch}"
- - integration-distribution-stream:
- stream: "{stream}"
- - integration-test-plan:
- test-plan: "{testplan}"
- - integration-test-suites:
- test-suites: "{test-suites}"
- - integration-patch-refspec:
- branch: "master"
- - integration-jdk-version:
- jdkversion: "{jre}"
- - distribution-karaf-version:
- karaf-version: "{karaf-version}"
- - string:
- name: OPENSTACK_BRANCH
- default: "{openstack-branch}"
- description: "Openstack version to use with devstack"
- - string:
- name: IPSEC_VXLAN_TUNNELS_ENABLED
- default: "{ipsec-vxlan-tunnels-enabled}"
- description: "Enable ipsec over vxlan support for all controllers and computes"
- - string:
- name: PUBLIC_BRIDGE
- default: "{public-bridge}"
- description: "Specifies public bridge for provider and external networking"
- - string:
- name: SECURITY_GROUP_MODE
- default: "{security-group-mode}"
- description: "Security Group Mode to be used in netvirt aclservice config"
- - string:
- name: ENABLE_ITM_DIRECT_TUNNELS
- default: "{enable-itm-direct-tunnels}"
- description: "Tunnel Configuration mode to be used in genius ifm config"
- - string:
- name: PUBLIC_PHYSICAL_NETWORK
- default: "{public-physical-network}"
- description: "provider physical network used for flat/vlan networks"
- - string:
- name: ODL_SNAT_MODE
- default: "{odl-snat-mode}"
- description: "The SNAT mode to be configured - options are conntrack/controller."
- - string:
- name: GROUP_ADD_MOD_ENABLED
- default: "{group-add-mod-enabled}"
- description: "The Group add mod enabled mode to be configured - options are true/false."
+++ /dev/null
----
-- project:
- name: netvirt-apex-image-management
- jobs:
- - "netvirt-apex-image-management"
-
- project: "netvirt"
- os-cloud: vex
- disable-job: true
-
-- job-template:
- name: "netvirt-apex-image-management"
- disable-job: false
-
- node: "centos7-apex-2c-2g"
- disabled: "{disable-job}"
-
- parameters:
- - string:
- name: OS_CLOUD
- default: "{os-cloud}"
- description: |
- The name of a cloud configuration in clouds.yaml. OS_CLOUD is a
- variable name that is significant to openstack client as a
- environment variable. Please refer to the documentation for
- further details.
- https://docs.openstack.org/developer/python-openstackclient/
-
- wrappers:
- # Listed after to override openstack-infra-wrappers clouds.yaml definition
- - config-file-provider:
- files:
- - file-id: clouds-yaml
- target: "$HOME/.config/openstack/clouds.yaml"
-
- builders:
- - integration-apex-image-manager
-
- triggers:
- - timed: "@weekly"
+++ /dev/null
----
-- project:
- name: netvirt-csit-apex-integration-oxygen
- jobs:
- - inttest-csit-apex-custom
-
- disable-job: true
-
- project: "netvirt"
- stream:
- - silicon:
- branch: "master"
- jre: "openjdk11"
- - aluminium:
- branch: "stable/aluminium"
- jre: "openjdk11"
- - magnesium:
- branch: "stable/magnesium"
- jre: "openjdk11"
- - sodium:
- branch: "stable/sodium"
-
- testplan: "{project}-{topology}-openstack.txt"
-
- test-suites: "openstack/connectivity/l2.robot
- openstack/connectivity/external_network.robot"
-
- topology: "1node"
-
- install-features: "odl-netvirt-openstack, odl-mdsal-trace"
-
- public-bridge: "br-datacentre"
- public-physical-network: "datacentre"
- odl-snat-mode: "conntrack"
-
- apex_controller_count: "1"
- apex_controller_flavor: "v1-standard-8"
- apex_controller_image: "ZZCI - OPNFV - apex - controller - 0"
-
- apex_compute_0_flavor: "v1-standard-4"
- apex_compute_0_image: "ZZCI - OPNFV - apex - compute - 0"
-
- apex_compute_1_flavor: "v1-standard-4"
- apex_compute_1_image: "ZZCI - OPNFV - apex - compute - 1"
-
- os-cmb-cnt: 0
- os-ctl-cnt: 1
- os-cmp-cnt: 2
-
- functionality:
- - upstream-snat-conntrack
- - gate-snat-conntrack:
- gate-job: "True"
-
- openstack:
- - queens:
- openstack-branch: "stable/queens"
- odl-maven-jobs-jdk11
- odl-maven-verify-jobs
- csit-gate-list: "netvirt-csit-1node-0cmb-1ctl-2cmp-apex-queens-gate-snat-conntrack-{stream}"
- block_on_csit: false
-
stream: silicon
branch: "master"
- odl-maven-jobs-jdk11
- odl-maven-verify-jobs
- csit-gate-list: "netvirt-csit-1node-0cmb-1ctl-2cmp-apex-queens-gate-snat-conntrack-{stream}"
- block_on_csit: false
-
stream: aluminium
branch: "stable/aluminium"
- odl-maven-jobs-jdk11
- odl-maven-verify-jobs
- csit-gate-list: "netvirt-csit-1node-0cmb-1ctl-2cmp-apex-queens-gate-snat-conntrack-{stream}"
- block_on_csit: false
-
stream: magnesium
branch: "stable/magnesium"
- odl-maven-jobs
- odl-maven-verify-jobs
- csit-gate-list: "netvirt-csit-1node-0cmb-1ctl-2cmp-apex-queens-gate-snat-conntrack-{stream}"
- block_on_csit: false
-
stream: sodium
branch: "stable/sodium"
+++ /dev/null
-#!/bin/bash
-# SPDX-License-Identifier: EPL-1.0
-##############################################################################
-# Copyright (c) 2018 Red Hat, Inc. and others.
-#
-# All rights reserved. This program and the accompanying materials
-# are made available under the terms of the Eclipse Public License v1.0
-# which accompanies this distribution, and is available at
-# http://www.eclipse.org/legal/epl-v10.html
-##############################################################################
-
-# Ensure we fail the job if any steps fail.
-set -x -o pipefail -o errexit
-
-virtualenv "/tmp/v/openstack"
-# shellcheck source=/tmp/v/openstack/bin/activate disable=SC1091
-source "/tmp/v/openstack/bin/activate"
-pip install --upgrade --quiet "pip<10.0.0" setuptools
-pip install --upgrade --quiet python-openstackclient
-pip freeze
-
-df -h
-
-mkdir -p /tmp/apex_snapshots
-pushd /tmp/apex_snapshots
-
-wget artifacts.opnfv.org/apex/queens/noha/snapshot.properties
-cat snapshot.properties
-# shellcheck disable=SC1091
-source snapshot.properties
-SNAPSHOT_FILENAME=$(basename "$OPNFV_SNAP_URL")
-
-wget --progress=dot:giga "$OPNFV_SNAP_URL"
-gunzip -c "$SNAPSHOT_FILENAME" > snapshots.tar
-
-# builder VMs don't have enough disk to handle a full un-tarring, so doing one
-# big file at a time and deleting it from the tarball as a workaround for now
-tar -tf snapshots.tar
-
-images=$(tar --list -f snapshots.tar | grep qcow2)
-for image in $images; do
- tar -xf snapshots.tar "$image"
- tar --delete --file=snapshots.tar "$image"
-done
-
-# get the ssh keys and node.yaml for uploading to swift at the end
-tar -xf snapshots.tar ./id_rsa
-tar -xf snapshots.tar ./node.yaml
-tar -xf snapshots.tar ./overcloudrc
-
-ls -altr
-
-# grab the right baremetal# for the controller(s) and compute(s)
-CONTROLLER_NODE=$(grep -E 'type|vNode-name' node.yaml | grep -E -A1 controller | tail -n1 | awk '{print $2}')
-COMPUTE_0_NODE=$(grep -E 'type|vNode-name' node.yaml | grep -E -A1 compute | tail -n1 | awk '{print $2}')
-COMPUTE_1_NODE=$(grep -E 'type|vNode-name' node.yaml | grep -E -A1 compute | head -n2 | tail -n1 | awk '{print $2}')
-
-# Customize images to work in ODL Vexxhost infra
-sudo yum install -y libguestfs-tools
-export LIBGUESTFS_BACKEND=direct
-virt-customize -a "$CONTROLLER_NODE.qcow2" \
- --run-command "crudini --set /var/lib/config-data/puppet-generated/neutron/etc/neutron/plugins/ml2/ml2_conf.ini ml2 physical_network_mtus datacentre:1458" \
- --run-command "crudini --set /var/lib/config-data/puppet-generated/neutron/etc/neutron/plugins/ml2/ml2_conf.ini ml2 path_mtu 1458" \
- --run-command "crudini --set /var/lib/config-data/puppet-generated/neutron/etc/neutron/neutron.conf '' global_physnet_mtu 1458" \
- --run-command "crudini --set /var/lib/config-data/puppet-generated/neutron/etc/neutron/dhcp_agent.ini '' debug true" \
-
-virt-customize -a "$COMPUTE_0_NODE.qcow2" \
- --run-command "crudini --set /var/lib/config-data/puppet-generated/nova_libvirt/etc/nova/nova.conf libvirt virt_type qemu"
-
-virt-customize -a "$COMPUTE_1_NODE.qcow2" \
- --run-command "crudini --set /var/lib/config-data/puppet-generated/nova_libvirt/etc/nova/nova.conf libvirt virt_type qemu"
-
-for image in $CONTROLLER_NODE $COMPUTE_0_NODE $COMPUTE_1_NODE
-do
- # Change interface MTU to account for default network mtu of 1458
- virt-customize -a "$image.qcow2" \
- --run-command 'sudo echo "MTU=\"1458\"" >> /etc/sysconfig/network-scripts/ifcfg-eth0' \
- --run-command 'sudo echo "MTU=\"1458\"" >> /etc/sysconfig/network-scripts/ifcfg-br-int' \
- --run-command 'sudo echo "MTU=\"1458\"" >> /etc/sysconfig/network-scripts/ifcfg-ovs-system' \
- --run-command "sudo crudini --set /etc/selinux/config '' SELINUX permissive"
-done
-
-popd
-
-openstack image list
-
-# clean out any zombie OPNFV - apex images that *may* be left over from troubled jobs
-openstack image list | grep -E 'OPNFV - apex.*new ' | awk '{print "openstack image delete",$2}' | sh || true
-
-qemu-img convert -f qcow2 -O raw "/tmp/apex_snapshots/$CONTROLLER_NODE.qcow2" "/tmp/apex_snapshots/$CONTROLLER_NODE.raw"
-rm "/tmp/apex_snapshots/$CONTROLLER_NODE.qcow2"
-qemu-img convert -f qcow2 -O raw "/tmp/apex_snapshots/$COMPUTE_0_NODE.qcow2" "/tmp/apex_snapshots/$COMPUTE_0_NODE.raw"
-rm "/tmp/apex_snapshots/$COMPUTE_0_NODE.qcow2"
-qemu-img convert -f qcow2 -O raw "/tmp/apex_snapshots/$COMPUTE_1_NODE.qcow2" "/tmp/apex_snapshots/$COMPUTE_1_NODE.raw"
-rm "/tmp/apex_snapshots/$COMPUTE_1_NODE.qcow2"
-
-# create .new images first, then we can delete the existing and rename .new
-# to existing to reduce the delta of when these images might be unavailable
-CONTROLLER_IMAGE_NAME="ZZCI - OPNFV - apex - controller - 0"
-COMPUTE_0_IMAGE_NAME="ZZCI - OPNFV - apex - compute - 0"
-COMPUTE_1_IMAGE_NAME="ZZCI - OPNFV - apex - compute - 1"
-
-openstack image create \
- --disk-format raw --container-format bare \
- --file "/tmp/apex_snapshots/$CONTROLLER_NODE.raw" "$CONTROLLER_IMAGE_NAME.new"
-openstack image create \
- --disk-format raw --container-format bare \
- --file "/tmp/apex_snapshots/$COMPUTE_0_NODE.raw" "$COMPUTE_0_IMAGE_NAME.new"
-openstack image create \
- --disk-format raw --container-format bare \
- --file "/tmp/apex_snapshots/$COMPUTE_1_NODE.raw" "$COMPUTE_1_IMAGE_NAME.new"
-
-# clean out any non ".new" OPNFV - apex images. In the case of a previously failed
-# or aborted apex management job, we can end up with multiple images with the same
-# name so being thorough here.
-openstack image list | grep -E 'OPNFV - apex' | grep -Ev 'new' | awk '{print "openstack image delete",$2}' | sh
-
-openstack image set --name "$CONTROLLER_IMAGE_NAME" "$CONTROLLER_IMAGE_NAME.new"
-openstack image set --tag "Date Uploaded: $(date)" "$CONTROLLER_IMAGE_NAME"
-openstack image set --tag "Apex Archive: $(basename "$OPNFV_SNAP_URL")" "$CONTROLLER_IMAGE_NAME"
-
-openstack image set --name "$COMPUTE_0_IMAGE_NAME" "$COMPUTE_0_IMAGE_NAME.new"
-openstack image set --tag "Date Uploaded: $(date)" "$COMPUTE_0_IMAGE_NAME"
-openstack image set --tag "Apex Archive: $(basename "$OPNFV_SNAP_URL")" "$COMPUTE_0_IMAGE_NAME"
-
-openstack image set --name "$COMPUTE_1_IMAGE_NAME" "$COMPUTE_1_IMAGE_NAME.new"
-openstack image set --tag "Date Uploaded: $(date)" "$COMPUTE_1_IMAGE_NAME"
-openstack image set --tag "Apex Archive: $(basename "$OPNFV_SNAP_URL")" "$COMPUTE_1_IMAGE_NAME"
-
-# Now that the images should be up, active and ready, we can update
-# the ssh key and node.yaml in swift
-openstack container create OPNFV-APEX-SNAPSHOTS
-openstack object create OPNFV-APEX-SNAPSHOTS /tmp/apex_snapshots/node.yaml --name node.yaml
-openstack object create OPNFV-APEX-SNAPSHOTS /tmp/apex_snapshots/id_rsa --name id_rsa
-openstack object create OPNFV-APEX-SNAPSHOTS /tmp/apex_snapshots/overcloudrc --name overcloudrc
-openstack object list OPNFV-APEX-SNAPSHOTS
-
-openstack image list
-
-df -h
-
- odl-maven-jobs-jdk11
- odl-maven-verify-jobs
- csit-gate-list: "netvirt-csit-1node-0cmb-1ctl-2cmp-apex-queens-gate-snat-conntrack-{stream}"
- block_on_csit: false
-
stream: silicon
project: "openflowplugin"
project-name: "openflowplugin"
- odl-maven-jobs-jdk11
- odl-maven-verify-jobs
- csit-gate-list: "netvirt-csit-1node-0cmb-1ctl-2cmp-apex-queens-gate-snat-conntrack-{stream}"
- block_on_csit: false
-
stream: aluminium
project: "openflowplugin"
project-name: "openflowplugin"
- odl-maven-jobs-jdk11
- odl-maven-verify-jobs
- csit-gate-list: "netvirt-csit-1node-0cmb-1ctl-2cmp-apex-queens-gate-snat-conntrack-{stream}"
- block_on_csit: false
-
stream: magnesium
project: "openflowplugin"
project-name: "openflowplugin"
- odl-maven-jobs
- odl-maven-verify-jobs
- csit-gate-list: "netvirt-csit-1node-0cmb-1ctl-2cmp-apex-queens-gate-snat-conntrack-{stream}"
- block_on_csit: false
-
stream: sodium
project: "openflowplugin"
project-name: "openflowplugin"
- odl-maven-jobs-jdk11
- odl-maven-verify-jobs
- csit-gate-list: "netvirt-csit-1node-0cmb-1ctl-2cmp-apex-queens-gate-snat-conntrack-{stream}"
- block_on_csit: false
-
stream: silicon
project: "ovsdb"
project-name: "ovsdb"
- odl-maven-jobs-jdk11
- odl-maven-verify-jobs
- csit-gate-list: "netvirt-csit-1node-0cmb-1ctl-2cmp-apex-queens-gate-snat-conntrack-{stream}"
- block_on_csit: false
-
stream: aluminium
project: "ovsdb"
project-name: "ovsdb"
- odl-maven-jobs-jdk11
- odl-maven-verify-jobs
- csit-gate-list: "netvirt-csit-1node-0cmb-1ctl-2cmp-apex-queens-gate-snat-conntrack-{stream}"
- block_on_csit: false
-
stream: magnesium
project: "ovsdb"
project-name: "ovsdb"
- odl-maven-jobs
- odl-maven-verify-jobs
- csit-gate-list: "netvirt-csit-1node-0cmb-1ctl-2cmp-apex-queens-gate-snat-conntrack-{stream}"
- block_on_csit: false
-
stream: sodium
project: "ovsdb"
project-name: "ovsdb"
- shell: !include-raw: opendaylight-infra-stack.sh
- shell: !include-raw-escape: opendaylight-infra-copy-ssh-keys.sh
-- builder:
- name: opendaylight-apex-stack
- # opendaylight-apex-stack.sh has a required variable {stack-template} that
- # must be passed into this macro.
- builders:
- - shell: !include-raw: opendaylight-infra-stack.sh
-
- builder:
name: distribute-build-url
# Place URL of the current run of a build job to a file at given path.
+++ /dev/null
----
-heat_template_version: 2016-04-08
-
-parameters:
- job_name:
- type: string
- description: Name of job running this template
- silo:
- type: string
- default: vm
- description: |
- String to identify a Jenkins Master
-
- Useful in scenarios where multiple Jenkins Masters might be
- creating VMs in the same cloud to uniquely distinguish them.
-
- TODO we should move to using tags once OpenStack Ocata which is
- available to us with support for server tags properties.
- index:
- type: number
- group:
- type: number
- default: "0"
- vm_flavor:
- type: string
- default: v1-standard-4
- description: OpenStack Flavor to use
- vm_image:
- type: string
- description: VM Image to spin up
- vm_network:
- type: string
- description: VM network
-
-resources:
- instance:
- type: "OS::Nova::Server"
- properties:
- image: { get_param: vm_image }
- flavor: { get_param: vm_flavor }
- name:
- str_replace:
- template: SILO-JOB_NAME-GROUP-VM_TYPE-INDEX
- params:
- "SILO": { get_param: silo }
- "JOB_NAME": { get_param: job_name }
- "VM_TYPE":
- str_split:
- - " - "
- - { get_param: vm_image }
- - 2
- "GROUP": { get_param: group }
- "INDEX": { get_param: index }
- networks:
- - network: { get_param: vm_network }
+++ /dev/null
----
-heat_template_version: 2016-04-08
-
-parameters:
- job_name:
- type: string
- description: Name of job running this template
- silo:
- type: string
- description: Jenkins Silo job is running in (typically releng|sandbox)
- vm_0_count:
- type: number
- description: Number of VMs for 1st VM type
- vm_0_flavor:
- type: string
- default: v1-standard-1
- description: Flavor to run instance on for 1st VM type
- vm_0_image:
- type: string
- description: Image to run instance on for 1st VM type
- vm_1_count:
- type: number
- description: Number of VMs for 2nd VM type
- vm_1_flavor:
- type: string
- default: v1-standard-1
- description: Flavor to run instance on for 2nd VM type
- vm_1_image:
- type: string
- description: Image to run instance on for 2nd VM type
- vm_2_count:
- type: number
- description: Number of VMs for 3rd VM type
- vm_2_flavor:
- type: string
- default: v1-standard-1
- description: Flavor to run instance on for 3rd VM type
- vm_2_image:
- type: string
- description: Image to run instance on for 3rd VM type
-
-resources:
- 192_network:
- type: OS::Neutron::Net
- properties:
- name:
- str_replace:
- template: SILO-JOB_NAME-NET
- params:
- "SILO": { get_param: silo }
- "JOB_NAME": { get_param: job_name }
- "NET": "APEX_192_network"
- port_security_enabled: false
-
- 192_subnet:
- type: OS::Neutron::Subnet
- properties:
- name:
- str_replace:
- template: SILO-JOB_NAME-SUBNET
- params:
- "SILO": { get_param: silo }
- "JOB_NAME": { get_param: job_name }
- "SUBNET": "APEX_192_subnet"
- network_id: { get_resource: 192_network }
- cidr: 192.0.2.0/24
- gateway_ip: 192.0.2.1
- allocation_pools:
- - start: 192.0.2.51
- end: 192.0.2.99
-
- vm_0_group:
- type: "OS::Heat::ResourceGroup"
- properties:
- count: { get_param: vm_0_count }
- resource_def:
- type: apex-server.yaml
- properties:
- job_name: { get_param: job_name }
- silo: { get_param: silo }
- index: "%index%"
- group: "0"
- vm_flavor: { get_param: vm_0_flavor }
- vm_image: { get_param: vm_0_image }
- vm_network: { get_resource: 192_network }
-
- vm_1_group:
- type: "OS::Heat::ResourceGroup"
- properties:
- count: { get_param: vm_1_count }
- resource_def:
- type: apex-server.yaml
- properties:
- job_name: { get_param: job_name }
- silo: { get_param: silo }
- index: "%index%"
- group: "1"
- vm_flavor: { get_param: vm_1_flavor }
- vm_image: { get_param: vm_1_image }
- vm_network: { get_resource: 192_network }
-
- vm_2_group:
- type: "OS::Heat::ResourceGroup"
- properties:
- count: { get_param: vm_2_count }
- resource_def:
- type: apex-server.yaml
- properties:
- job_name: { get_param: job_name }
- silo: { get_param: silo }
- index: "%index%"
- group: "2"
- vm_flavor: { get_param: vm_2_flavor }
- vm_image: { get_param: vm_2_image }
- vm_network: { get_resource: 192_network }