--- # TODO: Make all bash constants more readable. # e.g.: DISTRIBUTION_BRANCH instead of DISTROBRANCH. - parameter: name: integration-csit-parameters parameters: - opendaylight-infra-parameters: os-cloud: "{os-cloud}" project: "integration/test" branch: "{branch}" refspec: "refs/heads/{branch}" artifacts: "{archive-artifacts}" # CSIT - integration-bundle-url: bundle-url: "{bundle-url}" - integration-repo-url: repo-url: "{repo-url}" - integration-test-options: test-options: "{robot-options}" - integration-controller-parameters: controller-debug-map: "{debug-map}" controller-features: "{install-features}" controller-max-mem: "{controller-max-mem}" controller-scope: "{controller-scope}" controller-use-features-boot: "{use-features-boot}" feature-elasticsearch-attribute: "{elasticsearch-attribute}" - distribution-is-karaf-appl: karaf-based-controller: "{karaf-based-controller}" - distribution-karaf-version: karaf-version: "{karaf-version}" - parameter: name: integration-controller-parameters parameters: - string: name: CONTROLLERSCOPE default: "{controller-scope}" description: "Feature Only or Feature All test" - string: name: CONTROLLERFEATURES default: "{controller-features}" description: | Features to install in the controller separated by comma. (note: Do not include decanter* features used for harvesting JVM metrics. They are inserted in install scripts.) - string: name: CONTROLLERDEBUGMAP default: "{controller-debug-map}" description: | Key:Value map of debugs to enable on the controller separated by spaces. The module should not include org.opendaylight. The levels must be one of: ERROR, WARN, INFO, DEBUG, TRACE. (ex. ovsdb.lib:ERROR netvirt.vpnservice:INFO) - string: name: CONTROLLERMAXMEM default: "{controller-max-mem}" description: | Max Heap size to use when starting the controller. - string: name: ELASTICSEARCHATTRIBUTE default: "{feature-elasticsearch-attribute}" description: | Configure the period the script should harvest JVM metrics. Possible values: long (=120000) short (=5000) any number between 5000 and 120000 disabled (use for disabling harvesting) - string: name: USEFEATURESBOOT default: "{controller-use-features-boot}" description: "Boolean to toggle whether featuresBoot is appended or ignored" - parameter: name: integration-distribution-branch parameters: - string: name: DISTROBRANCH default: "{branch}" description: "Distribution GIT branch" - parameter: name: distribution-is-karaf-appl parameters: - string: name: IS_KARAF_APPL default: "{karaf-based-controller}" description: "Boolean to let pipeline know if the project has a different deployer" - parameter: name: distribution-karaf-version # Arguments: # karaf-version: Provide, otherwise defaults.yaml value is used. parameters: - string: name: KARAF_VERSION default: "{karaf-version}" description: "karaf3, karaf4, odl, controller, netconf affects ODL deploy scripts" - parameter: name: integration-bundle-url parameters: - string: name: BUNDLE_URL default: "{bundle-url}" description: "URL to karaf distribution zip" - parameter: name: integration-repo-url parameters: - string: name: REPO_URL default: "{repo-url}" description: "URL to feature repo artifact" - parameter: name: integration-jdk-version parameters: - string: name: JDKVERSION default: "{jdkversion}" description: "Parameter to indicate JAVA Version" - parameter: name: integration-distribution-stream parameters: - string: name: DISTROSTREAM default: "{stream}" description: "Distribution stream string, for suites to know which behavior to expect" - parameter: name: integration-script-plan parameters: - string: name: SCRIPTPLAN default: "{script-plan}" description: "Script plan we will run before downloading and un-archiving am ODL distribution" - parameter: name: integration-config-plan parameters: - string: name: CONFIGPLAN default: "{config-plan}" description: "Config plan we will run after un-archiving and starting an ODL distribution" - parameter: name: integration-stream-test-plan parameters: - string: name: STREAMTESTPLAN default: "{stream-test-plan}" description: "Stream-specific test plan we will run" - parameter: name: integration-test-plan parameters: - string: name: TESTPLAN default: "{test-plan}" description: "General test plan we will run unless stream-specific one is found" - parameter: name: integration-test-suites parameters: - string: name: SUITES default: "{test-suites}" description: | List of space-separated suites. Useful when running specific suites within a testplan. Do not include ${{WORKSPACE}}/test/csit/suites/ in the values: (ex. openstack/connectivity/l2.robot openstack/clustering/ha_l2.robot openstack/tempest/tempest.robot netvirt/elan/elan.robot). - parameter: name: integration-test-options parameters: - string: name: TESTOPTIONS default: "{test-options}" description: "Robot command options" - parameter: name: integration-distribution-git-url parameters: - string: name: DISTROGITURL default: "{git-url}/integration/distribution" description: "Distribution GIT URL (do not modify)" - parameter: name: integration-patch-refspec parameters: - string: name: PATCHREFSPEC default: "{branch}" description: "Integration Patch Refspec" # Macro: integration-gerrit-scm # Operation: this macro downloads a project gerrit # Used by: all csit jobs - scm: name: integration-gerrit-scm scm: - git: credentials-id: "opendaylight-jenkins-ssh" url: "$GIT_BASE" basedir: "{basedir}" refspec: "{refspec}" branches: - "origin/{branch}" skip-tag: true choosing-strategy: "gerrit" # Macro: integration-distribution-scm # Operation: this macro downloads the integration/distribution repo using distribution as basedir # Used by: integration-patch-test-{stream} job template - scm: name: integration-distribution-scm scm: - git: credentials-id: "opendaylight-jenkins-ssh" basedir: "distribution" url: "$DISTROGITURL" refspec: "" branches: - "origin/{branch}" skip-tag: true # Macro: integration-trigger-patch-submitted # This is a single macro to use for all csit jobs who vote on every (relevant) patch set. # Only 'recheck' trigger word is supported, it always triggers the full set of relevant jobs, # in order to prevent Jenkins from starting only a subset and still voting Verified+1. # Contrary to gerrit-trigger-patch-submitted, this votes +1 also on unstable runs. # Arguments: # gerrit-server-name: name of gerrit server to listen to # project: pattern to match triggering projects # branch: triggering branch name # files: pattern to match triggering filepaths - trigger: name: integration-trigger-patch-submitted triggers: - gerrit: server-name: "{gerrit-server-name}" trigger-on: - patchset-created-event: exclude-drafts: true exclude-trivial-rebase: false exclude-no-code-change: false - draft-published-event - comment-added-contains-event: comment-contains-value: '^Patch Set[ ]+[0-9]+:([ ]+|[\n]+)(recheck|reverify)$' override-votes: true gerrit-build-unstable-verified-value: +1 gerrit-build-unstable-codereview-value: 0 projects: - project-compare-type: ANT project-pattern: "{project}" branches: - branch-compare-type: ANT branch-pattern: "**/{branch}" file-paths: - compare-type: ANT pattern: "{files}" # Macro: integration-csit-only # Operation: this macro sets a trigger on reverse job list # Used by: csit -only- job templates - trigger: name: integration-csit-trigger triggers: - reverse: jobs: "{jobs}" result: "success" # Macro: integration-openstack-controller-mininet # Operation: this macro will spin the controller and mininet vms # Used by: {project}-csit-* job templates - wrapper: name: integration-openstack-controller-mininet wrappers: - openstack: instances: - cloud-name: "ODLRPC" template-name: "{controller-image}" count: "{controller-vms}" - cloud-name: "ODLRPC" template-name: "{mininet-image}" count: "{mininet-vms}" # Macro: integration-openstack-controller-devstack # Operation: this macro will spin the controller and devstack vms # Used by: {project}-csit-openstack job templates - wrapper: name: integration-openstack-controller-devstack wrappers: - openstack: instances: - cloud-name: "ODLRPC" template-name: "{controller-image}" count: "{controller-vms}" - cloud-name: "ODLRPC" template-name: "{openstack-image}" count: "{openstack-vms}" ################# # Shell Scripts # ################# # Macro: integration-get-slave-addresses # Operation: this macro gets the IP addresses of the dynamic vms # Used by: {project}-csit-* job templates - builder: name: integration-get-slave-addresses builders: - shell: !include-raw: integration-get-slave-addresses.sh - builder: name: integration-set-variables # Operation: Construct variables not requiring online detection and export using set_variables.env # Required bash variables: # KARAF_VERSION: 'karaf4' or 'karaf3'. Use distribution-karaf-version macro to set. builders: - shell: !include-raw: integration-set-variables.sh - inject: properties-file: set_variables.env - builder: name: integration-detect-variables # Operation: Construct variables including online detection and export using detect_variables.env # This calls integration-set-variables internally. # Required bash variables: # KARAF_VERSION: 'karaf4' or 'karaf3'. Use distribution-karaf-version macro to set. builders: - integration-set-variables - shell: !include-raw: integration-detect-variables.sh - inject: properties-file: detect_variables.env # Macro: integration-deply-controller-run-test # Operation: this macro deploys single contoller and runs test # Used by: {project}-csit-1node-* job templates - builder: name: integration-deploy-controller-run-test builders: - shell: !include-raw: integration-deploy-controller-run-test.sh # Macro: integration-deply-non-karaf-controller-run-test # Operation: this macro deploys non-karaf controller application and run test # Used by: {project}-csit-1node-* job templates - builder: name: integration-run-test builders: - shell: !include-raw: integration-run-test.sh # Macro: integration-configure-clustering # Operation: this macro configures the clustering # Used by: {project}-csit-3node-* job templates - builder: name: integration-configure-clustering builders: - shell: !include-raw: integration-configure-clustering.sh # Macro: integration-start-cluster-run-test # Operation: this macro starts the 3-node cluster and runs test # Used by: {project}-csit-3node-* job templates - builder: name: integration-start-cluster-run-test builders: - shell: !include-raw: integration-start-cluster-run-test.sh # Macro: integration-install-robotframework # Operation: Installs robotframework using pip to a virtualenv # Used by: {project}-csit-* job templates - builder: name: integration-install-robotframework builders: - shell: !include-raw: integration-install-robotframework.sh # Macro: integration-cleanup-tmp # Operation: Cleans up temporary files created by build # Used by: {project}-csit-* job templates - builder: name: integration-cleanup-tmp builders: - shell: !include-raw: integration-cleanup-tmp.sh # Macro: integration-multipatch-builder # Operation: checks out multiple patches and builds custom distribution # Used by: integration-multipatch-test-* job templates - builder: name: integration-multipatch-builder builders: - shell: !include-raw: multipatch-distribution.sh # Macro: integration-cleanup-workspace # Operation: Cleans up files possibly left there by the previous build # Used by: {project}-csit-* job templates - builder: name: integration-cleanup-workspace builders: - shell: !include-raw: integration-cleanup-workspace.sh # Macro: integration-rebase-patch # Operation: For cloned project, rebase checked-out Gerrit patch onto $DISTROBRANCH. # Used by: integration-patch-distribution-* job template - builder: name: integration-rebase-gerrit-patch builders: - shell: !include-raw: integration-rebase-gerrit-patch.sh # Macro: integration-install-common-functions # Operation: Copy the common-functions.sh script to csit nodes # Used by: {project}-csit-* job templates - builder: name: integration-install-common-functions builders: - shell: !include-raw: - copy-common-functions.sh - common-functions.sh - builder: name: integration-deploy-openstack-run-test builders: - shell: !include-raw: - integration-deploy-openstack-run-test.sh - builder: name: integration-compare-distributions # Operation: will compare an already created distribtion (probably last # published to nexus) to the current distribution being created. # The comparison is done to find the list of patches that are new to the # current distribution. # Required bash variables: # KARAF_ARTIFACT: 'karaf' or 'distribution-karaf'. Use integration-set-variables. builders: - shell: !include-raw: integration-compare-distributions.sh - builder: name: integration-upload-distribution # Upload distribution karaf artifact to Nexus SNAPSHOT repo for testing. # Arguments: # {dist-pom} path to pom file containing the version # Required bash variables: # KARAF_VERSION: 'karaf4' or 'karaf3'. # Use distribution-karaf-version macro to customize. builders: - config-file-provider: files: - file-id: global-settings variable: GLOBAL_SETTINGS_FILE - file-id: integration-settings variable: SETTINGS_FILE - integration-set-variables - inject: properties-content: BUNDLE_POM={dist-pom} - shell: !include-raw-escape: integration-upload-distribution.sh - inject: properties-file: integration-upload-distribution.env ############## # Publishers # ############## - publisher: name: integration-csit-publish publishers: - robot: output-path: "" other-files: "{obj:robot_other_files}" unstable-threshold: "{robot-unstable-threshold}" pass-threshold: "{robot-pass-threshold}" only-critical: false - integration-csit-archive-build - integration-csit-run-postscript - publisher: name: integration-csit-archive-build publishers: - postbuildscript: builders: - role: BOTH build-on: - ABORTED - FAILURE - NOT_BUILT - SUCCESS - UNSTABLE build-steps: - shell: | #!/bin/bash echo "Archiving csit artifacts" cd $WORKSPACE mkdir -p ./archives for i in `seq 1 ${NUM_ODL_SYSTEM}`; do NODE_FOLDER="./archives/odl_${i}" mkdir -p ${NODE_FOLDER} mv odl${i}_* ${NODE_FOLDER} || true mv karaf_${i}_*_threads* ${NODE_FOLDER} || true mv *_${i}.png ${NODE_FOLDER} || true mv /tmp/odl${i}_* ${NODE_FOLDER} || true mv gclogs-${i} ${NODE_FOLDER} || true done curl --output robot-plugin.zip "$BUILD_URL/robot/report/*zip*/robot-plugin.zip" unzip -d ./archives robot-plugin.zip mv *.log *.log.gz *.csv *.png ./archives || true # Don't fail if file missing # TODO: Can the following line ever fail? find . -type f -name *.hprof -print0 \ | tar -cvf - --null -T - | xz --threads=0 > ./archives/hprof.tar.xz # TODO: Tweak the compression level if better ratio (or speed) is needed. mark-unstable-if-failed: true - publisher: name: integration-csit-run-postscript publishers: - postbuildscript: builders: - role: BOTH build-on: - ABORTED - FAILURE - NOT_BUILT - SUCCESS - UNSTABLE build-steps: - config-file-provider: files: - file-id: "odl-elastic-cloud" target: "${HOME}/.netrc" - shell: | #!/bin/bash #The following script runs any configurable postplan stored in test/csit/postplans. if [ -f "${WORKSPACE}/test/csit/postplans/${TESTPLAN}" ]; then echo "postplan exists!!!" echo "Changing the postplan path..." script_name=${WORKSPACE}/test/csit/postplans/${TESTPLAN} cat ${script_name} | sed "s:integration:${WORKSPACE}:" > postplan.txt cat postplan.txt egrep -v '(^[[:space:]]*#|^[[:space:]]*$)' postplan.txt | while read -r line ; do echo "Executing ${line}..." ${line} done fi rm ${HOME}/.netrc mark-unstable-if-failed: true - publisher: name: integration-csit-collect-netvirt-logs publishers: - postbuildscript: builders: - role: BOTH build-on: - ABORTED - FAILURE - NOT_BUILT - SUCCESS - UNSTABLE build-steps: - inject: properties-file: "slave_addresses.txt" - shell: | #!/bin/bash source /tmp/common-functions.sh ${BUNDLEFOLDER} collect_netvirt_logs mark-unstable-if-failed: true