string(defaultValue: '', description: '', name: 'UPSTREAM_JOB_NUMBER'),
string(defaultValue: 'dpkg1', description: '', name: 'GPG_KEY_NAME'),
string(defaultValue: 'artifactory-osm', description: '', name: 'ARTIFACTORY_SERVER'),
+ string(defaultValue: 'osm-stage_4', description: '', name: 'DOWNSTREAM_STAGE_NAME'),
booleanParam(defaultValue: false, description: '', name: 'SAVE_CONTAINER_ON_FAIL'),
booleanParam(defaultValue: false, description: '', name: 'SAVE_CONTAINER_ON_PASS'),
+ booleanParam(defaultValue: false, description: '', name: 'DO_STAGE_4'),
+ booleanParam(defaultValue: true, description: '', name: 'DO_BUILD'),
+ booleanParam(defaultValue: false, description: '', name: 'DO_INSTALL'),
+ booleanParam(defaultValue: false, description: '', name: 'DO_SMOKE'),
+ booleanParam(defaultValue: false, description: '', name: 'SAVE_ARTIFACTS_OVERRIDE'),
])
])
+def uninstall_osm(stackName) {
+ sh """
+ export OSM_USE_LOCAL_DEVOPS=true
+ export PATH=$PATH:/snap/bin
+ installers/full_install_osm.sh -y -s ${stackName} --test --nolxd --nodocker --nojuju --nohostports --nohostclient --uninstall
+ """
+}
+
node("${params.NODE}") {
sh 'env'
tag_or_branch = params.GERRIT_BRANCH.replaceAll(/\./,"")
- container_name_prefix = "osm-${tag_or_branch}"
- container_name = "${container_name_prefix}-${BUILD_NUMBER}"
stage("Checkout") {
checkout scm
ci_helper = load "jenkins/ci-pipelines/ci_helper.groovy"
+ def upstream_main_job = params.UPSTREAM_SUFFIX
+
+ // upstream jobs always use merged artifacts
+ upstream_main_job += '-merge'
+ container_name_prefix = "osm-${tag_or_branch}"
+ container_name = "${container_name_prefix}"
+ if ( JOB_NAME.contains('merge') ) {
+ container_name += "-merge"
+ }
+ container_name += "-${BUILD_NUMBER}"
+
// Copy the artifacts from the upstream jobs
stage("Copy Artifacts") {
// cleanup any previous repo
sh 'rm -rf repo'
- if ( params.UPSTREAM_SUFFIX ) {
-
- dir("repo") {
- // grab all stable upstream builds based on the
- // given target UPSTREAM_SUFFIX
-
- dir("${RELEASE}") {
- def list = ["SO", "UI", "RO", "openvim", "osmclient"]
- for (component in list) {
- step ([$class: 'CopyArtifact',
- projectName: "${component}${params.UPSTREAM_SUFFIX}/${GERRIT_BRANCH}"])
-
- // grab the build name/number
- //options = get_env_from_build('build.env')
- build_num = ci_helper.get_env_value('build.env','BUILD_NUMBER')
- //build_num = sh(returnStdout:true, script: "cat build.env | awk -F= '/BUILD_NUMBER/{print \$2}'").trim()
- ci_helper.get_archive(params.ARTIFACTORY_SERVER,component,GERRIT_BRANCH, "${component}-stage_2 :: ${GERRIT_BRANCH}", build_num)
-
- // cleanup any prevously defined dists
- sh "rm -rf dists"
- }
-
- // check if an upstream artifact based on specific build number has been requested
- // This is the case of a merge build and the upstream merge build is not yet complete (it is not deemed
- // a successful build yet). The upstream job is calling this downstream job (with the its build artifiact)
- if ( params.UPSTREAM_JOB_NAME ) {
- step ([$class: 'CopyArtifact',
- projectName: "${params.UPSTREAM_JOB_NAME}",
- selector: [$class: 'SpecificBuildSelector', buildNumber: "${params.UPSTREAM_JOB_NUMBER}"]
- ])
-
- //options = get_env_from_build('build.env')
- // grab the build name/number
- //build_num = sh(returnStdout:true, script: "cat build.env | awk -F= '/BUILD_NUMBER/{print \$2}'").trim()
- build_num = ci_helper.get_env_value('build.env','BUILD_NUMBER')
- component = ci_helper.get_mdg_from_project(ci_helper.get_env_value('build.env','GERRIT_PROJECT'))
-
- ci_helper.get_archive(params.ARTIFACTORY_SERVER,component,GERRIT_BRANCH, "${component}-stage_2 :: ${GERRIT_BRANCH}", build_num)
-
- sh "rm -rf dists"
- }
-
- // sign all the components
- for (component in list) {
- sh "dpkg-sig --sign builder -k ${GPG_KEY_NAME} pool/${component}/*"
- }
-
- // now create the distro
- for (component in list) {
- sh "mkdir -p dists/${params.REPO_DISTRO}/${component}/binary-amd64/"
- sh "apt-ftparchive packages pool/${component} > dists/${params.REPO_DISTRO}/${component}/binary-amd64/Packages"
- sh "gzip -9fk dists/${params.REPO_DISTRO}/${component}/binary-amd64/Packages"
- }
-
- // create and sign the release file
- sh "apt-ftparchive release dists/${params.REPO_DISTRO} > dists/${params.REPO_DISTRO}/Release"
- sh "gpg --yes -abs -u ${GPG_KEY_NAME} -o dists/${params.REPO_DISTRO}/Release.gpg dists/${params.REPO_DISTRO}/Release"
-
- // copy the public key into the release folder
- // this pulls the key from the home dir of the current user (jenkins)
- sh "cp ~/${REPO_KEY_NAME} ."
+ dir("repo") {
+ // grab all stable upstream builds based on the
+
+ dir("${RELEASE}") {
+ def list = ["RO", "openvim", "osmclient", "IM", "devops", "MON", "N2VC", "NBI", "common", "LCM"]
+ for (component in list) {
+ step ([$class: 'CopyArtifact',
+ projectName: "${component}${upstream_main_job}/${GERRIT_BRANCH}"])
+
+ // grab the build name/number
+ //options = get_env_from_build('build.env')
+ build_num = ci_helper.get_env_value('build.env','BUILD_NUMBER')
+
+ // grab the archives from the stage_2 builds (ie. this will be the artifacts stored based on a merge)
+ ci_helper.get_archive(params.ARTIFACTORY_SERVER,component,GERRIT_BRANCH, "${component}${upstream_main_job} :: ${GERRIT_BRANCH}", build_num)
+
+ // cleanup any prevously defined dists
+ sh "rm -rf dists"
}
- // start an apache server to serve up the images
- http_server_name = "${container_name}-apache"
- pwd = sh(returnStdout:true, script: 'pwd').trim()
- repo_base_url = ci_helper.start_http_server(pwd,http_server_name)
+ // check if an upstream artifact based on specific build number has been requested
+ // This is the case of a merge build and the upstream merge build is not yet complete (it is not deemed
+ // a successful build yet). The upstream job is calling this downstream job (with the its build artifiact)
+ if ( params.UPSTREAM_JOB_NAME ) {
+ step ([$class: 'CopyArtifact',
+ projectName: "${params.UPSTREAM_JOB_NAME}",
+ selector: [$class: 'SpecificBuildSelector', buildNumber: "${params.UPSTREAM_JOB_NUMBER}"]
+ ])
+
+ //options = get_env_from_build('build.env')
+ // grab the build name/number
+ //build_num = sh(returnStdout:true, script: "cat build.env | awk -F= '/BUILD_NUMBER/{print \$2}'").trim()
+ build_num = ci_helper.get_env_value('build.env','BUILD_NUMBER')
+ component = ci_helper.get_mdg_from_project(ci_helper.get_env_value('build.env','GERRIT_PROJECT'))
+
+ // the upstream job name contains suffix with the project. Need this stripped off
+ def project_without_branch = params.UPSTREAM_JOB_NAME.split('/')[0]
+
+ ci_helper.get_archive(params.ARTIFACTORY_SERVER,component,GERRIT_BRANCH, "${project_without_branch} :: ${GERRIT_BRANCH}", build_num)
+
+ sh "rm -rf dists"
+ }
+
+ // sign all the components
+ for (component in list) {
+ sh "dpkg-sig --sign builder -k ${GPG_KEY_NAME} pool/${component}/*"
+ }
+
+ // now create the distro
+ for (component in list) {
+ sh "mkdir -p dists/${params.REPO_DISTRO}/${component}/binary-amd64/"
+ sh "apt-ftparchive packages pool/${component} > dists/${params.REPO_DISTRO}/${component}/binary-amd64/Packages"
+ sh "gzip -9fk dists/${params.REPO_DISTRO}/${component}/binary-amd64/Packages"
+ }
+
+ // create and sign the release file
+ sh "apt-ftparchive release dists/${params.REPO_DISTRO} > dists/${params.REPO_DISTRO}/Release"
+ sh "gpg --yes -abs -u ${GPG_KEY_NAME} -o dists/${params.REPO_DISTRO}/Release.gpg dists/${params.REPO_DISTRO}/Release"
+
+ // copy the public key into the release folder
+ // this pulls the key from the home dir of the current user (jenkins)
+ sh "cp ~/${REPO_KEY_NAME} ."
+
+ // merge the change logs
+ sh """
+ rm -f changelog/changelog-osm.html
+ [ ! -d changelog ] || for mdgchange in \$(ls changelog); do cat changelog/\$mdgchange >> changelog/changelog-osm.html; done
+ """
}
+ // start an apache server to serve up the images
+ http_server_name = "${container_name}-apache"
+
+ pwd = sh(returnStdout:true, script: 'pwd').trim()
+ repo_base_url = ci_helper.start_http_server(pwd,http_server_name)
}
}
error = null
+ if ( params.DO_BUILD ) {
+ stage("Build") {
+ sh "make -j4 -C docker CMD_DOCKER_ARGS= TAG=${container_name} RELEASE=${params.RELEASE} REPOSITORY_BASE=${repo_base_url} REPOSITORY_KEY=${params.REPO_KEY_NAME} REPOSITORY=${params.REPO_DISTRO}"
+ }
+ }
try {
- stage("Install") {
+ if ( params.DO_INSTALL ) {
+ stage("Install") {
- //will by default always delete containers on complete
- //sh "jenkins/system/delete_old_containers.sh ${container_name_prefix}"
+ //will by default always delete containers on complete
+ //sh "jenkins/system/delete_old_containers.sh ${container_name_prefix}"
- commit_id = ''
- repo_distro = ''
- repo_key_name = ''
- release = ''
+ commit_id = ''
+ repo_distro = ''
+ repo_key_name = ''
+ release = ''
- if ( params.COMMIT_ID )
- {
- commit_id = "-b ${params.COMMIT_ID}"
- }
+ if ( params.COMMIT_ID )
+ {
+ commit_id = "-b ${params.COMMIT_ID}"
+ }
- if ( params.REPO_DISTRO )
- {
- repo_distro = "-r ${params.REPO_DISTRO}"
- }
+ if ( params.REPO_DISTRO )
+ {
+ repo_distro = "-r ${params.REPO_DISTRO}"
+ }
+
+ if ( params.REPO_KEY_NAME )
+ {
+ repo_key_name = "-k ${params.REPO_KEY_NAME}"
+ }
- if ( params.REPO_KEY_NAME )
- {
- repo_key_name = "-k ${params.REPO_KEY_NAME}"
+ if ( params.RELEASE )
+ {
+ release = "-R ${params.RELEASE}"
+ }
+
+ sh """
+ export PATH=$PATH:/snap/bin
+ installers/full_install_osm.sh -y -s ${container_name} --test --nolxd --nodocker --nojuju --nohostports --nohostclient \
+ --nodockerbuild -t ${container_name} \
+ ${commit_id} \
+ ${repo_distro} \
+ ${repo_base_url} \
+ ${repo_key_name} \
+ ${release} \
+ ${params.BUILD_FROM_SOURCE}
+ """
}
+ }
- if ( params.RELEASE )
- {
- release = "-R ${params.RELEASE}"
+ if ( params.DO_SMOKE ) {
+ stage("Smoke") {
+ ci_helper.systest_run(container_name, 'smoke')
+ junit '*.xml'
}
-
- sh """
- export OSM_USE_LOCAL_DEVOPS=true
- jenkins/host/start_build system --build-container ${container_name} \
- ${commit_id} \
- ${repo_distro} \
- ${repo_base_url} \
- ${repo_key_name} \
- ${release} \
- ${params.BUILD_FROM_SOURCE}
- """
}
- stage("Test") {
- ci_helper.systest_run(container_name, 'smoke')
- junit '*.xml'
+ stage_4_archive = false
+ if ( params.DO_STAGE_4 ) {
+ stage("stage_4") {
+ def downstream_params = [
+ string(name: 'CONTAINER_NAME', value: container_name),
+ string(name: 'NODE', value: NODE_NAME.split()[0]),
+ ]
+ stage_4_result = build job: "${params.DOWNSTREAM_STAGE_NAME}/${GERRIT_BRANCH}", parameters: downstream_params, propagate: false
+ currentBuild.result = stage_4_result.result
+
+ if ( stage_4_result.getResult().equals('SUCCESS') ) {
+ stage_4_archive = true;
+ }
+ }
}
- stage("Archive") {
- sh "echo ${container_name} > build_version.txt"
- archiveArtifacts artifacts: "build_version.txt", fingerprint: true
+ // override to save the artifacts
+ if ( params.SAVE_ARTIFACTS_OVERRIDE || stage_4_archive ) {
+ stage("Archive") {
+ sh "echo ${container_name} > build_version.txt"
+ archiveArtifacts artifacts: "build_version.txt", fingerprint: true
- // Archive the tested repo
- dir("repo/${RELEASE}") {
- ci_helper.archive(params.ARTIFACTORY_SERVER,RELEASE,GERRIT_BRANCH,'tested')
+ // Archive the tested repo
+ dir("repo/${RELEASE}") {
+ ci_helper.archive(params.ARTIFACTORY_SERVER,RELEASE,GERRIT_BRANCH,'tested')
+ }
}
}
}
catch(caughtError) {
+ println("Caught error!")
error = caughtError
currentBuild.result = 'FAILURE'
}
finally {
sh "docker stop ${http_server_name}"
+ sh "docker rm ${http_server_name}"
- if (error) {
- if ( !params.SAVE_CONTAINER_ON_FAIL ) {
- sh "lxc delete ${container_name} --force"
+ if ( params.DO_INSTALL ) {
+ if (error) {
+ if ( !params.SAVE_CONTAINER_ON_FAIL ) {
+ uninstall_osm container_name
+ }
+ throw error
}
- throw error
- }
- else {
- if ( !params.SAVE_CONTAINER_ON_PASS ) {
- sh "lxc delete ${container_name} --force"
+ else {
+ if ( !params.SAVE_CONTAINER_ON_PASS ) {
+ uninstall_osm container_name
+ }
}
}
}