commit | bf4cfe5dffaf029a09c55a8d6e7404f511c82527 | [log] [tgz] |
---|---|---|
author | Zack Williams <zdw@opennetworking.org> | Mon May 14 16:29:57 2018 -0700 |
committer | Zack Williams <zdw@opennetworking.org> | Mon May 14 16:29:57 2018 -0700 |
tree | 3c2a7a8d4387ba2bea0a19388b3727dd42476f31 | |
parent | e6f414c0ae29148b2e59086f4ee42ee24fb8a973 [diff] |
[CORD-3038] Lint check on new docs repo Change-Id: I38e24a67088ba0df0a29fe2c237ddc36b304f1ee
This repo holds configuration for the Jenkins testing infrastructure used by CORD.
The best way to work with this repo is to check it out with repo
, per these instructions: Downloading testing and QA repositories
NOTE: This repo uses git submodules. If you have trouble with the tests or other tasks, please run:
git submodule init && git submodule update
to obtain these submodules, asrepo
won't do this automatically for you.
LF mailing list for release engineering
The #lf-releng
channel on Freenode IRC is usually well attendend.
When writing jobs, there are some things that JJB should be used to handle, and some things that should be put in external scripts or Jenkinsfile.
Some things that are good to put in a JJB job:
JJB Jobs shold not:
!include-escape
or !include-raw-escape
.When adding a new git repo that needs tests:
Create a new file in jjb/verify
named <reponame>.yaml
Create a project using the name of the repo, and a job-group section with a list of jobs-template id
s to invoke.
Optional: If you have more than one job that applies to the repo, add a dependency-jobs
variable to each item the job-group
jobs
list to control the order of jobs to invoke. Note that this is a string with the name of the jobs as created in Jenkins, not the job-template
id.
To create jobs that are usable by multiple repos, you want to create a job-template that can be used by multiple jobs.
Most job-template
s are kept in jjb/*.yaml
. See lint.yaml
or api-test.yaml
for examples.
Every job-template
must have at least a name
(which creates the name of the job in Jenkins) and an id
item (referred to in the job-group
), as well as several modules that invoke Jenkins functionality, or macros
(see below, and in the docs) that customize or provide defaults for those modules.
Default values can be found in jjb/defaults.yaml
. These can be used in projects
, jobs
, job-templates
.
NOTE: Defaults don't work with
macros
- all parameters must be passed to every macro invocation.
If you need to customize how a Jenkins module is run, consider creating a reusable macro. These are generally put in jjb/cord-macros.yaml
, and have names matching cord-infra-*
.
See also global-jjb/jjb/lf-macros.yaml
for more macros to use (these have name matching lf-infra-*
).
There are a few useful macros defined in jjb/cord-macros.yml
cord-infra-properties
- sets build discarder settingscord-infra-gerrit-repo-scm
- checks out the entire source tree with the repo
toolcord-infra-gerrit-repo-patch
- checks out a patch to a git repo within a checked out repo source tree (WIP, doesn't work yet)cord-infra-gerrit-trigger-patchset
- triggers build on gerrit new patchset, draft publishing, comments, etc.cord-infra-gerrit-trigger-merge
- triggers build on gerrit mergeJJB job definitions can be tested by running make test
, which will create a python virtualenv, install jenkins-job-builder in it, then try building all the job files, which are put in job-configs
and can be inspected.
The output of this is somewhat difficult to decipher, sometimes requiring you to go through the python backtrace to figure out where the error occurred in the jenkins-job-builder source code.
There is also a make lint
target which will run yamllint
on all the JJB YAML files, which can catch a variety of formatting errors.
If you're writing a new shell script, it's a good idea to test it with shellcheck before including it - failing to heed those messages then using !include-escape
to add it to the job may lead to hard to debug problems with the job definition.
The Jenkins executors are spun up automatically in EC2, and torn down after jobs have completed. Some are "one shot" and others (usually static or lint checks) are re-used for run multiple jobs.
The AMI images used for these executors built with Packer and most of the local configuration happens in packer/provision/basebuild.sh
. If you need a new tool installed in the executor, you would add the steps to install it here. It's verified, and when merged generates a new AMI image.
NOTE: Future builds won't automatically use the new AMI - you have to manually set the instance
AMI ID
on jenkins in Global Config > Cloud > Amazon EC2. The new AMI ID can be found near the end of the logs of the run of ci-management-packer-merge--basebuild.
If you create a new cloud instance type, make sure to set both the Security group names
and Subnet ID for VPC
or it will fail to instantiate.