Table of contents:
- Getting started
- Build the project
- Generating step framework
- Best practices for writing piper-go steps
- Testing
- Debugging
- Release
- Pipeline Configuration
- Security Setup
- Ramp up your development environment
- Get familiar with Go language
- Create a GitHub account
- Setup GitHub access via SSH
- Create and checkout a repo fork
- Optional: Get Jenkins related environment
- Optional: Get familiar with Jenkins Pipelines as Code
First you need to set up an appropriate development environment:
- Install Go, see GO Getting Started
- Install an IDE with Go plugins, see for example Go in Visual Studio Code
In order to get yourself started, there is a lot of useful information out there.
As a first step to take we highly recommend the Golang documentation, especially A Tour of Go.
We have a strong focus on high quality software and contributions without adequate tests will not be accepted. There is an excellent resource which teaches Go using a test-driven approach: Learn Go with Tests
The project uses Go modules. Thus please make sure to NOT checkout the project into your GOPATH
.
To check out this repository:
- Create your own fork of this repo
- Clone it to your machine, for example like:
mkdir -p ${HOME}/projects/jenkins-library
cd ${HOME}/projects
git clone [email protected]:${YOUR_GITHUB_USERNAME}/jenkins-library.git
cd jenkins-library
git remote add upstream [email protected]:sap/jenkins-library.git
git remote set-url --push upstream no_push
If you want to contribute also to the Jenkins-specific parts like
- Jenkins library step
- Jenkins pipeline integration
you need to do the following in addition:
- Install Groovy
- Install Maven
- Get a local Jenkins installed: Use for example cx-server
The Jenkins related parts depend on
You should get familiar with these concepts for contributing to the Jenkins-specific parts.
Use Docker:
docker build -t piper:latest .
You can extract the binary using Docker means to your local filesystem:
docker create --name piper piper:latest
docker cp piper:/build/piper .
docker rm piper
The steps are generated based on the yaml files in resources/metadata/
with the following command from the root of the project:
go generate
The yaml format is kept pretty close to Tekton's task format. Where the Tekton format was not sufficient some extenstions have been made.
Examples are:
- matadata - longDescription
- spec - inputs - secrets
- spec - containers
- spec - sidecars
There are certain extensions:
-
aliases allow alternative parameter names also supporting deeper configuration structures. Example
-
resources allow to read for example from a shared
commonPipelineEnvironment
which contains information which has been provided by a previous step in the pipeline via an output. Example -
secrets allow to specify references to Jenkins credentials which can be used in the
groovy
library. Example -
outputs allow to write to dedicated outputs like
- Influx metrics. Example
- Sharing data via
commonPipelineEnvironment
which can be used by another step as input
-
conditions allow for example to specify in which case a certain container is used (depending on a configuration parameter). Example
Implementing a new step starts by adding a new yaml file in resources/metadata/
and running
the step generator. This creates most of the boiler-plate code for the
step's implementation in cmd/
. There are four files per step based on the name given within the yaml:
cmd/<step>.go
- contains the skeleton of your step implementation.cmd/<step>_test.go
- write your unit tests here.cmd/<step>_generated.go
- contains the generated boiler plate code, and a dedicated type definition for your step's options.cmd/<step>_generated_test.go
- contains a simple unit test for the generated part.
You never edit in the generated parts. If you need to make changes, you make them in the yaml and re-run the step generator (which will of course not overwrite your implementation).
The file cmd/<step>.go
initially contains two functions:
func step(options stepOptions, telemetryData *telemetry.CustomData) {
err := runStep(&options, telemetryData)
if err != nil {
log.Entry().WithError(err).Fatal("step execution failed")
}
}
func runStep(options *stepOptions, telemetryData *telemetry.CustomData) error {
}
The separation into these two functions facilitates unit tests and mocking. From your tests, you could call
runStep()
with mocking instances of needed objects, while inside step()
, you create runtime instances of these
objects.
Logging is done via the sirupsen/logrus framework. It can conveniently be accessed through:
import (
"github.com/SAP/jenkins-library/pkg/log"
)
func myStep ...
...
log.Entry().Info("This is my info.")
...
}
If a fatal error occurs your code should act similar to:
...
if err != nil {
log.Entry().
WithError(err).
Fatal("failed to execute step ...")
}
Calling Fatal
results in an os.Exit(0)
and before exiting some cleanup actions (e.g. writing output data,
writing telemetry data if not deactivated by the user, ...) are performed.
In order to better understand the root cause of errors that occur, we wrap errors like
f, err := os.Open(path)
if err != nil {
return errors.Wrapf(err, "open failed for %v", path)
}
defer f.Close()
We use github.com/pkg/errors for that.
It has proven a good practice to bubble up errors until the runtime entry function and only there exit via the logging framework (see also logging).
For errors, we have a convenience function to set a pre-defined category once an error occurs:
log.SetErrorCategory(log.ErrorCompliance)
Error categories are defined in pkg/log/ErrorCategory
.
With writing a fatal error
log.Entry().WithError(err).Fatal("the error message")
the category will be written into the file errorDetails.json
and can be used from there in the further pipeline flow.
Writing the file is handled by pkg/log/FatalHook
.
Unit tests are done using basic golang
means.
Additionally, we encourage you to use github.com/stretchr/testify/assert in order to have slimmer assertions if you like. A good pattern to follow is this:
func TestNameOfFunctionUnderTest(t *testing.T) {
t.Run("A description of the test case", func(t *testing.T) {
// init
// test
// assert
})
t.Run("Another test case", func(t *testing.T) {
// init
// test
// assert
})
}
This will also structure the test output for better readability.
Tests should be written only for the code of your step implementation, while any external functionality should be mocked, in order to test all code paths including the error cases.
There are (at least) two approaches for this:
In this approach you declare an interface that contains every external function used within your step that you need to be able to mock. In addition, you declare a struct which holds the data you need during runtime, and implement the interface with the "real" functions. Here is an example to illustrate:
import (
"github.com/SAP/jenkins-library/pkg/piperutils"
)
type myStepUtils interface {
fileExists(path string) (bool, error)
fileRead(path string) ([]byte, error)
}
type myUtilsData struct {
fileUtils piperutils.Files
}
func (u *myUtilsData) fileExists(path string) (bool, error) {
return u.fileUtils.FileExists(path)
}
func (u *myUtilsData) fileRead(path string) ([]byte, error) {
return u.fileUtils.FileRead(path)
}
Then you create the runtime version of the utils data in your top-level entry function and
pass it to your run*()
function:
func step(options stepOptions, _ *telemetry.CustomData) {
utils := myUtilsData{
fileUtils: piperutils.Files{},
}
err := runStep(&options, &utils)
...
}
func runStep(options *stepOptions, utils myStepUtils) error {
...
exists, err := utils.fileExists(path)
...
}
In your tests, you would provide a mocking implementation of this interface and pass
instances of that to the functions under test. To better illustrate this, here is an example
for the interface above implemented in the <step>_test.go
file:
type mockUtilsBundle struct {
files map[string][]byte
}
func newMockUtilsBundle() mockUtilsBundle {
utils := mockUtilsBundle{}
utils.files = map[string][]byte{}
return utils
}
func (m *mockUtilsBundle) fileExists(path string) (bool, error) {
content := m.files[path]
return content != nil, nil
}
func (m *mockUtilsBundle) fileRead(path string) ([]byte, error) {
content := m.files[path]
if content == nil {
return nil, fmt.Errorf("could not read '%s': %w", path, os.ErrNotExist)
}
return content, nil
}
// This is how it would be used in tests:
func TestSomeFunction() {
t.Run("Happy path", func(t *testing.T) {
// init
utils := newMockUtilsBundle()
utils.files["some/path/file.xml"] = []byte(´content of the file´)
// test
err := someFunction(&utils)
// assert
assert.NoError(t, err)
})
t.Run("Error path", func(t *testing.T) {
// init
utils := newMockUtilsBundle()
// test
err := someFunction(&utils)
// assert
assert.EqualError(t, err, "could not read 'some/path/file.xml'")
})
}
An alternative approach are global function pointers:
import (
FileUtils "github.com/SAP/jenkins-library/pkg/piperutils"
)
var fileUtilsExists = FileUtils.FileExists
func someFunction(options *stepOptions) error {
...
exists, err := fileUtilsExists(path)
...
}
In your tests, you can then simply set the function pointer to a mocking implementation:
func TestSomeFunction() {
t.Run("Happy path", func(t *testing.T) {
// init
originalFileExists := fileUtilsExists
fileUtilsExists = func(filename string) (bool, error) {
return true, nil
}
defer fileUtilsExists = originalFileExists
// test
err := someFunction(...)
// assert
assert.NoError(t, err)
})
t.Run("Error path", func(t *testing.T) {
// init
originalFileExists := fileUtilsExists
fileUtilsExists = func(filename string) (bool, error) {
return false, errors.New("something happened")
}
defer fileUtilsExists = originalFileExists
// test
err := someFunction(...)
// assert
assert.EqualError(t, err, "something happened")
})
}
Both approaches have their own benefits. Global function pointers require less preparation in the actual implementation and give great flexibility in the tests, while mocking interfaces tend to result in more code re-use and slim down the tests. The mocking implementation of a utils interface can facilitate implementations of related functions to be based on shared data.
Tests that can be executed in parallel should be marked as such.
With the command t.Parallel()
the test framework can be notified that this test can run in parallel, and it can start running the next test.
(Example in Stackoverflow)
Therefore, this command shall be called at the beginning of a test method and also in each t.Run()
sub tests.
See also the documentation for t.Parallel()
and t.Run()
.
func TestMethod(t *testing.T) {
t.Parallel() // indicates that this method can run parallel to other methods
t.Run("sub test 1", func(t *testing.T){
t.Parallel() // indicates that this sub test can run parallel to other sub tests
// execute test
})
t.Run("sub test 2", func(t *testing.T){
t.Parallel() // indicates that this sub test can run parallel to other sub tests
// execute test
})
}
Go will first execute the non-parallelized tests in sequence and afterwards execute all the parallel tests in parallel, limited by the default number of parallel executions.
It is important that tests executed in parallel use the variable values actually meant to be visible to them.
Especially in table tests, it can happen easily that a variable injected into the t.Run()
-closure via the outer scope is changed before or while the closure executes.
To prevent this, it is possible to create shadowing instances of variables in the body of the test loop.
(See blog about it.)
At the minimum, you need to capture the test case value from the loop iteration variable, by shadowing this variable in the loop body.
Inside the t.Run()
closure, this shadow copy is visible, and cannot be overwritten by later loop iterations.
If you do not make this shadowing copy, what is visible in the closure is the variable which gets re-assigned with a new value in each loop iteration.
The value of this variable is then not fixed for the test run.
func TestMethod(t *testing.T) {
t.Parallel() // indicates that this method can parallel to other methods
testCases := []struct {
Name string
}{
{
Name: "Name1"
},
{
Name: "Name2"
},
}
for _, testCase := range testCases { // testCase defined here is re-assigned in each iteration
testCase := testCase // define new variable within loop to detach from overwriting of the outer testCase variable by next loop iteration
// The same variable name "testCase" is used for convenience.
t.Run(testCase.Name, func(t *testing.T) {
t.Parallel() // indicates that this sub test can run parallel to other sub tests
// execute test
})
}
}
Debugging can be initiated with VS code fairly easily. Compile the binary with specific compiler flags to turn off optimizations go build -gcflags "all=-N -l" -o piper.exe
.
Modify the launch.json
located in folder .vscode
of your project root to point with program
exactly to the binary that you just built with above command - must be an absolute path. Add any arguments required for the execution of the Piper step to args
. What is separated with a blank on the command line must go into a separate string.
{
// Use IntelliSense to learn about possible attributes.
// Hover to view descriptions of existing attributes.
// For more information, visit: https://go.microsoft.com/fwlink/?linkid=830387
"version": "0.2.0",
"configurations": [
{
"name": "Launch",
"type": "go",
"request": "launch",
"mode": "exec",
"program": "C:/CF@HCP/git/jenkins-library-public/piper.exe",
"env": {},
"args": ["checkmarxExecuteScan", "--password", "abcd", "--username", "1234", "--projectName", "testProject4711", "--serverUrl", "https://cx.server.com/"]
}
]
}
Finally, set your breakpoints and use the Launch
button in the VS code UI to start debugging.
Releases are performed using Project "Piper" Action.
We release on schedule (once a week) and on demand.
To perform a release, the respective action must be invoked for which a convenience script is available in contrib/perform-release.sh
.
It requires a personal access token for GitHub with repo
scope.
Example usage PIPER_RELEASE_TOKEN=THIS_IS_MY_TOKEN contrib/perform-release.sh
.
The pipeline configuration is organized in a hierarchical manner and configuration parameters are incorporated from multiple sources. In general, there are four sources for configurations:
- Directly passed step parameters
- Project specific configuration placed in
.pipeline/config.yml
- Custom default configuration provided in
customDefaults
parameter of the project config or passed as parameter to the stepsetupCommonPipelineEnvironment
- Default configuration from Piper library
For more information and examples on how to configure a project, please refer to the configuration documentation.
The configuration of a project is, as of now, resolved separately for Groovy and Go steps. There are, however, dependencies between the steps responsible for resolving the configuration. The following provides an overview of the central components and their dependencies.
The step setupCommonPipelineEnvironment
initializes the commonPipelineEnvironment
and DefaultValueCache
.
Custom default configurations can be provided as parameters to setupCommonPipelineEnvironment
or via the customDefaults
parameter in project configuration.
The DefaultValueCache
caches the resolved (custom) default pipeline configuration and the list of configurations that contributed to the result.
On initialization, it merges the provided custom default configurations with the default configuration from Piper library, as per the hierarchical order.
Note, the list of configurations cached by DefaultValueCache
is used to pass path to the (custom) default configurations of each Go step.
It only contains the paths of configurations which are not provided via customDefaults
parameter of the project configuration, since the Go layer already resolves configurations provided via customDefaults
parameter independently.
You can find additional hints at documentation/developer-hints
Here some hints and tricks are described to enhance the security within the development process.
In git, commits can be signed to guarantee that that changes were made by the person named in the commit. The name and email used for commits can be easily modified in the local git setup and afterwards it cannot be distinguished anymore if the commit was done by the real person or by some potential attacker.
In Windows, this can be done via GnuPG. Download and install the tool. Via the manager tool Kleopatra a new key pair can be easily created with a little wizard. Make sure that the name and email are the ones used in your git.
The public key must then be added to the github's GPG section. The private key should be kept in a backup as this signature is bound to you and not your machine.
The only thing left are some changes in the .gitconfig file. The file shall be located in your user directory. It might look something like the following. All parts that are not relevant for signing were removed.
[user]
name = My Name
email = [email protected]
# Hash or email of you GPG key
signingkey = D3CF72CC4006DE245C049566242831AEEE9DA2DD
[commit]
# enable signing for commits
gpgsign = true
[tag]
# enable signing for tags (note the capital S)
gpgSign = true
[gpg]
# Windows was not able to find the private key. Setting the gpg command to use solved this.
program = C:\\Program Files (x86)\\GnuPG\\bin\\gpg.exe
Add the three to four lines to you git config and this will do the necessary such that all your commits will be signed.