Skip to content

ondrejsika/gitlab-ci-training

Repository files navigation

Ondrej Sika (sika.io) | [email protected] | join slack, open slack

Gitlab CI Training

About Course

Any Questions?

Write me mail to [email protected]

Course

About Me - Ondrej Sika

Freelance DevOps Engineer, Consultant & Lecturer

  • Complete DevOps Pipeline
  • Open Source / Linux Stack
  • Cloud & On-Premise
  • Technologies: Git, Gitlab, Gitlab CI, Docker, Kubernetes, Terraform, Prometheus, ELK / EFK, Rancher, Proxmox

Star, Create Issues, Fork, and Contribute

Feel free to star this repository or fork it.

If you found bug, create issue or pull request.

Also feel free to propose improvements by creating issues.

DevOps Live (Czech only)

https://devopslive.cz

DevOps Kniha (Czech only)

https://kniha.sika.io

Chat

For sharing links & "secrets".

Agenda

What is CI

Git

Git is a version control system for tracking changes in computer files and coordinating work on those files among multiple people.

Gitlab

From project planning and source code management to CI/CD and monitoring, GitLab is a complete DevOps platform, delivered as a single application.

CI

In software engineering, continuous integration is the practice of merging all developer working copies to a shared mainline several times a day.

Usage of CI

Automatization of:

  • build process
  • testing - unit tests, integration tests, linting, formating
  • deployment - dev, staging, production

Gitlab CI Features

  • Integration into Gitlab (CE & EE)
  • Versioned (YAML file in repository)
  • Easy to scale
  • Docker & Kubernetes support

Gitlab CI Architecture

gitlab-ci-architecture

Gitlab Runner

Runs on every platform - Linux, Mac, Windows, Docker, Kubernetes

How to install & configure:

My scripts to bootstrap Gitlab Runner in Docker: https://github.com/ondrejsika/ondrejsika-gitlab-runner

Setup Gitlab CI

Demo Gitlab

Setup

You can setup Gitlab on Digital Ocean & Cloudflare using Terraform

https://github.com/ondrejsika/terraform-demo-gitlab

Login

SSH Password

Password is asdfasdf1234 for both: Girlab on [email protected] and Runner on [email protected]

Install Runners into Gitlab

Go to Admin -> Runners

And Set up a shared Runner manually

Setup Gitlab Runner using slu

Install slu

curl -fsSL https://raw.githubusercontent.com/sikalabs/slu/master/install.sh | sudo sh

From Vault

slu login --url https://vault-slu.sikalabs.io --user gitlab-ci-sikademo --password gitlab-ci-sikademo
slu gitlab-ci setup-runner --gitlab sikademo

From CLI arguments

slu gitlab-ci setup-runner --gitlab-url https://gitlab.sikademo.com --token <token>

[DEPRECATED] Setup runner by shell scripts

git clone https://github.com/sikalabs/sikalabs-gitlab-runner
cd sikalabs-gitlab-runner
./create-runner.sh
./register-runner.sh https://gitlab.sikademo.com/ h94VrzQzZnJ_va3hxGGW
./set-concurrency.sh

Gitlab Workflow (CI) VS Code Extension

https://marketplace.visualstudio.com/items?itemName=GitLab.gitlab-workflow

CI Jobs

Gitlab CI YAML

Configuration of everything aroud Gitlab CI is stored inside Git repository in file .gitlab-ci.yml

If you don't know YAML format, check out this simple YAML tutorial - https://learnxinyminutes.com/docs/yaml/

Here is a Gitlab CI YAML reference - https://docs.gitlab.com/ee/ci/yaml/

First Job

Create new projet and create there file .gitlab-ci.yml with following content:

# .gitlab-ci.yml

job:
  script: echo Hello World!

Push to Gitlab and check it out.

Job

Jobs are smalles units which can be executed by Gitlab CI. Here are samples of common job configurations.

Jobs are top level object in Gitlab CI YAML files instead of few keywords. Keywords are: image, services, stages, types, before_script, after_script, variables, cache

Script

Every job require script - it's a shell scrip which will be executed by job. Script can be string or list of strings.

# .gitlab-ci.yml

job1:
  script: echo Hello World!
job2:
  script:
    - echo Hello World!
    - echo Ahoj Svete!

Tags

You can choose specific set of runners to execute the job

# .gitlab-ci.yml

build:
  tags:
    - build
  script: echo Build!

deploy:
  tags:
    - kubernetes
  script: echo Deploy!

Stages

You can define order of jobs by stages. You can define stages and their order. Jobs in same stage run in parallel and after CI finishes all job in stage, then start jobs from next.

# .gitlab-ci.yml

stages:
  - build
  - test

build:
  stage: build
  script: echo Build!

test1:
  stage: test
  script: echo Test1!

test2:
  stage: test
  script: echo Test2!

.pre, .post stages

Default stages which runs before and after our user-defined stages. No need to specify them in stages.

stages:
  - build
  - test

pre:
  stage: .pre
  script: echo Pre ...

build:
  stage: build
  script: echo Build!

test:
  stage: test
  script: echo Test!

post:
  stage: .post
  script: echo Post ...

Needs

# .gitlab-ci.yml

stages:
  - build
  - test
  - deploy

linux build:
  stage: build
  script: sleep 10 && echo Done

mac build:
  stage: build
  script: sleep 20 && echo Done

lint:
  stage: test
  needs: []
  script: echo Done

linux unit tests:
  stage: test
  needs:
    - linux build
  script: echo Done

linux e2e tests:
  stage: test
  needs:
    - linux build
  script: sleep 10 && echo Done

mac unit tests:
  stage: test
  needs:
    - mac build
  script: sleep 5 && echo Done

mac e2e tests:
  stage: test
  needs:
    - mac build
  script: sleep 30 && echo Done

release linux:
  stage: deploy
  script: "echo Done"
  needs:
    - lint
    - linux unit tests
    - linux e2e tests

release mac:
  stage: deploy
  script: "echo Done"
  needs:
    - lint
    - mac unit tests
    - mac e2e tests

Needs in same stage

stages:
  - test

test1:
  stage: test
  script: echo Test1!

test2:
  stage: test
  script: echo Test2!
  needs:
    - test1

Before & After Script

You can define script which will be executed befor and after job script. You can define those script globally or per job.

# .gitlab-ci.yml

before_script:
  - echo Global before

after_script:
  - echo Global after

job1:
  script: echo Job1!

job2:
  script: echo Job2! && false

job3:
  before_script:
    - echo Local before
  after_script: []
  script: echo Job3!

Cleanup after script

# .gitlab-ci.yml

job1:
  script:
    - echo deploy!
    - echo test! && false
    - echo cleanup!

job2:
  script:
    - echo deploy!
    - echo test! && false
  after_script:
    - echo cleanup!

Job status in after script

job1:
  script:
    - echo deploy!
    - echo test!
  after_script:
    - echo $CI_JOB_STATUS

job2:
  script:
    - echo deploy!
    - echo test! && false
  after_script:
    - echo $CI_JOB_STATUS

CI_JOB_STATUS

When

You can control when do you want run your jobs. By default, jobs are executed automatically when previous stage succeed. You can specify another condition, you can run jobs manually, always or on error.

# .gitlab-ci.yml

stages:
  - build
  - test

build:
  stage: build
  script: echo Run build ...
  # script: echo Run build ... && false

test:
  stage: test
  script: echo Run test ...

deploy:
  stage: test
  script: echo Run deploy ...
  when: manual

diagnostics:
  stage: test
  script: echo Run diagnostics ...
  when: on_failure

reporting:
  stage: test
  script: echo Run CI reporting ...
  when: always

Example with following jobs after manual job

stages:
  - deploy prod
  - test prod

deploy prod:
  stage: deploy prod
  when: manual
  allow_failure: false
  # allow_failure: true
  script:
    - echo Deploy ...

test prod:
  stage: test prod
  script:
    - echo Test ...

Allow Failure

You can specify flag allow_failure to true, job can fail but pipeline will succeed.

# .gitlab-ci.yml

test:
  script: echo test ... && false
  allow_failure: true

Manual jobs are allowed to fail by default, if you want to disallow failure, you have to set allow_failure to false.

# .gitlab-ci.yml

test:
  when: manual
  script: echo test ... && false

test2:
  when: manual
  script: echo test ... && false
  allow_failure: false

Retry

# .gitlab-ci.yml

image: sikalabs/ci

test1:
  script: '[ "$(slu random int --max 2)" == "0" ]'
  retry: 2

test2:
  script: '[ "$(slu random int --max 2)" == "0" ]'
  retry: 2

test3:
  script: '[ "$(slu random int --max 2)" == "0" ]'
  retry: 2

test4:
  script: slu chaos-monkey random-status-code --verbose
  retry: 2

test5:
  script: slu chaos-monkey random-status-code --verbose
  retry: 2

test6:
  script: slu chaos-monkey random-status-code --verbose
  retry: 2

Variables

Gitlab CI offers you lots of usable variables like:

  • CI
  • CI_PROJECT_PATH_SLUG, CI_COMMIT_REF_SLUG
  • CI_COMMIT_SHORT_SHA
  • CI_PIPELINE_SOURCE
  • CI_COMMIT_TAG, CI_COMMIT_BRANCH, CI_DEFAULT_BRANCH
  • CI_PIPELINE_ID, CI_JOB_ID
  • CI_REGISTRY, CI_REGISTRY_USER, CI_REGISTRY_PASSWORD
  • CI_JOB_STATUS

See all varibles: https://docs.gitlab.com/ce/ci/variables/predefined_variables.html#variables-reference

You can define own variables in:

  • Predefined CI Variables
  • Group CI Settings
  • Project CI Setting
  • Globally in CI YAML
  • In job in CI YAML

You can define varible in Settings -> CI / CD -> Variables. Same for project and group. You can define for example connection to your Kubernetes cluster, AWS credentials, ...

Variables can be defined as:

  • Masked - Value is hidden from the CI output. You probably dont want to show any credential, even development one.
  • Protected - Protected variable appears only in jobs on protected branches. If developers can't push to protected branches, there have no chance to get production deployment keys or deploy to production. After code has been merged to master (protected), then protected variables appears and you can deploy to production.

Example job:

# .gitlab-ci.yml

variables:
  XXX_GLOBAL: global

job1:
  script: env | grep XXX

job2:
  variables:
    XXX_LOCAL: local
  script: env | grep XXX

job3:
  variables:
    XXX_FOO: foo
    XXX_BAR: bar
    XXX_FOO_BAR: $XXX_FOO-$XXX_BAR
  script: env | grep XXX

job4:
  variables:
    TAG: $CI_COMMIT_REF_SLUG-$CI_COMMIT_SHORT_SHA-$CI_PIPELINE_ID
    IMAGE: $CI_REGISTRY_IMAGE:$TAG
    BASE_HOST: sikademo.com
    HOST: $CI_PROJECT_PATH_SLUG-$CI_COMMIT_REF_SLUG.$BASE_HOST
  script:
    - echo docker build . -t $IMAGE
    - echo helm install ... --set host=$HOST

Variables with Terraform

You can manage project or group variables using Terraform.

resource "gitlab_project_variable" "base_host" {
  project  = local.project_id
  key      = "BASE_HOST"
  value    = "k8s.sikademo.com
}

See example ./terraform-gitlab-ci-variables

You can manage entire Gitlab with Terraform ;)

Rules

You can specify another condition when you can run jobs. You can specify branches and tags on which you want to run your jobs or not.

Simple example:

# .gitlab-ci.yml

push:
  rules:
    - if: $CI_PIPELINE_SOURCE == "push"
  script:
    - echo push

run:
  rules:
    - if: $CI_PIPELINE_SOURCE == "web"
  script:
    - echo run

More advanced example:

# .gitlab-ci.yml

stages:
  - unit_test
  - integration_test
  - build
  - deploy

unit_test:
  stage: unit_test
  script: echo Unit Test ...
  rules:
    - if: $CI_COMMIT_BRANCH
    - if: $CI_COMMIT_TAG
    - if: $CI_PIPELINE_SOURCE == "merge_request_event"

integration_test:
  stage: integration_test
  script: echo Integration Test ...
  rules:
    - if: $CI_PIPELINE_SOURCE == "merge_request_event"

build:
  stage: build
  script: echo Build ...
  rules:
    - if: $CI_COMMIT_BRANCH == $CI_DEFAULT_BRANCH
    - if: $CI_COMMIT_TAG

deploy:
  stage: deploy
  script: echo Deploy ...
  rules:
    - if: $CI_COMMIT_BRANCH == $CI_DEFAULT_BRANCH
      when: manual
      allow_failure: false
    - if: $CI_COMMIT_TAG

Rules Changes

You can run job when are changes is some files. That's great for monorepos.

# .gitlab-ci.yml

Build Frontend:
  script: echo Build Frontend ...
  rules:
    - changes:
        - frontend/**

Build Backend:
  script: echo Build Backend ...
  rules:
    - changes:
        - backend/**

If && Changes Together

Build Frontend:
  script: echo Build Frontend ...
  rules:
    - if: $CI_COMMIT_BRANCH == $CI_DEFAULT_BRANCH || $CI_MERGE_REQUEST_ID
      changes:
        - frontend/**/*

Build Backend:
  script: echo Build Backend ...
  rules:
    - if: $CI_COMMIT_BRANCH == $CI_DEFAULT_BRANCH || $CI_MERGE_REQUEST_ID
      changes:
        - backend/**/*

Merge Request Pipelines

stages:
  - build
  - test
  - deploy

build:
  stage: build
  script: echo build
  rules:
    - if: $CI_COMMIT_BRANCH
    - if: $CI_MERGE_REQUEST_ID

test:
  stage: test
  script: echo test
  rules:
    - if: $CI_MERGE_REQUEST_ID

deploy:
  stage: deploy
  script: echo deploy
  rules:
    - if: $CI_COMMIT_BRANCH == $CI_DEFAULT_BRANCH

Workflow

Workflow Name

# .gitlab-ci.yml

workflow:
  name: Pipeline $CI_COMMIT_REF_NAME $CI_COMMIT_SHORT_SHA $CI_COMMIT_TITLE

Cache

Cache is used to specify a list of files and directories which should be cached between jobs. You can only use paths that are within the project workspace.

If cache is defined outside the scope of jobs, it means it is set globally and all jobs will use that definition.

You can specify cache key for better caching of different branches

# .gitlab-ci.yml

job:
  cache:
    key: ${CI_COMMIT_REF_SLUG}
    paths:
      - now
  script:
    - date >> now
    - cat now

More about caches: https://docs.gitlab.com/ce/ci/caching/index.html

Git Clean Flags

Instead of cache, you can use faster option how to keep some data in repository. If you are using cache, your repository is cleaned by git clean -ffdx, which means remove all ignored & untracked files and directories. Then your cache is downloaded and extracted into your working directory. This way is not fastest way if you have thousands files in cache (like node modules, for example).

Instead of cache, you can specify variable GIT_CLEAN_FLAGS and use own git clean arguments. If you want to ignore node modules from git clean, you can set -ffdx -e .yarn-cache.

# .gitlab-ci.yml

variables:
  GIT_CLEAN_FLAGS: -ffdx -e .yarn-cache

job:
  script:
    - yarn install --frozen-lockfile --cache-folder .yarn-cache --prefer-offline
    - yarn build

Artifacts

Artifacts is used to specify a list of files and directories which should be attached to the job when it succeeds, fails, or always.

The artifacts will be sent to GitLab after the job finishes and will be available for download in the GitLab UI.

Artifact are also distributed to jobs in next stages.

# .gitlab-ci.yml

stages:
  - build
  - test

build:
  stage: build
  script:
    - mkdir -p out
    - echo '<h1>Hello World!</h1>' > out/index.html
  artifacts:
    paths:
      - out

test:
  stage: test
  script:
    - cat out/index.html

When the job succed, you can browse and download job from Gitlab.

More about artifacts: https://docs.gitlab.com/ce/user/project/pipelines/job_artifacts.html

Gitlab Pages

Enable Gitlab Pages in /etc/gitlab/gitlab.rb

pages_external_url "http://pages.sikademo.com"

Gitlab pages MUST be on http (NOT https) or you have to provide wildcard certificate for *.pages.sikademo.com. See: https://gitlab.com/gitlab-org/gitlab/-/issues/376130

Artifacts Dependencies

By default, all artifacts will be passed to jobs in following stages. If you want artifact only from specific jobs, you can use dependencies to choose which articact you want.

Fists usage is when artifact are on same path and there will be colision, see:

# .gitlab-ci.yml

stages:
  - build
  - test

build_A:
  stage: build
  script: mkdir -p out && echo '<h1>Hello from Project A!</h1>' > out/index.html
  artifacts:
    paths:
      - out

build_B:
  stage: build
  script: mkdir -p out && echo '<h1>Hello from Project B!</h1>' > out/index.html
  artifacts:
    paths:
      - out

test A:
  stage: test
  script: cat out/index.html
  dependencies:
    - build_A

test B:
  stage: test
  script: cat out/index.html
  dependencies:
    - build_B

Or you can use dependencies when you have lots of artifact and dont want to slow down your jobs by downloading unnecessary artifacts.

JUnit Test Reports

Docs

# test_example.py
def test_ok():
    assert True

def test_err():
    assert False
# .gitlab-ci.yml
test:
  image: ondrejsika/pytest
  script:
    - pytest --junitxml=report.xml
  artifacts:
    reports:
      junit: report.xml

Docker

  • Fully supported
  • Easiest way how to create build environment
  • Easiest way how to run and distribute your software

If you have a Docker or Kubernetes executor (we have) you can define image where you want to run your job.

You can specify image globally or in job:

# .gitlab-ci.yml

image: sikalabs/ci

build:
  image: node
  script:
    - yarn
    - yarn run build

deploy:
  script:
    - kubectl apply -f kubernetes.yml

You can also run docker commands from job because we have added Docker socket there. See here and here.

# .gitlab-ci.yml

variables:
  TAG: $CI_COMMIT_REF_SLUG-$CI_COMMIT_SHORT_SHA-$CI_PIPELINE_ID
  IMAGE: $CI_REGISTRY_IMAGE:$TAG

job:
  script:
    - docker login $CI_REGISTRY -u $CI_REGISTRY_USER -p $CI_REGISTRY_PASSWORD
    - docker build -t $IMAGE .
    - docker push $IMAGE

Docker Image sikalabs/ci

That is my image which I use for most of CI jobs.

It contains lots of common tools like git, zip, curl, wget, Docker client, Docker Compose, Kubernetes client, ...

You can see the repository on Github - https://github.com/sikalabs/sikalabs-container-images/tree/master/ci

Kaniko

https://github.com/GoogleContainerTools/kaniko

variables:
  GIT_CLEAN_FLAGS: -ffdx -e .kaniko-cache
  TAG: $CI_COMMIT_REF_SLUG-$CI_COMMIT_SHORT_SHA-$CI_PIPELINE_ID
  IMAGE: $CI_REGISTRY_IMAGE:$TAG

job:
  image:
    name: gcr.io/kaniko-project/executor:v1.19.2-debug
    entrypoint: [""]
  script:
    - echo "{\"auths\":{\"$CI_REGISTRY\":{\"username\":\"$CI_REGISTRY_USER\",\"password\":\"$CI_REGISTRY_PASSWORD\"}}}" > /kaniko/.docker/config.json
    - /kaniko/executor
      --context .
      --dockerfile ./Dockerfile
      --destination $IMAGE
      --cache=true
      --cache-dir .kaniko-cache
      --cache-repo $CI_REGISTRY_IMAGE/cache

matrix

# .gitlab-ci.yml

image: golang:1.22-bullseye

build:
  parallel:
    matrix:
      - GOOS:
          - linux
          - darwin
          - windows
        GOARCH:
          - amd64
          - arm64
  artifacts:
    paths:
      - hello-${GOOS}-${GOARCH}
  script:
    - go build -o hello-${GOOS}-${GOARCH} main.go
// main.go

package main

import "fmt"

func main() {
	fmt.Println("Hello World")
}

Environments

Environment is used to define that a job deploys to a specific environment.

If environment is specified and no environment under that name exists, a new one will be created automatically.

# .gitlab-ci.yml

variables:
  BASE_HOST: dev.company.com
  HOST: $CI_PROJECT_PATH_SLUG-$CI_COMMIT_REF_SLUG.$BASE_HOST

deploy:
  script: echo 'Deploy!'
  environment:
    name: $CI_COMMIT_REF_SLUG
    url: https://$HOST

Stop Environmet

You can define stop job for the environment which you can stop manually from Gitlab. If stop job is defined, Gitlab automatically stop unnecessary environments, like envirments created from merged or deleted branches.

# .gitlab-ci.yml

deploy:
  script: echo Deploy!
  environment:
    name: $CI_COMMIT_REF_SLUG
    url: https://$CI_COMMIT_REF_SLUG.dev.company.com
    on_stop: stop

stop:
  variables:
    GIT_STRATEGY: none
  script: echo Stop!
  when: manual
  environment:
    name: $CI_COMMIT_REF_SLUG
    action: stop

Environment: Auto Stop In

https://docs.gitlab.com/ee/ci/yaml/#environmentauto_stop_in

deploy_review:
  script: echo Deploy for review!
  environment:
    name: $CI_COMMIT_REF_SLUG
    url: https://$CI_COMMIT_REF_SLUG.dev.company.com
    on_stop: stop_rewiew
    auto_stop_in: 1 day

Extends

# .gitlab-ci.yml

stages:
  - deploy dev
  - deploy prod

.dev:
  stage: deploy dev
  variables:
    ENV: dev

.prod:
  stage: deploy prod
  variables:
    ENV: prod

.deploy:
  script: echo Deploy $ENV

deploy dev:
  extends:
    - .deploy
    - .dev

deploy prod:
  extends:
    - .deploy
    - .prod

Includes

Docs

# .gitlab-ci.yml

stages:
  - build
  - deploy

include:
  - .gitlab-ci-build.yml
  - .gitlab-ci-deploy.yml
# .gitlab-ci-build.yml

build:
  stage: build
  script: echo build
# .gitlab-ci-deploy.yml

deploy:
  stage: deploy
  script: echo deploy

Child Pipelines

Docs

# .gitlab-ci.yml

pipeline_a:
  trigger:
    include: a/.gitlab-ci.yml
    strategy: depend
  rules:
    - changes:
        - a/**

pipeline_b:
  trigger:
    include: b/.gitlab-ci.yml
    strategy: depend
  rules:
    - changes:
        - b/**
# a/.gitlab-ci.yml

stages:
  - build
  - deploy

build:
  stage: build
  script: echo Build service A

deploy:
  stage: deploy
  script: echo Deploy service A
# b/.gitlab-ci.yml

stages:
  - build
  - deploy

build:
  stage: build
  script: echo Build service B

deploy:
  stage: deploy
  script: echo Deploy service B

Generated Pipelines

# .gitlab-ci.yml
stages:
  - generate
  - pipeline

generate:
  stage: generate
  image: python:3.7-slim
  script: python generate-gitlab-ci.py
  artifacts:
    paths:
      - .gitlab-ci.generated.yml

pipeline:
  stage: pipeline
  trigger:
    include:
      - artifact: .gitlab-ci.generated.yml
        job: generate
    strategy: depend
# generate-gitlab-ci.py
import json

SERVICES = (
    "foo",
    "bar",
    "baz",
)

def make_service(name):
    return {
        "build-%s" % name: {
            "stage": "build",
            "script":[
                "echo Build %s" % name,
            ]
        },
        "deploy-%s" % name: {
            "stage": "deploy",
            "script":[
                "echo Deploy %s" % name,
            ]
        }
    }

with open(".gitlab-ci.generated.yml", "w") as f:
    pipeline = {}
    pipeline.update({
        "stages": ["build", "deploy"]
    })
    for service in SERVICES:
        pipeline.update(make_service(service))
    f.write(json.dumps(pipeline))

Multi Project Pipelines

Docs

Project foo depends on library bar. If you make any change in library bar you have to trigger pipeline in project foo. ondrejsika/foo is full path to project where you want to run pipeline.

# .gitlab-ci.yml (library bar)
job:
  script: Do something

triger-pipelines:
  trigger: ondrejsika/foo

Terraform Pipeline

image: registry.gitlab.com/gitlab-org/terraform-images/stable:latest
variables:
  TF_ROOT: ${CI_PROJECT_DIR}
  TF_ADDRESS: ${CI_API_V4_URL}/projects/${CI_PROJECT_ID}/terraform/state/${CI_PROJECT_NAME}

cache:
  paths:
    - ${TF_ROOT}/.terraform

before_script:
  - cd ${TF_ROOT}

stages:
  - validate
  - plan
  - apply

validate:
  stage: validate
  script:
    - gitlab-terraform init
    - gitlab-terraform validate

plan:
  stage: plan
  script:
    - gitlab-terraform plan
    - gitlab-terraform plan-json
  artifacts:
    name: plan
    paths:
      - ${TF_ROOT}/plan.cache
    reports:
      terraform: ${TF_ROOT}/plan.json

apply:
  stage: apply
  script:
    - gitlab-terraform apply
  dependencies:
    - plan
  rules:
    - if: $CI_COMMIT_BRANCH == $CI_DEFAULT_BRANCH
      when: manual
      allow_failure: false

Inspired by https://gitlab.com/gitlab-org/configure/examples/gitlab-terraform-aws/-/blob/master/.gitlab-ci.yml

Scheduled Pipelines

You can schedule pipeline.

nightly:
  script: echo Nightly build ...
  rules:
    - if: $CI_PIPELINE_SOURCE == "schedule"
      when: never

Schedule setting in gitlab.rb

If you want to run scheduled pipelines in less than hour, you have to set custom pipeline_schedule_worker_cron in /etc/gitlab/gitlab.rb

gitlab_rails['pipeline_schedule_worker_cron'] = "*/5 * * * *"

Scheduled Pipelines using Terraform

resource "gitlab_pipeline_schedule" "nightly" {
  project     = local.project_id
  description = "nightly builds"
  ref         = "master"
  cron        = "0 0 * * *"
}

resource "gitlab_pipeline_schedule_variable" "nightly" {
  project              = gitlab_pipeline_schedule.nightly.project
  pipeline_schedule_id = gitlab_pipeline_schedule.nightly.id
  key                  = "NIGHTLY"
  value                = "yes"
}

Gitlab Runner in Kubernetes using Operator and Helm

Using sikalabs/gitlab-runner chart.

Install OLM

curl -sL https://github.com/operator-framework/operator-lifecycle-manager/releases/download/v0.23.1/install.sh | bash -s v0.23.1

Install Gitlab Runner Operator

kubectl create -f https://operatorhub.io/install/gitlab-runner-operator.yaml

Create Runner

helm upgrade --install \
  gitlab-runner \
  --repo https://helm.sikalabs.io gitlab-runner \
  --namespace gitlab-runner \
  --create-namespace \
  --set gitlabUrl=https://gitlab.sikademo.com \
  --set runnerRegistrationToken=h94VrzQzZnJ_va3hxGGW

Git Tag from CI

tag:
  when: manual
  script:
    - slu git tag-next-calver
    - git remote set-url origin https://token:$COMMIT_TOKEN@$CI_SERVER_HOST/$CI_PROJECT_PATH.git
    - git push origin `slu git get-last-calver`

Gitlab Feature Flags

See example in ondrejsika/golang-examples.

Resources

Docs:

Examples

Thank you! & Questions?

That's it. Do you have any questions? Let's go for a beer!

Do you like the course? Tweet about it <3

Please, tweet something with @ondrejsika. Thanks :)

Ondrej Sika

Next time?

Wanna to go for a beer or do some work together? Just book me :)

Training Sessions

2024-11-21 O2

2024-02-12 Alpiq

2023-06-19

2023-05-15 Tieto

2023-02-27 Siemens