You can use GitLab CI/CD with Docker to create Docker images. For example, you can create a Docker image of your application, test it, and push it to a container registry.
To run Docker commands in your CI/CD jobs, you must configure GitLab Runner to support docker
commands. This method requires privileged
mode.
If you want to build Docker images without enabling privileged
mode on the runner, you can use a Docker alternative.
To enable Docker commands for your CI/CD jobs, you can use:
Use the shell executorTo include Docker commands in your CI/CD jobs, you can configure your runner to use the shell
executor. In this configuration, the gitlab-runner
user runs the Docker commands, but needs permission to do so.
Install GitLab Runner.
Register a runner. Select the shell
executor. For example:
sudo gitlab-runner register -n \
--url "https://gitlab.com/" \
--registration-token REGISTRATION_TOKEN \
--executor shell \
--description "My Runner"
On the server where GitLab Runner is installed, install Docker Engine. View a list of supported platforms.
Add the gitlab-runner
user to the docker
group:
sudo usermod -aG docker gitlab-runner
Verify that gitlab-runner
has access to Docker:
sudo -u gitlab-runner -H docker info
In GitLab, add docker info
to .gitlab-ci.yml
to verify that Docker is working:
default:
before_script:
- docker info
build_image:
script:
- docker build -t my-docker-image .
- docker run my-docker-image /script/to/run/tests
You can now use docker
commands (and install Docker Compose if needed).
When you add gitlab-runner
to the docker
group, you effectively grant gitlab-runner
full root permissions. For more information, see security of the docker
group.
“Docker-in-Docker” (dind
) means:
The Docker image includes all of the docker
tools and can run the job script in context of the image in privileged mode.
You should use Docker-in-Docker with TLS enabled, which is supported by GitLab.com instance runners.
You should always pin a specific version of the image, like docker:24.0.5
. If you use a tag like docker:latest
, you have no control over which version is used. This can cause incompatibility problems when new versions are released.
You can use the Docker executor to run jobs in a Docker container.
Docker-in-Docker with TLS enabled in the Docker executorThe Docker daemon supports connections over TLS. TLS is the default in Docker 19.03.12 and later.
This task enables --docker-privileged
, which effectively disables the container’s security mechanisms and exposes your host to privilege escalation. This action can cause container breakout. For more information, see runtime privilege and Linux capabilities.
To use Docker-in-Docker with TLS enabled:
Install GitLab Runner.
Register GitLab Runner from the command line. Use docker
and privileged
mode:
sudo gitlab-runner register -n \
--url "https://gitlab.com/" \
--registration-token REGISTRATION_TOKEN \
--executor docker \
--description "My Docker Runner" \
--tag-list "tls-docker-runner" \
--docker-image "docker:24.0.5" \
--docker-privileged \
--docker-volumes "/certs/client"
docker:24.0.5
image (if none is specified at the job level). To start the build and service containers, it uses the privileged
mode. If you want to use Docker-in-Docker, you must always use privileged = true
in your Docker containers./certs/client
for the service and build container, which is needed for the Docker client to use the certificates in that directory. For more information, see the Docker image documentation.The previous command creates a config.toml
entry similar to the following example:
[[runners]]
url = "https://gitlab.com/"
token = TOKEN
executor = "docker"
[runners.docker]
tls_verify = false
image = "docker:24.0.5"
privileged = true
disable_cache = false
volumes = ["/certs/client", "/cache"]
[runners.cache]
[runners.cache.s3]
[runners.cache.gcs]
You can now use docker
in the job script. You should include the docker:24.0.5-dind
service:
default:
image: docker:24.0.5
services:
- docker:24.0.5-dind
before_script:
- docker info
variables:
# When you use the dind service, you must instruct Docker to talk with
# the daemon started inside of the service. The daemon is available
# with a network connection instead of the default
# /var/run/docker.sock socket. Docker 19.03 does this automatically
# by setting the DOCKER_HOST in
# https://github.com/docker-library/docker/blob/d45051476babc297257df490d22cbd806f1b11e4/19.03/docker-entrypoint.sh#L23-L29
#
# The 'docker' hostname is the alias of the service container as described at
# https://docs.gitlab.com/ee/ci/services/#accessing-the-services.
#
# Specify to Docker where to create the certificates. Docker
# creates them automatically on boot, and creates
# `/certs/client` to share between the service and job
# container, thanks to volume mount from config.toml
DOCKER_TLS_CERTDIR: "/certs"
build:
stage: build
tags:
- tls-docker-runner
script:
- docker build -t my-docker-image .
- docker run my-docker-image /script/to/run/tests
Directories defined in volumes = ["/certs/client", "/cache"]
in the Docker-in-Docker with TLS enabled in the Docker executor approach are persistent between builds. If multiple CI/CD jobs using a Docker executor runner have Docker-in-Docker services enabled, then each job writes to the directory path. This approach might result in a conflict.
To address this conflict, use a Unix socket on a volume shared between the Docker-in-Docker service and the build container. This approach improves performance and establishes a secure connection between the service and client.
The following is a sample config.toml
with temporary volume shared between build and service containers:
[[runners]]
url = "https://gitlab.com/"
token = TOKEN
executor = "docker"
[runners.docker]
image = "docker:24.0.5"
privileged = true
volumes = ["/runner/services/docker"] # Temporary volume shared between build and service containers.
The Docker-in-Docker service creates a docker.sock
. The Docker client connects to docker.sock
through a Docker Unix socket volume.
job:
variables:
# This variable is shared by both the DinD service and Docker client.
# For the service, it will instruct DinD to create `docker.sock` here.
# For the client, it tells the Docker client which Docker Unix socket to connect to.
DOCKER_HOST: "unix:///runner/services/docker/docker.sock"
services:
- docker:24.0.5-dind
image: docker:24.0.5
script:
- docker version
Docker-in-Docker with TLS disabled in the Docker executor
Sometimes there are legitimate reasons to disable TLS. For example, you have no control over the GitLab Runner configuration that you are using.
Register GitLab Runner from command line. Use docker
and privileged
mode:
sudo gitlab-runner register -n \
--url "https://gitlab.com/" \
--registration-token REGISTRATION_TOKEN \
--executor docker \
--description "My Docker Runner" \
--tag-list "no-tls-docker-runner" \
--docker-image "docker:24.0.5" \
--docker-privileged
The previous command creates a config.toml
entry similar to the following example:
[[runners]]
url = "https://gitlab.com/"
token = TOKEN
executor = "docker"
[runners.docker]
tls_verify = false
image = "docker:24.0.5"
privileged = true
disable_cache = false
volumes = ["/cache"]
[runners.cache]
[runners.cache.s3]
[runners.cache.gcs]
Include the docker:24.0.5-dind
service in the job script:
default:
image: docker:24.0.5
services:
- docker:24.0.5-dind
before_script:
- docker info
variables:
# When using dind service, you must instruct docker to talk with the
# daemon started inside of the service. The daemon is available with
# a network connection instead of the default /var/run/docker.sock socket.
#
# The 'docker' hostname is the alias of the service container as described at
# https://docs.gitlab.com/ee/ci/docker/using_docker_images.html#accessing-the-services
#
DOCKER_HOST: tcp://docker:2375
#
# This instructs Docker not to start over TLS.
DOCKER_TLS_CERTDIR: ""
build:
stage: build
tags:
- no-tls-docker-runner
script:
- docker build -t my-docker-image .
- docker run my-docker-image /script/to/run/tests
You might need to configure proxy settings to use the docker push
command.
For more information, see Proxy settings when using dind service.
Use the Kubernetes executor with Docker-in-DockerYou can use the Kubernetes executor to run jobs in a Docker container.
Docker-in-Docker with TLS enabled in KubernetesTo use Docker-in-Docker with TLS enabled in Kubernetes:
Using the Helm chart, update the values.yml
file to specify a volume mount.
runners:
tags: "tls-dind-kubernetes-runner"
config: |
[[runners]]
[runners.kubernetes]
image = "ubuntu:20.04"
privileged = true
[[runners.kubernetes.volumes.empty_dir]]
name = "docker-certs"
mount_path = "/certs/client"
medium = "Memory"
Include the docker:24.0.5-dind
service in the job:
default:
image: docker:24.0.5
services:
- name: docker:24.0.5-dind
variables:
HEALTHCHECK_TCP_PORT: "2376"
before_script:
- docker info
variables:
# When using dind service, you must instruct Docker to talk with
# the daemon started inside of the service. The daemon is available
# with a network connection instead of the default
# /var/run/docker.sock socket.
DOCKER_HOST: tcp://docker:2376
#
# The 'docker' hostname is the alias of the service container as described at
# https://docs.gitlab.com/ee/ci/services/#accessing-the-services.
#
# Specify to Docker where to create the certificates. Docker
# creates them automatically on boot, and creates
# `/certs/client` to share between the service and job
# container, thanks to volume mount from config.toml
DOCKER_TLS_CERTDIR: "/certs"
# These are usually specified by the entrypoint, however the
# Kubernetes executor doesn't run entrypoints
# https://gitlab.com/gitlab-org/gitlab-runner/-/issues/4125
DOCKER_TLS_VERIFY: 1
DOCKER_CERT_PATH: "$DOCKER_TLS_CERTDIR/client"
build:
stage: build
tags:
- tls-dind-kubernetes-runner
script:
- docker build -t my-docker-image .
- docker run my-docker-image /script/to/run/tests
To use Docker-in-Docker with TLS disabled in Kubernetes, you must adapt the previous example to:
[[runners.kubernetes.volumes.empty_dir]]
section from the values.yml
file.2376
to 2375
with DOCKER_HOST: tcp://docker:2375
.DOCKER_TLS_CERTDIR: ""
.For example:
Using the Helm chart, update the values.yml
file:
runners:
tags: "no-tls-dind-kubernetes-runner"
config: |
[[runners]]
[runners.kubernetes]
image = "ubuntu:20.04"
privileged = true
You can now use docker
in the job script. You should include the docker:24.0.5-dind
service:
default:
image: docker:24.0.5
services:
- name: docker:24.0.5-dind
variables:
HEALTHCHECK_TCP_PORT: "2375"
before_script:
- docker info
variables:
# When using dind service, you must instruct Docker to talk with
# the daemon started inside of the service. The daemon is available
# with a network connection instead of the default
# /var/run/docker.sock socket.
DOCKER_HOST: tcp://docker:2375
#
# The 'docker' hostname is the alias of the service container as described at
# https://docs.gitlab.com/ee/ci/services/#accessing-the-services.
#
# This instructs Docker not to start over TLS.
DOCKER_TLS_CERTDIR: ""
build:
stage: build
tags:
- no-tls-dind-kubernetes-runner
script:
- docker build -t my-docker-image .
- docker run my-docker-image /script/to/run/tests
Docker-in-Docker is the recommended configuration, but you should be aware of the following issues:
The docker-compose
command: This command is not available in this configuration by default. To use docker-compose
in your job scripts, follow the Docker Compose installation instructions.
Cache: Each job runs in a new environment. Because every build gets its own instance of the Docker engine, concurrent jobs do not cause conflicts. However, jobs can be slower because there’s no caching of layers. See Docker layer caching.
Storage drivers: By default, earlier versions of Docker use the vfs
storage driver, which copies the file system for each job. Docker 17.09 and later use --storage-driver overlay2
, which is the recommended storage driver. See Using the OverlayFS driver for details.
Root file system: Because the docker:24.0.5-dind
container and the runner container do not share their root file system, you can use the job’s working directory as a mount point for child containers. For example, if you have files you want to share with a child container, you could create a subdirectory under /builds/$CI_PROJECT_PATH
and use it as your mount point. For a more detailed explanation, see issue #41227.
variables:
MOUNT_POINT: /builds/$CI_PROJECT_PATH/mnt
script:
- mkdir -p "$MOUNT_POINT"
- docker run -v "$MOUNT_POINT:/mnt" my-docker-image
To use Docker commands in your CI/CD jobs, you can bind-mount /var/run/docker.sock
into the build container. Docker is then available in the context of the image.
If you bind the Docker socket you can’t use docker:24.0.5-dind
as a service. Volume bindings also affect services, making them incompatible.
To mount the Docker socket with the Docker executor, add "/var/run/docker.sock:/var/run/docker.sock"
to the Volumes in the [runners.docker]
section.
To mount /var/run/docker.sock
while registering your runner, include the following options:
sudo gitlab-runner register \
--non-interactive \
--url "https://gitlab.com/" \
--registration-token REGISTRATION_TOKEN \
--executor "docker" \
--description "docker-runner" \
--tag-list "socket-binding-docker-runner" \
--docker-image "docker:24.0.5" \
--docker-volumes "/var/run/docker.sock:/var/run/docker.sock"
The previous command creates a config.toml
entry similar to the following example:
[[runners]]
url = "https://gitlab.com/"
token = RUNNER_TOKEN
executor = "docker"
[runners.docker]
tls_verify = false
image = "docker:24.0.5"
privileged = false
disable_cache = false
volumes = ["/var/run/docker.sock:/var/run/docker.sock", "/cache"]
[runners.cache]
Insecure = false
Use Docker in the job script:
default:
image: docker:24.0.5
before_script:
- docker info
build:
stage: build
tags:
- socket-binding-docker-runner
script:
- docker build -t my-docker-image .
- docker run my-docker-image /script/to/run/tests
To mount the Docker socket with the Kubernetes executor, add "/var/run/docker.sock"
to the Volumes in the [[runners.kubernetes.volumes.host_path]]
section.
To specify a volume mount, update the values.yml
file by using Helm chart.
runners:
tags: "socket-binding-kubernetes-runner"
config: |
[[runners]]
[runners.kubernetes]
image = "ubuntu:20.04"
privileged = false
[runners.kubernetes]
[[runners.kubernetes.volumes.host_path]]
host_path = '/var/run/docker.sock'
mount_path = '/var/run/docker.sock'
name = 'docker-sock'
read_only = true
Use Docker in the job script:
default:
image: docker:24.0.5
before_script:
- docker info
build:
stage: build
tags:
- socket-binding-kubernetes-runner
script:
- docker build -t my-docker-image .
- docker run my-docker-image /script/to/run/tests
When you use Docker socket binding, you avoid running Docker in privileged mode. However, the implications of this method are:
When you share the Docker daemon, you effectively disable the container’s security mechanisms and expose your host to privilege escalation. This can cause container breakout. For example, if you run docker rm -f $(docker ps -a -q)
in a project, it removes the GitLab Runner containers.
Concurrent jobs might not work. If your tests create containers with specific names, they might conflict with each other.
Any containers created by Docker commands are siblings of the runner, rather than children of the runner. This might cause complications for your workflow.
Sharing files and directories from the source repository into containers might not work as expected. Volume mounting is done in the context of the host machine, not the build container. For example:
docker run --rm -t -i -v $(pwd)/src:/home/app/src test-image:latest run_app_tests
You do not need to include the docker:24.0.5-dind
service, like you do when you use the Docker-in-Docker executor:
default:
image: docker:24.0.5
before_script:
- docker info
build:
stage: build
script:
- docker build -t my-docker-image .
- docker run my-docker-image /script/to/run/tests
For complex Docker-in-Docker setups like Code Quality scanning using CodeClimate, you must match host and container paths for proper execution. For more details, see Use private runners for CodeClimate-based scanning.
Use Docker pipe bindingWindows Containers run Windows executables compiled for the Windows Server kernel and userland (either windowsservercore or nanoserver). To build and run Windows containers, a Windows system with container support is required. For more information, see Windows Containers.
To use Docker pipe binding, you must install and run a Docker Engine on the host Windows Server operating system. For more information, see Install Docker Community Edition (CE) on Windows Server.
To use Docker commands in your Windows-based container CI/CD jobs, you can bind-mount \\.\pipe\docker_engine
into the launched executor container. Docker is then available in the context of the image.
The Docker pipe binding in Windows is similar to Docker socket binding in Linux and have similar Known issues as Known issues with Docker socket binding.
A mandatory prerequisite for usage of Docker pipe binding is a Docker Engine installed and running on the host Windows Server operating system. See: Install Docker Community Edition (CE) on Windows Server
Use the Docker executor with Docker pipe bindingYou can use the Docker executor to run jobs in a Windows-based container.
To mount the Docker pipe with the Docker executor, add "\\.\pipe\docker_engine:\\.\pipe\docker_engine"
to the Volumes in the [runners.docker]
section.
To mount "\\.\pipe\docker_engine
while registering your runner, include the following options:
.\gitlab-runner.exe register \
--non-interactive \
--url "https://gitlab.com/" \
--registration-token REGISTRATION_TOKEN \
--executor "docker-windows" \
--description "docker-windows-runner"
--tag-list "docker-windows-runner" \
--docker-image "docker:25-windowsservercore-ltsc2022" \
--docker-volumes "\\.\pipe\docker_engine:\\.\pipe\docker_engine"
The previous command creates a config.toml
entry similar to the following example:
[[runners]]
url = "https://gitlab.com/"
token = RUNNER_TOKEN
executor = "docker-windows"
[runners.docker]
tls_verify = false
image = "docker:25-windowsservercore-ltsc2022"
privileged = false
disable_cache = false
volumes = ["\\.\pipe\docker_engine:\\.\pipe\docker_engine"]
[runners.cache]
Insecure = false
Use Docker in the job script:
default:
image: docker:25-windowsservercore-ltsc2022
before_script:
- docker version
- docker info
build:
stage: build
tags:
- docker-windows-runner
script:
- docker build -t my-docker-image .
- docker run my-docker-image /script/to/run/tests
You can use the Kubernetes executor to run jobs in a Windows-based container.
To use Kubernetes executor for Windows-based containers, you must include Windows nodes in your Kubernetes cluster. For more information, see Windows containers in Kubernetes.
You can use Runner operating in a Linux environment but targeting Windows nodes
To mount the Docker pipe with the Kubernetes executor, add "\\.\pipe\docker_engine"
to the Volumes in the [[runners.kubernetes.volumes.host_path]]
section.
To specify a volume mount, update the values.yml
file by using Helm chart.
runners:
tags: "kubernetes-windows-runner"
config: |
[[runners]]
executor = "kubernetes"
# The FF_USE_POWERSHELL_PATH_RESOLVER feature flag has to be enabled for PowerShell
# to resolve paths for Windows correctly when Runner is operating in a Linux environment
# but targeting Windows nodes.
[runners.feature_flags]
FF_USE_POWERSHELL_PATH_RESOLVER = true
[runners.kubernetes]
[[runners.kubernetes.volumes.host_path]]
host_path = '\\.\pipe\docker_engine'
mount_path = '\\.\pipe\docker_engine'
name = 'docker-pipe'
read_only = true
[runners.kubernetes.node_selector]
"kubernetes.io/arch" = "amd64"
"kubernetes.io/os" = "windows"
"node.kubernetes.io/windows-build" = "10.0.20348"
Use Docker in the job script:
default:
image: docker:25-windowsservercore-ltsc2022
before_script:
- docker version
- docker info
build:
stage: build
tags:
- kubernetes-windows-runner
script:
- docker build -t my-docker-image .
- docker run my-docker-image /script/to/run/tests
When you migrate from dockerd
to containerd
, the AWS EKS bootstrapping script Start-EKSBootstrap.ps1
stops and disables the Docker Service. To work around this issue, rename the Docker Service after you Install Docker Community Edition (CE) on Windows Server with this script:
Write-Output "Rename the just installed Docker Engine Service from docker to dockerd"
Write-Output "because the Start-EKSBootstrap.ps1 stops and disables the docker Service as part of migration from dockerd to containerd"
Stop-Service -Name docker
dockerd --register-service --service-name dockerd
Start-Service -Name dockerd
Write-Output "Ready to do Docker pipe binding on Windows EKS Node! :-)"
Known issues with Docker pipe binding
Docker pipe binding has the same set of security and isolation issues as the Known issues with Docker socket binding.
Enable registry mirror fordocker:dind
service
When the Docker daemon starts inside the service container, it uses the default configuration. You might want to configure a registry mirror for performance improvements and to ensure you do not exceed Docker Hub rate limits.
The service in the.gitlab-ci.yml
file
You can append extra CLI flags to the dind
service to set the registry mirror:
services:
- name: docker:24.0.5-dind
command: ["--registry-mirror", "https://registry-mirror.example.com"] # Specify the registry mirror to use
The service in the GitLab Runner configuration file
If you are a GitLab Runner administrator, you can specify the command
to configure the registry mirror for the Docker daemon. The dind
service must be defined for the Docker or Kubernetes executor.
Docker:
[[runners]]
...
executor = "docker"
[runners.docker]
...
privileged = true
[[runners.docker.services]]
name = "docker:24.0.5-dind"
command = ["--registry-mirror", "https://registry-mirror.example.com"]
Kubernetes:
[[runners]]
...
name = "kubernetes"
[runners.kubernetes]
...
privileged = true
[[runners.kubernetes.services]]
name = "docker:24.0.5-dind"
command = ["--registry-mirror", "https://registry-mirror.example.com"]
The Docker executor in the GitLab Runner configuration file
If you are a GitLab Runner administrator, you can use the mirror for every dind
service. Update the configuration to specify a volume mount.
For example, if you have a /opt/docker/daemon.json
file with the following content:
{
"registry-mirrors": [
"https://registry-mirror.example.com"
]
}
Update the config.toml
file to mount the file to /etc/docker/daemon.json
. This mounts the file for every container created by GitLab Runner. The configuration is detected by the dind
service.
[[runners]]
...
executor = "docker"
[runners.docker]
image = "alpine:3.12"
privileged = true
volumes = ["/opt/docker/daemon.json:/etc/docker/daemon.json:ro"]
The Kubernetes executor in the GitLab Runner configuration file
If you are a GitLab Runner administrator, you can use the mirror for every dind
service. Update the configuration to specify a ConfigMap volume mount.
For example, if you have a /tmp/daemon.json
file with the following content:
{
"registry-mirrors": [
"https://registry-mirror.example.com"
]
}
Create a ConfigMap with the content of this file. You can do this with a command like:
kubectl create configmap docker-daemon --namespace gitlab-runner --from-file /tmp/daemon.json
You must use the namespace that the Kubernetes executor for GitLab Runner uses to create job pods.
After the ConfigMap is created, you can update the config.toml
file to mount the file to /etc/docker/daemon.json
. This update mounts the file for every container created by GitLab Runner. The dind
service detects this configuration.
[[runners]]
...
executor = "kubernetes"
[runners.kubernetes]
image = "alpine:3.12"
privileged = true
[[runners.kubernetes.volumes.config_map]]
name = "docker-daemon"
mount_path = "/etc/docker/daemon.json"
sub_path = "daemon.json"
Authenticate with registry in Docker-in-Docker
When you use Docker-in-Docker, the standard authentication methods do not work, because a fresh Docker daemon is started with the service. You should authenticate with registry.
Make Docker-in-Docker builds faster with Docker layer cachingWhen using Docker-in-Docker, Docker downloads all layers of your image every time you create a build. You can make your builds faster with Docker layer caching.
Use the OverlayFS driverThe instance runners on GitLab.com use the overlay2
driver by default.
By default, when using docker:dind
, Docker uses the vfs
storage driver, which copies the file system on every run. You can avoid this disk-intensive operation by using a different driver, for example overlay2
.
Ensure a recent kernel is used, preferably >= 4.2
.
Check whether the overlay
module is loaded:
sudo lsmod | grep overlay
If you see no result, then the module is not loaded. To load the module, use:
If the module loaded, you must make sure the module loads on reboot. On Ubuntu systems, do this by adding the following line to /etc/modules
:
You can enable the driver for each project individually by using the DOCKER_DRIVER
CI/CD variable in .gitlab-ci.yml
:
variables:
DOCKER_DRIVER: overlay2
Use the OverlayFS driver for every project
If you use your own runners, you can enable the driver for every project by setting the DOCKER_DRIVER
environment variable in the [[runners]]
section of the config.toml
file:
environment = ["DOCKER_DRIVER=overlay2"]
If you’re running multiple runners, you must modify all configuration files.
Read more about the runner configuration and using the OverlayFS storage driver.
Docker alternativesYou can build container images without enabling privileged mode on your runner:
To use Buildah with GitLab CI/CD, you need a runner with one of the following executors:
In this example, you use Buildah to:
In the last step, Buildah uses the Dockerfile
under the root directory of the project to build the Docker image. Finally, it pushes the image to the project’s container registry:
build:
stage: build
image: quay.io/buildah/stable
variables:
# Use vfs with buildah. Docker offers overlayfs as a default, but Buildah
# cannot stack overlayfs on top of another overlayfs filesystem.
STORAGE_DRIVER: vfs
# Write all image metadata in the docker format, not the standard OCI format.
# Newer versions of docker can handle the OCI format, but older versions, like
# the one shipped with Fedora 30, cannot handle the format.
BUILDAH_FORMAT: docker
FQ_IMAGE_NAME: "$CI_REGISTRY_IMAGE/test"
before_script:
# GitLab container registry credentials taken from the
# [predefined CI/CD variables](../variables/_index.md#predefined-cicd-variables)
# to authenticate to the registry.
- echo "$CI_REGISTRY_PASSWORD" | buildah login -u "$CI_REGISTRY_USER" --password-stdin $CI_REGISTRY
script:
- buildah images
- buildah build -t $FQ_IMAGE_NAME
- buildah images
- buildah push $FQ_IMAGE_NAME
If you are using GitLab Runner Operator deployed to an OpenShift cluster, try the tutorial for using Buildah to build images in rootless container.
Use the GitLab container registryAfter you’ve built a Docker image, you can push it to the GitLab container registry.
Troubleshootingopen //./pipe/docker_engine: The system cannot find the file specified
The following error might appear when you run a docker
command in the PowerShell script to access the mounted Docker pipe:
PS C:\> docker version
Client:
Version: 25.0.5
API version: 1.44
Go version: go1.21.8
Git commit: 5dc9bcc
Built: Tue Mar 19 15:06:12 2024
OS/Arch: windows/amd64
Context: default
error during connect: this error may indicate that the docker daemon is not running: Get "http://%2F%2F.%2Fpipe%2Fdocker_engine/v1.44/version": open //./pipe/docker_engine: The system cannot find the file specified.
The error indicates that the Docker Engine is not running on the Windows EKS Node and the Docker pipe binding could not be used in the Windows-based Executor container.
To solve the problem, use the workaround described in Use the Kubernetes executor with Docker pipe binding.
RetroSearch is an open source project built by @garambo | Open a GitHub Issue
Search and Browse the WWW like it's 1997 | Search results from DuckDuckGo
HTML:
3.2
| Encoding:
UTF-8
| Version:
0.7.4