Job artifact troubleshooting for administrators
- Tier: Free, Premium, Ultimate
- Offering: GitLab Self-Managed
When administering job artifacts, you might encounter the following issues.
Job artifacts using too much disk space
Job artifacts can fill up your disk space quicker than expected. Some possible reasons are:
- Users have configured job artifacts expiration to be longer than necessary.
- The number of jobs run, and hence artifacts generated, is higher than expected.
- Job logs are larger than expected, and have accumulated over time.
- The file system might run out of inodes because empty directories are left behind by artifact housekeeping. The Rake task for orphaned artifact files removes these.
- Artifact files might be left on disk and not deleted by housekeeping. Run the Rake task for orphaned artifact files to remove these. This script should always find work to do, as it also removes empty directories (see above).
- Artifact housekeeping was changed significantly, and you might need to enable a feature flag to use the updated system.
- The keep latest artifacts from most recent success jobs feature is enabled.
In these and other cases, identify the projects most responsible for disk space usage, figure out what types of artifacts are using the most space, and in some cases, manually delete job artifacts to reclaim disk space.
Artifacts housekeeping
Artifacts housekeeping is the process that identifies which artifacts are expired and can be deleted.
Housekeeping disabled in GitLab 15.0 to 15.2
Artifact housekeeping was significantly improved in GitLab 15.0, introduced behind feature flags disabled by default. The flags were enabled by default in GitLab 15.3.
If artifacts housekeeping does not seem to be working in GitLab 15.0 to GitLab 15.2, you should check if the feature flags are enabled.
To check if the feature flags are enabled:
-
Start a Rails console.
-
Check if the feature flags are enabled.
Feature.enabled?(:ci_detect_wrongly_expired_artifacts) Feature.enabled?(:ci_update_unlocked_job_artifacts) Feature.enabled?(:ci_job_artifacts_backlog_work)
-
If any of the feature flags are disabled, enable them:
Feature.enable(:ci_detect_wrongly_expired_artifacts) Feature.enable(:ci_update_unlocked_job_artifacts) Feature.enable(:ci_job_artifacts_backlog_work)
These changes include switching artifacts from unlocked
to locked
if
they should be retained.
unknown
status
Artifacts with Artifacts created before housekeeping was updated have a status of unknown
. After they expire,
these artifacts are not processed by the new housekeeping.
You can check the database to confirm if your instance has artifacts with the unknown
status:
-
Start a database console:
sudo gitlab-psql
Helm chart (Kubernetes)
# Find the toolbox pod
kubectl --namespace <namespace> get pods -lapp=toolbox
# Connect to the PostgreSQL console
kubectl exec -it <toolbox-pod-name> -- /srv/gitlab/bin/rails dbconsole --include-password --database main
Docker
sudo docker exec -it <container_name> /bin/bash
gitlab-psql
Self-compiled (source)
sudo -u git -H psql -d gitlabhq_production
-
Run the following query:
select expire_at, file_type, locked, count(*) from ci_job_artifacts where expire_at is not null and file_type != 3 group by expire_at, file_type, locked having count(*) > 1;
If records are returned, then there are artifacts which the housekeeping job is unable to process. For example:
expire_at | file_type | locked | count
-------------------------------+-----------+--------+--------
2021-06-21 22:00:00+00 | 1 | 2 | 73614
2021-06-21 22:00:00+00 | 2 | 2 | 73614
2021-06-21 22:00:00+00 | 4 | 2 | 3522
2021-06-21 22:00:00+00 | 9 | 2 | 32
2021-06-21 22:00:00+00 | 12 | 2 | 163
Artifacts with locked status 2
are unknown
. Check
issue #346261
for more details.
unknown
artifacts
Clean up The Sidekiq worker that processes all unknown
artifacts is enabled by default in
GitLab 15.3 and later. It analyzes the artifacts returned by the above database query and
determines which should be locked
or unlocked
. Artifacts are then deleted
by that worker if needed.
The worker can be enabled on GitLab Self-Managed:
-
Start a Rails console.
-
Check if the feature is enabled.
Feature.enabled?(:ci_job_artifacts_backlog_work)
-
Enable the feature, if needed:
Feature.enable(:ci_job_artifacts_backlog_work)
The worker processes 10,000 unknown
artifacts every seven minutes, or roughly two million
in 24 hours.
There is a related ci_job_artifacts_backlog_large_loop_limit
feature flag
which causes the worker to process unknown
artifacts
in batches that are five times larger.
This flag is not recommended for use.
@final
artifacts not deleted from object store
An issue in GitLab 16.1 and 16.2 caused @final
artifacts to not be deleted from object storage.
Administrators of GitLab instances that ran GitLab 16.1 or 16.2 for some time could see an increase in object storage used by artifacts. Follow this procedure to check for and remove these artifacts.
Removing the files is a two stage process:
List orphaned job artifacts
Linux package (Omnibus)
sudo gitlab-rake gitlab:cleanup:list_orphan_job_artifact_final_objects
Docker
docker exec -it <container-id> bash
gitlab-rake gitlab:cleanup:list_orphan_job_artifact_final_objects
Either write to a persistent volume mounted in the container, or when the command completes: copy the output file out of the session.
Self-compiled (source)
sudo -u git -H bundle exec rake gitlab:cleanup:list_orphan_job_artifact_final_objects RAILS_ENV=production
Helm chart (Kubernetes)
# find the pod
kubectl get pods --namespace <namespace> -lapp=toolbox
# open the Rails console
kubectl exec -it -c toolbox <toolbox-pod-name> bash
gitlab-rake gitlab:cleanup:list_orphan_job_artifact_final_objects
When the command complete, copy the file out of the session onto persistent storage.
The Rake task has some additional features that apply to all types of GitLab deployment:
-
Scanning object storage can be interrupted. Progress is recorded in Redis, this is used to resume scanning artifacts from that point.
-
By default, the Rake task generates a CSV file:
/opt/gitlab/embedded/service/gitlab-rails/tmp/orphan_job_artifact_final_objects.csv
-
Set an environment variable to specify a different filename:
# Packaged GitLab sudo su - FILENAME='custom_filename.csv' gitlab-rake gitlab:cleanup:list_orphan_job_artifact_final_objects
-
If the output file exists already (the default, or the specified file) it appends entries to the file.
-
Each row contains the fields
object_path,object_size
comma separated, with no file header. For example:35/13/35135aaa6cc23891b40cb3f378c53a17a1127210ce60e125ccf03efcfdaec458/@final/1a/1a/5abfa4ec66f1cc3b681a4d430b8b04596cbd636f13cdff44277211778f26,201
Delete orphaned job artifacts
Linux package (Omnibus)
sudo gitlab-rake gitlab:cleanup:delete_orphan_job_artifact_final_objects
Docker
docker exec -it <container-id> bash
gitlab-rake gitlab:cleanup:delete_orphan_job_artifact_final_objects
- Copy the output file out of the session when the command completes, or write it to a volume that has been mounted by the container.
Self-compiled (source)
sudo -u git -H bundle exec rake gitlab:cleanup:delete_orphan_job_artifact_final_objects RAILS_ENV=production
Helm chart (Kubernetes)
# find the pod
kubectl get pods --namespace <namespace> -lapp=toolbox
# open the Rails console
kubectl exec -it -c toolbox <toolbox-pod-name> bash
gitlab-rake gitlab:cleanup:delete_orphan_job_artifact_final_objects
- When the command complete, copy the file out of the session onto persistent storage.
The following applies to all types of GitLab deployment:
- Specify the input filename using the
FILENAME
variable. By default the script looks for:/opt/gitlab/embedded/service/gitlab-rails/tmp/orphan_job_artifact_final_objects.csv
- As the script deletes files, it writes out a CSV file with the deleted files:
-
the file is in the same directory as the input file
-
the filename is prefixed with
deleted_from--
. For example:deleted_from--orphan_job_artifact_final_objects.csv
. -
The rows in the file are:
object_path,object_size,object_generation/version
, for example:35/13/35135aaa6cc23891b40cb3f378c53a17a1127210ce60e125ccf03efcfdaec458/@final/1a/1a/5abfa4ec66f1cc3b681a4d430b8b04596cbd636f13cdff44277211778f26,201,1711616743796587
-
List projects and builds with artifacts with a specific expiration (or no expiration)
Using a Rails console, you can find projects that have job artifacts with either:
- No expiration date.
- An expiration date more than 7 days in the future.
Similar to deleting artifacts, use the following example time frames and alter them as needed:
7.days.from_now
10.days.from_now
2.weeks.from_now
3.months.from_now
1.year.from_now
Each of the following scripts also limits the search to 50 results with .limit(50)
, but this number can also be changed as needed:
# Find builds & projects with artifacts that never expire
builds_with_artifacts_that_never_expire = Ci::Build.with_downloadable_artifacts.where(artifacts_expire_at: nil).limit(50)
builds_with_artifacts_that_never_expire.find_each do |build|
puts "Build with id #{build.id} has artifacts that don't expire and belongs to project #{build.project.full_path}"
end
# Find builds & projects with artifacts that expire after 7 days from today
builds_with_artifacts_that_expire_in_a_week = Ci::Build.with_downloadable_artifacts.where('artifacts_expire_at > ?', 7.days.from_now).limit(50)
builds_with_artifacts_that_expire_in_a_week.find_each do |build|
puts "Build with id #{build.id} has artifacts that expire at #{build.artifacts_expire_at} and belongs to project #{build.project.full_path}"
end
List projects by total size of job artifacts stored
List the top 20 projects, sorted by the total size of job artifacts stored, by running the following code in the Rails console:
sudo gitlab-rake gitlab:cleanup:list_orphan_job_artifact_final_objects
```0
You can change the number of projects listed by modifying `.limit(20)` to the
number you want.
### List largest artifacts in a single project
List the 50 largest job artifacts in a single project by running the following
code in the [Rails console](../operations/rails_console.md):
```shell
sudo gitlab-rake gitlab:cleanup:list_orphan_job_artifact_final_objects
```1
You can change the number of job artifacts listed by modifying `.limit(50)` to
the number you want.
### List artifacts in a single project
List the artifacts for a single project, sorted by artifact size. The output includes the:
- ID of the job that created the artifact
- artifact size
- artifact file type
- artifact creation date
- on-disk location of the artifact
```shell
sudo gitlab-rake gitlab:cleanup:list_orphan_job_artifact_final_objects
```2
To change the number of job artifacts listed, change the number in `limit(50)`.
### Delete old builds and artifacts
These commands remove data permanently. Before running them in a production environment,
you should try them in a test environment first and make a backup of the instance
that can be restored if needed.
#### Delete old artifacts for a project
This step also erases artifacts that users have [chosen to keep](../../ci/jobs/job_artifacts.md#with-an-expiry):
```shell
sudo gitlab-rake gitlab:cleanup:list_orphan_job_artifact_final_objects
```3
In [GitLab 15.3 and earlier](https://gitlab.com/gitlab-org/gitlab/-/issues/372537), use the following instead:
```shell
sudo gitlab-rake gitlab:cleanup:list_orphan_job_artifact_final_objects
```4
#### Delete old artifacts instance wide
This step also erases artifacts that users have [chosen to keep](../../ci/jobs/job_artifacts.md#with-an-expiry):
```shell
sudo gitlab-rake gitlab:cleanup:list_orphan_job_artifact_final_objects
```5
In [GitLab 15.3 and earlier](https://gitlab.com/gitlab-org/gitlab/-/issues/372537), use the following instead:
```shell
sudo gitlab-rake gitlab:cleanup:list_orphan_job_artifact_final_objects
```6
#### Delete old job logs and artifacts for a project
```shell
sudo gitlab-rake gitlab:cleanup:list_orphan_job_artifact_final_objects
```7
#### Delete old job logs and artifacts instance wide
```shell
sudo gitlab-rake gitlab:cleanup:list_orphan_job_artifact_final_objects
```8
In [GitLab 15.3 and earlier](https://gitlab.com/gitlab-org/gitlab/-/issues/369132), replace
`Ci::BuildEraseService.new(build, admin_user).execute` with `build.erase(erased_by: admin_user)`.
`1.year.ago` is a Rails [`ActiveSupport::Duration`](https://api.rubyonrails.org/classes/ActiveSupport/Duration.html) method.
Start with a long duration to reduce the risk of accidentally deleting artifacts that are still in use.
Rerun the deletion with shorter durations as needed, for example `3.months.ago`, `2.weeks.ago`, or `7.days.ago`.
The method `erase_erasable_artifacts!` is synchronous, and upon execution the artifacts are immediately removed;
they are not scheduled by a background queue.
### Deleting artifacts does not immediately reclaim disk space
When artifacts are deleted, the process occurs in two phases:
1. **Marked as ready for deletion**: `Ci::JobArtifact` records are removed from the database and
converted to `Ci::DeletedObject` records with a future `pick_up_at` timestamp.
1. **Remove from storage**: The artifact files remain on disk until the `Ci::ScheduleDeleteObjectsCronWorker` worker
processes the `Ci::DeletedObject` records and physically removes the files.
The removal is deliberately limited to prevent overwhelming system resources:
- The worker runs once per hour, at the 16-minute mark.
- It processes objects in batches with a maximum of 20 concurrent jobs.
- Each deleted object has a `pick_up_at` timestamp that determines when it becomes
eligible for physical deletion
For large-scale deletions, the physical cleanup can take a significant amount of time
before disk space is fully reclaimed. Cleanup could take several days for very large deletions.
If you need to reclaim disk space quickly, you can expedite artifact deletion.
#### Expedite artifact removal
If you need to reclaim disk space quickly after deleting a large number of artifacts,
you can bypass the standard scheduling limitations and expedite the deletion process.
These commands put significant load on your system if you are deleting a large number of artifacts.
```shell
sudo gitlab-rake gitlab:cleanup:list_orphan_job_artifact_final_objects
```9
### Delete old pipelines
These commands remove data permanently. Before running them in a production environment,
consider seeking guidance from a Support Engineer. You should also try them in a test environment first
and make a backup of the instance that can be restored if needed.
Deleting a pipeline also removes that pipeline's:
- Job artifacts
- Job logs
- Job metadata
- Pipeline metadata
Removing job and pipeline metadata can help reduce the size of the CI tables in the database.
The CI tables are usually the largest tables in an instance's database.
#### Delete old pipelines for a project
```shell
docker exec -it <container-id> bash
gitlab-rake gitlab:cleanup:list_orphan_job_artifact_final_objects
```0
#### Delete old pipelines instance-wide
```shell
docker exec -it <container-id> bash
gitlab-rake gitlab:cleanup:list_orphan_job_artifact_final_objects
```1
## Job artifact upload fails with error 500
If you are using object storage for artifacts and a job artifact fails to upload,
review:
- The job log for an error message similar to:
```plaintext
WARNING: Uploading artifacts as "archive" to coordinator... failed id=12345 responseStatus=500 Internal Server Error status=500 token=abcd1234
-
The workhorse log for an error message similar to:
{"error":"MissingRegion: could not find region configuration","level":"error","msg":"error uploading S3 session","time":"2021-03-16T22:10:55-04:00"}
In both cases, you might need to add region
to the job artifact object storage configuration.
500 Internal Server Error (Missing file)
Job artifact upload fails with Bucket names that include folder paths are not supported with consolidated object storage.
For example, bucket/path
. If a bucket name has a path in it, you might receive an error similar to:
docker exec -it <container-id> bash
gitlab-rake gitlab:cleanup:list_orphan_job_artifact_final_objects
```2
If a job artifact fails to upload with the above error when using consolidated object storage, make sure you are [using separate buckets](../object_storage.md#use-separate-buckets) for each data type.
## Job artifacts fail to upload with `FATAL: invalid argument` when using Windows mount
If you are using a Windows mount with CIFS for job artifacts, you may see an
`invalid argument` error when the runner attempts to upload artifacts:
```shell
docker exec -it <container-id> bash
gitlab-rake gitlab:cleanup:list_orphan_job_artifact_final_objects
```3
To work around this issue, you can try:
- Switching to an ext4 mount instead of CIFS.
- Upgrading to at least Linux kernel 5.15 which contains a number of important bug fixes
relating to CIFS file leases.
- For older kernels, using the `nolease` mount option to disable file leasing.
For more information, [see the investigation details](https://gitlab.com/gitlab-org/gitlab/-/issues/389995).
## Usage quota shows incorrect artifact storage usage
Sometimes the [artifacts storage usage](../../user/storage_usage_quotas.md) displays an incorrect
value for the total storage space used by artifacts. To recalculate the artifact
usage statistics for all projects in the instance, you can run this background script:
```shell
docker exec -it <container-id> bash
gitlab-rake gitlab:cleanup:list_orphan_job_artifact_final_objects
```4
The `https://example.com/path/file.csv` file must list the project IDs for
all projects for which you want to recalculate artifact storage usage. Use this format for the file:
```shell
docker exec -it <container-id> bash
gitlab-rake gitlab:cleanup:list_orphan_job_artifact_final_objects
```5
The artifact usage value can fluctuate to `0` while the script is running. After
recalculation, usage should display as expected again.
## Artifact download flow diagrams
The following flow diagrams illustrate how job artifacts work. These
diagrams assume object storage is configured for job artifacts.
### Proxy download disabled
With [`proxy_download` set to `false`](../object_storage.md), GitLab
redirects the runner to download artifacts from object storage with a
pre-signed URL. It is usually faster for runners to fetch from the
source directly so this configuration is generally recommended. It
should also reduce bandwidth usage because the data does not have to be
fetched by GitLab and sent to the runner. However, it does require
giving runners direct access to object storage.
The request flow looks like:
```shell
docker exec -it <container-id> bash
gitlab-rake gitlab:cleanup:list_orphan_job_artifact_final_objects
```6
In this diagram:
1. First, the runner attempts to fetch a job artifact by using the
`GET /api/v4/jobs/:id/artifacts` endpoint. The runner attaches the
`direct_download=true` query parameter on the first attempt to indicate
that it is capable of downloading from object storage directly. Direct
downloads can be disabled in the runner configuration via the
[`FF_USE_DIRECT_DOWNLOAD` feature flag](https://docs.gitlab.com/runner/configuration/feature-flags.html).
This flag is set to `true` by default.
1. The runner sends the GET request using HTTP Basic Authentication
with the `gitlab-ci-token` username and an auto-generated
CI/CD job token as the password. This token is generated by GitLab and
given to the runner at the start of a job.
1. The GET request gets passed to the GitLab API, which looks
up the token in the database and finds the user who triggered the job.
1. In steps 5-8:
- If the user has access to the build, then GitLab generates
a presigned URL and sends a 302 Redirect with the `Location` set to that
URL. The runner follows the 302 Redirect and downloads the artifacts.
- If the job cannot be found or the user does not have access to the job,
then the API returns 401 Unauthorized.
The runner does not retry if it receives the following HTTP status codes:
- 200 OK
- 401 Unauthorized
- 403 Forbidden
- 404 Not Found
However, if the runner receives any other status code, such as a 500 error,
it re-attempts to download the artifacts two more times, sleeping 1 second
between each attempt. The subsequent attempts omit `direct_download=true`.
### Proxy download enabled
If `proxy_download` is `true`, GitLab always fetches the
artifacts from object storage and send the data to the runner, even if
the runner sends the `direct_download=true` query parameter. Proxy
downloads might be desirable if runners have restricted network access.
The following diagram is similar to the disabled proxy download example,
except at steps 6-9, GitLab does not send a 302 Redirect to the
runner. Instead, GitLab instructs Workhorse to fetch the data and stream
it back to the runner. From the runner perspective, the original GET
request to `/api/v4/jobs/:id/artifacts` returns the binary data
directly.
```shell
docker exec -it <container-id> bash
gitlab-rake gitlab:cleanup:list_orphan_job_artifact_final_objects
```7
## `413 Request Entity Too Large` error
If the artifacts are too large, the job might fail with the following error:
```shell
docker exec -it <container-id> bash
gitlab-rake gitlab:cleanup:list_orphan_job_artifact_final_objects
```8
You might need to:
- Increase the [maximum artifacts size](../settings/continuous_integration.md#maximum-artifacts-size).
- If you are using NGINX as a proxy server, increase the file upload size limit which is limited to 1 MB by default.
Set a higher value for `client-max-body-size` in the NGINX configuration file.