Communities

Writing
Writing
Codidact Meta
Codidact Meta
The Great Outdoors
The Great Outdoors
Photography & Video
Photography & Video
Scientific Speculation
Scientific Speculation
Cooking
Cooking
Electrical Engineering
Electrical Engineering
Judaism
Judaism
Languages & Linguistics
Languages & Linguistics
Software Development
Software Development
Mathematics
Mathematics
Christianity
Christianity
Code Golf
Code Golf
Music
Music
Physics
Physics
Linux Systems
Linux Systems
Power Users
Power Users
Tabletop RPGs
Tabletop RPGs
Community Proposals
Community Proposals
tag:snake search within a tag
answers:0 unanswered questions
user:xxxx search by author id
score:0.5 posts with 0.5+ score
"snake oil" exact phrase
votes:4 posts with 4+ votes
created:<1w created < 1 week ago
post_type:xxxx type of post
Search help
Notifications
Mark all as read See all your notifications »
Q&A

Welcome to Software Development on Codidact!

Will you help us build our independent community of developers helping developers? We're small and trying to grow. We welcome questions about all aspects of software development, from design to code to QA and more. Got questions? Got answers? Got code you'd like someone to review? Please join us.

Keep failed CronJob Pods around in Kubernetes

+1
−0

It seems like pods created by a Kubernetes CronJob (or rather Job) are deleted as soon as they fail. This makes it difficult to figure out what went wrong (unless you have Prometheus or some other log aggregator in use).

Is there a way to keep the pods of the failed Jobs around for a while to facilitate post mortem analysis?

History
Why does this post require attention from curators or moderators?
You might want to add some details to your flag.
Why should this post be closed?

0 comment threads

1 answer

+1
−0

There are a few fields that play into this, but the gist of it is to set the Pod's restartPolicy to Never. This doesn't mean that the Job only tries once. Rather, instead of restarting the container inside the Pod on failure, a new Pod will get scheduled until the Job's backoffLimit is reached (default being 6 at the moment). Finally the CronJob's failedJobsHistoryLimit decides whether the failed Job together with its Pods will be kept around. It defaults to 1.

Here's an example of a debuggable CronJob, adapted from the official example (CC BY 4.0, The Kubernetes Authors):

apiVersion: batch/v1
kind: CronJob
metadata:
  name: hello
spec:
  schedule: "* * * * *"
  jobTemplate:
    spec:
      template:
        spec:
          containers:
            - name: hello
              image: busybox:1.28
              imagePullPolicy: IfNotPresent
              command:
                - /bin/sh
                - -c
                - date; echo Hello from the Kubernetes cluster
            # This is the key bit.
            # Defaults are okay for the other discussed fields.
            restartPolicy: Never
History
Why does this post require attention from curators or moderators?
You might want to add some details to your flag.

0 comment threads

Sign up to answer this question »