Replies: 3 comments 9 replies
-
@toast-gear I see the pods are again in terminating state today, even though the autoscaling is removed. I see this in action-controller manager logs 2022-10-26T08:19:17Z INFO actions-runner-controller.runnerpod Runner pod is annotated to wait for completion, and the runner container is not restarting {"runnerpod": "actions-runner/mule-runner-new-bqg48-jqgnp"} |
Beta Was this translation helpful? Give feedback.
-
I can't read your log due to no indentation and no formatting. please post it in code block by using ```(trible backticks) as follows apiVersion: actions.summerwind.dev/v1alpha1
kind: RunnerDeployment
metadata:
name: example-runnerdeploy
spec:
replicas: 1
template:
spec:
repository: mumoshu/actions-runner-controller-ci |
Beta Was this translation helpful? Give feedback.
-
We're locking this discussion because it has not had recent activity and/or other members have asked for more information to assist you but received no response. Thank you for helping us maintain a productive and tidy community for all our members. |
Beta Was this translation helpful? Give feedback.
-
Hi Team,
I am using ARC, with two running deployments one for java apps and another for mulesoft apps. Whenever we autoscaling happens only the mulesoft runners are going into terminating state. I have read the discussions on pods getting stuck in terminating state but am not able to resolve my issue. Hope you can guide me here.
I am using the latest version ARC. 1.26
My runnerdeployment:
apiVersion: actions.summerwind.dev/v1alpha1
kind: RunnerDeployment
metadata:
name: mule-runner
namespace: actions-runner
spec:
replicas: 6
template:
spec:
organization: my-org
image: custom-image-here
labels:
- mule-runner
kubectl get pod -o yaml "runnerpod" -n namespace
apiVersion: v1
kind: Pod
metadata:
annotations:
actions-runner-controller/token-expires-at: "2022-10-19T17:23:42Z"
actions-runner/id: "2007"
actions-runner/runner-completion-wait-start-timestamp: "2022-10-20T16:13:27Z"
actions-runner/unregistration-failure-message: Bad request - Runner "mule-runner-h6chf-9s5d9"
is still running a job"
actions-runner/unregistration-start-timestamp: "2022-10-20T16:13:27Z"
kubernetes.io/psp: eks.privileged
sync-time: "2022-10-19T16:23:42Z"
creationTimestamp: "2022-10-19T16:23:42Z"
deletionGracePeriodSeconds: 0
deletionTimestamp: "2022-10-20T16:12:59Z"
finalizers:
labels:
actions-runner: ""
actions-runner-controller/inject-registration-token: "true"
pod-template-hash: 6c5fd57df9
runner-deployment-name: mule-runner
runner-template-hash: b554d6c8
name: mule-runner-h6chf-9s5d9
namespace: actions-runner
ownerReferences:
blockOwnerDeletion: true
controller: true
kind: Runner
name: mule-runner-h6chf-9s5d9
uid: 3286b78e-d6fd-417f-bc33-6b2caa9af704
resourceVersion: "18492717"
uid: b2bab42c-b3e6-4d02-9a91-cdb809e25bd3
spec:
containers:
value: my-org
value: mule-runner
value: "true"
value: "false"
value: https://github.com/
value: /runner/_work
value: "true"
value: "false"
value: tcp://localhost:2376
value: "1"
value: /certs/client
value: mule-runner-h6chf-9s5d9
value: A3IXUG6DBAMH35N6FNTWT73DKAZJ5AVPNFXHG5DBNRWGC5DJN5XF62LEZYA4HTDDWFUW443UMFWGYYLUNFXW4X3UPFYGLN2JNZ2GKZ3SMF2GS33OJFXHG5DBNRWGC5DJN5XA
image: my-cutom-image
imagePullPolicy: Always
name: runner
resources: {}
securityContext:
privileged: false
terminationMessagePath: /dev/termination-log
terminationMessagePolicy: File
volumeMounts:
name: runner
name: work
name: certs-client
readOnly: true
name: kube-api-access-m6v46
readOnly: true
value: /certs
image: docker:dind
imagePullPolicy: IfNotPresent
name: docker
resources: {}
securityContext:
privileged: true
terminationMessagePath: /dev/termination-log
terminationMessagePolicy: File
volumeMounts:
name: runner
name: certs-client
name: work
name: kube-api-access-m6v46
readOnly: true
dnsPolicy: ClusterFirst
enableServiceLinks: true
nodeName: ip-xxxx-us-east-2.compute.internal
preemptionPolicy: PreemptLowerPriority
priority: 0
restartPolicy: Never
schedulerName: default-scheduler
securityContext: {}
serviceAccount: default
serviceAccountName: default
terminationGracePeriodSeconds: 30
tolerations:
key: node.kubernetes.io/not-ready
operator: Exists
tolerationSeconds: 300
key: node.kubernetes.io/unreachable
operator: Exists
tolerationSeconds: 300
volumes:
name: runner
name: work
name: certs-client
projected:
defaultMode: 420
sources:
expirationSeconds: 3607
path: token
items:
path: ca.crt
name: kube-root-ca.crt
items:
apiVersion: v1
fieldPath: metadata.namespace
path: namespace
status:
conditions:
lastTransitionTime: "2022-10-19T16:35:52Z"
status: "True"
type: Initialized
lastTransitionTime: "2022-10-20T16:07:53Z"
status: "False"
type: Ready
lastTransitionTime: "2022-10-19T16:35:59Z"
status: "True"
type: ContainersReady
lastTransitionTime: "2022-10-19T16:35:52Z"
status: "True"
type: PodScheduled
containerStatuses:
image: docker:dind
imageID: docker-pullable://docker@sha256:999fc127a51b8a86593ff9ba2518f14cbd18555849f8927fd56fa82395effe16
lastState: {}
name: docker
ready: true
restartCount: 0
started: true
state:
running:
startedAt: "2022-10-19T16:35:56Z"
image: 854311749298.dkr.ecr.us-east-2.amazonaws.com/jb-autoscale-runner-mule-image:dev-0.217.0
imageID: docker-pullable://854311749298.dkr.ecr.us-east-2.amazonaws.com/jb-autoscale-runner-mule-image@sha256:12a31f0f7214f8788670b5975edea48f9159f637fe068d2b6c4ca3f8c36ca756
lastState: {}
name: runner
ready: true
restartCount: 0
started: true
state:
running:
startedAt: "2022-10-19T16:35:56Z"
hostIP: xxxx
phase: Running
podIP: xxxx
podIPs:
qosClass: BestEffort
startTime: "2022-10-19T16:35:52Z"
I see the pods are in terminating state. Please suggest what is wrong with my deployment and am not able to get any logs since they are in terminating state.
The action runner controller log:
[11:30] Sabhika Roofi
2022-10-25T06:00:43Z INFO actions-runner-controller.runnerpod Runner pod is annotated to wait for completion, and the runner container is not restarting {"runnerpod": "actions-runner/mule-runner-g9x6v-2qq7s"}2022-10-25T06:00:43Z INFO actions-runner-controller.runnerpod Runner pod is annotated to wait for completion, and the runner container is not restarting {"runnerpod": "actions-runner/mule-runner-g9x6v-h2sqj"}2022-10-25T06:00:43Z INFO actions-runner-controller.runnerpod Runner pod is annotated to wait for completion, and the runner container is not restarting {"runnerpod": "actions-runner/mule-runner-h6chf-gn4pf"}2022-10-25T06:00:43Z INFO actions-runner-controller.runnerpod Runner pod is annotated to wait for completion, and the runner container is not restarting {"runnerpod": "actions-runner/mule-runner-h6chf-xbdxf"}2022-10-25T06:00:43Z INFO actions-runner-controller.runnerpod Runner pod is annotated to wait for completion, and the runner container is not restarting {"runnerpod": "actions-runner/mule-runner-h6chf-klmvz"}
AWS EKS - 1.22
Node group - Managed node group
Autoscaling - automatic scheduled autoscaling
@mumoshu @toast-gear Please assist me. This is causing lot of jobs to be in queued status.
Beta Was this translation helpful? Give feedback.
All reactions