Airflow task killed randomly with error Executor reports task instance <TaskInstance: XXXXX [queued]> finished (failed) although the task says its queued. (Info: None) Was the task killed externally? #28359
Unanswered
MauriSanfi
asked this question in
Q&A
Replies: 1 comment 4 replies
-
Yep. Reported the user to Github as spammer. |
Beta Was this translation helpful? Give feedback.
3 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi everyone,
I'm using Airflow 2.4.2 with the Kubernetes executor and I'm facing a strange issue. Every now and then some pods seem to get stuck while initializing, and after some minutes I receive an alert about the task being killed externally. This is a sample of the logs that I get when this happens.
Just for clarification, the main airflow components are deployed in a different node pool than the one I use for the tasks.
This error happens at random hours, for random DAGs. Autoscaling is enabled.
The logs from the cluster are not super useful, I can't actually relate the logs with this error. But here's what I got:
Failed to create pod sandbox: rpc error: code = 2 desc = failed to start sandbox container for pod XXXXX: operation timeout: context deadline exceeded
Failed to create pod sandbox: rpc error: code = Unknown desc = failed to set up sandbox container
Any ideas about what to check for and where?
Thanks in advance!
Beta Was this translation helpful? Give feedback.
All reactions