thousands-car-79657
10/04/2024, 5:23 PMflytekit to 1.13.7 (from 1.10.3) while the backend flyte version was 1.13.1, I got the following system error running remote workflow
RuntimeExecutionError: max number of system retry attempts [51/50] exhausted. Last known status message: worker error(s) encountered: [0]: failed at Node[n0]. RuntimeExecutionError: failed during plugin execution, caused by: failed to execute handle for plugin [sidecar]: [BadTaskSpecification] invalid TaskSpecification, primary container [primary] not defined
Could it be caused by flytekit version being ahead of flyte version? How does version compatibility work?prehistoric-leather-97354
10/04/2024, 5:31 PMthousands-car-79657
10/04/2024, 5:35 PMprehistoric-leather-97354
10/04/2024, 5:37 PMprimaryprehistoric-leather-97354
10/04/2024, 5:37 PMkubectl for your cluster?thousands-car-79657
10/04/2024, 5:38 PMthousands-car-79657
10/04/2024, 5:39 PMmap_taskprehistoric-leather-97354
10/04/2024, 5:39 PMkubectl -n datology-development describe pods <failing pod name> | grep primary_container_namethousands-car-79657
10/04/2024, 5:41 PMhigh-accountant-32689
10/04/2024, 5:55 PMmap_task to use array nodes. Just to unblock you, you can still import the legacy map task from here, but I'd love to understand what broke in your case.high-accountant-32689
10/04/2024, 5:57 PMthousands-car-79657
10/04/2024, 6:02 PMthousands-car-79657
10/04/2024, 6:09 PMk -n staging describe pods fves1p1y-0 | grep primary_
primary_container_name: primaryprehistoric-leather-97354
10/04/2024, 6:11 PMthousands-car-79657
10/04/2024, 6:18 PMhigh-accountant-32689
10/04/2024, 6:25 PMthousands-car-79657
10/04/2024, 9:49 PMprimary_container_name. If the task_config of the task is not overwritten, primary_container_name will be the pod name. But if it’s overwritten with flytekitplugins.pod.task.Pod, then primary_container_name will be primary, which causes the error.thousands-car-79657
10/04/2024, 11:49 PMflytekitplugins-pod. Once I switched to PodTemplate (as advised here), it worked!