Skip to end of metadata
Go to start of metadata

You are viewing an old version of this content. View the current version.

Compare with Current View Version History

« Previous Version 2 Current »

Sometimes you may see flink-jobmanager crashing and not coming up, which means one of the checkpoint snapshots is corrupted. To fix it you can use one of the following methods:

  1. Scale down flink-jobmanager deployment to 0.

  2. Scale down flink-taskmanager deployment to 0.

  3. Go to storage/configmaps

  4. Delete

    1. gv-flink-cluster-config-map

    2. gv-flink-*-config-map

  5. Wait 15-20 sec

  6. Scale up flink-taskmanager deployment to 1.

  7. Scale up flink-jobmanager deployment to 1.

kubectl scale --replicas=0 deployment/flink-jobmanager
kubectl scale --replicas=0 deployment/flink-taskmanager
kubectl get configmap -n default
kubectl delete configmap gv-flink-cluster-config-map -n default
kubectl delete configmap gv-flink-*-config-map -n default     <--- insert config map name from previous command
kubectl scale --replicas=1 deployment/flink-jobmanager
kubectl scale --replicas=1 deployment/flink-taskmanager
  • No labels