Skip to main content
Version: v2.8

Troubleshooting Controlplane Nodes

This section applies to nodes with the controlplane role.

Check if the Controlplane Containers are Running

There are three specific containers launched on nodes with the controlplane role:

  • kube-apiserver
  • kube-controller-manager
  • kube-scheduler

The containers should have status Up. The duration shown after Up is the time the container has been running.

docker ps -a -f=name='kube-apiserver|kube-controller-manager|kube-scheduler'

Example output:

CONTAINER ID        IMAGE                                COMMAND                  CREATED             STATUS              PORTS               NAMES
26c7159abbcc rancher/hyperkube:v1.11.5-rancher1 "/opt/rke-tools/en..." 3 hours ago Up 3 hours kube-apiserver
f3d287ca4549 rancher/hyperkube:v1.11.5-rancher1 "/opt/rke-tools/en..." 3 hours ago Up 3 hours kube-scheduler
bdf3898b8063 rancher/hyperkube:v1.11.5-rancher1 "/opt/rke-tools/en..." 3 hours ago Up 3 hours kube-controller-manager

Controlplane Container Logging

note

If you added multiple nodes with the controlplane role, both kube-controller-manager and kube-scheduler use a leader election process to determine the leader. Only the current leader will log the performed actions. See Kubernetes leader election how to retrieve the current leader.

The logging of the containers can contain information on what the problem could be.

docker logs kube-apiserver
docker logs kube-controller-manager
docker logs kube-scheduler

RKE2 Server Logging

If Rancher provisions an RKE2 cluster that can't communicate with Rancher, you can run this command on a server node in the downstream cluster to get the RKE2 server logs:

journalctl -u rke2-server -f