Skip to content

Instantly share code, notes, and snippets.

@askb
Created December 19, 2021 04:36
Show Gist options
  • Star 0 You must be signed in to star a gist
  • Fork 0 You must be signed in to fork a gist
  • Save askb/cb1bb7843e2cbdfd33043f0a9056632c to your computer and use it in GitHub Desktop.
Save askb/cb1bb7843e2cbdfd33043f0a9056632c to your computer and use it in GitHub Desktop.
Debugging ODL K8s cluster nodes notReady
kubectl get po -A -o wide
NAMESPACE NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE READINESS GATES
default sdnc-opendaylight-0 0/1 Pending 0 37m <none> <none> <none> <none>
kube-system calico-kube-controllers-7b67cb9dd4-vnfm8 0/1 Pending 0 44m <none> <none> <none> <none>
kube-system calico-node-9kwq4 1/1 Running 0 40m 10.0.0.50 sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-node-1 <none> <none>
kube-system calico-node-jvpgh 1/1 Running 0 40m 10.0.0.59 sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-node-0 <none> <none>
kube-system calico-node-wp484 1/1 Running 0 44m 10.0.0.243 sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-master-0 <none> <none>
kube-system coredns-57995474d5-6qbjh 1/1 Running 0 44m 10.100.45.130 sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-master-0 <none> <none>
kube-system coredns-57995474d5-xtcxb 1/1 Running 0 44m 10.100.45.129 sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-master-0 <none> <none>
kube-system csi-cinder-controllerplugin-0 5/5 Running 0 43m 10.100.45.128 sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-master-0 <none> <none>
kube-system dashboard-metrics-scraper-7674b9d54f-dnqsl 0/1 Pending 0 44m <none> <none> <none> <none>
kube-system k8s-keystone-auth-8cjmz 1/1 Running 0 43m 10.0.0.243 sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-master-0 <none> <none>
kube-system kube-dns-autoscaler-7967dcdbd7-m5ct6 0/1 Pending 0 44m <none> <none> <none> <none>
kube-system kubernetes-dashboard-bfc6ccfdf-t4jl2 0/1 Pending 0 44m <none> <none> <none> <none>
kube-system openstack-cloud-controller-manager-vh77v 0/1 ImagePullBackOff 0 44m 10.0.0.243 sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-master-0 <none> <none>
kubectl get events --sort-by='.metadata.creationTimestamp'
LAST SEEN TYPE REASON OBJECT MESSAGE
41m Normal Starting node/sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-master-0 Starting kube-proxy.
41m Normal RegisteredNode node/sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-master-0 Node sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-master-0 event: Registered Node sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-master-0 in Controller
41m Normal NodeReady node/sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-master-0 Node sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-master-0 status is now: NodeReady
38m Normal NodeAllocatableEnforced node/sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-node-1 Updated Node Allocatable limit across pods
38m Normal NodeHasSufficientPID node/sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-node-1 Node sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-node-1 status is now: NodeHasSufficientPID
38m Normal NodeHasNoDiskPressure node/sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-node-1 Node sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-node-1 status is now: NodeHasNoDiskPressure
38m Normal NodeHasSufficientMemory node/sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-node-1 Node sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-node-1 status is now: NodeHasSufficientMemory
38m Normal Starting node/sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-node-1 Starting kubelet.
38m Normal Starting node/sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-node-0 Starting kubelet.
38m Normal RegisteredNode node/sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-node-1 Node sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-node-1 event: Registered Node sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-node-1 in Controller
38m Normal NodeAllocatableEnforced node/sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-node-0 Updated Node Allocatable limit across pods
38m Normal NodeHasSufficientPID node/sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-node-0 Node sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-node-0 status is now: NodeHasSufficientPID
38m Normal NodeHasNoDiskPressure node/sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-node-0 Node sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-node-0 status is now: NodeHasNoDiskPressure
38m Normal NodeHasSufficientMemory node/sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-node-0 Node sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-node-0 status is now: NodeHasSufficientMemory
38m Normal Starting node/sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-node-1 Starting kube-proxy.
38m Normal Starting node/sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-node-0 Starting kube-proxy.
37m Normal RegisteredNode node/sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-node-0 Node sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-node-0 event: Registered Node sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-node-0 in Controller
37m Normal NodeReady node/sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-node-1 Node sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-node-1 status is now: NodeReady
37m Normal NodeReady node/sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-node-0 Node sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-node-0 status is now: NodeReady
34m Warning FailedScheduling pod/sdnc-opendaylight-0 0/3 nodes are available: 1 node(s) had taint {node-role.kubernetes.io/master: }, that the pod didn't tolerate, 2 node(s) had taint {node.cloudprovider.kubernetes.io/uninitialized: true}, that the pod didn't tolerate.
34m Normal SuccessfulCreate statefulset/sdnc-opendaylight create Pod sdnc-opendaylight-0 in StatefulSet sdnc-opendaylight successful
12m Warning FailedScheduling pod/sdnc-opendaylight-0 0/3 nodes are available: 1 node(s) had taint {node-role.kubernetes.io/master: }, that the pod didn't tolerate, 2 node(s) had taint {node.cloudprovider.kubernetes.io/uninitialized: true}, that the pod didn't tolerate.
4m38s Warning FailedGetScale horizontalpodautoscaler/sdnc-opendaylight deployments/scale.apps "sdnc-opendaylight" not found
kubectl get nodes --show-labels
NAME STATUS ROLES AGE VERSION LABELS
sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-master-0 Ready master 39m v1.21.1 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/os=linux,kubernetes.io/arch=amd64,kubernetes.io/hostname=sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-master-0,kubernetes.io/os=linux,magnum.openstack.org/nodegroup=default-master,magnum.openstack.org/role=master,node-role.kubernetes.io/master=
sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-node-0 Ready <none> 35m v1.21.1 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/os=linux,kubernetes.io/arch=amd64,kubernetes.io/hostname=sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-node-0,kubernetes.io/os=linux,magnum.openstack.org/nodegroup=default-worker,magnum.openstack.org/role=worker
sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-node-1 Ready <none> 35m v1.21.1 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/os=linux,kubernetes.io/arch=amd64,kubernetes.io/hostname=sandbox-packaging-k8s-odl-depl-ycser6ofq7ic-node-1,kubernetes.io/os=linux,magnum.openstack.org/nodegroup=default-worker,magnum.openstack.org/role=worker
> kubectl describe pods
Name: sdnc-opendaylight-0
Namespace: default
Priority: 0
Node: <none>
Labels: app.kubernetes.io/instance=sdnc
app.kubernetes.io/name=opendaylight
controller-revision-hash=sdnc-opendaylight-5c9c85f8c4
statefulset.kubernetes.io/pod-name=sdnc-opendaylight-0
Annotations: kubernetes.io/psp: magnum.privileged
Status: Pending
IP:
IPs: <none>
Controlled By: StatefulSet/sdnc-opendaylight
Init Containers:
updatevolperm:
Image: busybox
Port: <none>
Host Port: <none>
Command:
chown
8181
/data
Environment: <none>
Mounts:
/data from odlvol (rw)
/var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-st6bd (ro)
Containers:
opendaylight:
Image: nexus3.opendaylight.org:10001/opendaylight/opendaylight:14.2.0
Port: 8181/TCP
Host Port: 0/TCP
Command:
bash
-c
bash -x /scripts/startodl.sh
Readiness: tcp-socket :8181 delay=30s timeout=1s period=10s #success=1 #failure=3
Environment:
FEATURES: odl-restconf,odl-restconf-all
JAVA_HOME: /opt/openjdk-11/
JAVA_OPTS: -Xms512m -Xmx2048m
EXTRA_JAVA_OPTS: -XX:+UseG1GC -XX:MaxGCPauseMillis=100 -XX:ParallelGCThreads=3 -XX:+ParallelRefProcEnabled -XX:+UseStringDeduplication
Mounts:
/data from odlvol (rw)
/scripts from scripts (rw)
/var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-st6bd (ro)
Conditions:
Type Status
PodScheduled False
Volumes:
scripts:
Type: ConfigMap (a volume populated by a ConfigMap)
Name: sdnc-opendaylight
Optional: false
odlvol:
Type: EmptyDir (a temporary directory that shares a pod's lifetime)
Medium:
SizeLimit: <unset>
kube-api-access-st6bd:
Type: Projected (a volume that contains injected data from multiple sources)
TokenExpirationSeconds: 3607
ConfigMapName: kube-root-ca.crt
ConfigMapOptional: <nil>
DownwardAPI: true
QoS Class: BestEffort
Node-Selectors: <none>
Tolerations: node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Warning FailedScheduling 30m default-scheduler 0/3 nodes are available: 1 node(s) had taint {node-role.kubernetes.io/master: }, that the pod didn't tolerate, 2 node(s) had taint {node.cloudprovider.kubernetes.io/uninitialized: true}, that the pod didn't tolerate.
Warning FailedScheduling 30m default-scheduler 0/3 nodes are available: 1 node(s) had taint {node-role.kubernetes.io/master: }, that the pod didn't tolerate, 2 node(s) had taint {node.cloudprovider.kubernetes.io/uninitialized: true}, that the pod didn't tolerate.
@askb
Copy link
Author

askb commented Dec 19, 2021

> kubectl describe nodes
Name:               sandbox-packaging-k8s-odl-depl-gpysc6ygvjci-master-0
Roles:              master
Labels:             beta.kubernetes.io/arch=amd64
                    beta.kubernetes.io/os=linux
                    kubernetes.io/arch=amd64
                    kubernetes.io/hostname=sandbox-packaging-k8s-odl-depl-gpysc6ygvjci-master-0
                    kubernetes.io/os=linux
                    magnum.openstack.org/nodegroup=default-master
                    magnum.openstack.org/role=master
                    node-role.kubernetes.io/master=
Annotations:        node.alpha.kubernetes.io/ttl: 0
                    projectcalico.org/IPv4Address: 10.0.0.96/24
                    volumes.kubernetes.io/controller-managed-attach-detach: true
CreationTimestamp:  Sun, 19 Dec 2021 03:11:03 +0000
Taints:             node-role.kubernetes.io/master:NoSchedule
                    node.cloudprovider.kubernetes.io/uninitialized=true:NoSchedule
Unschedulable:      false
Lease:
  HolderIdentity:  sandbox-packaging-k8s-odl-depl-gpysc6ygvjci-master-0
  AcquireTime:     <unset>
  RenewTime:       Sun, 19 Dec 2021 03:26:50 +0000
Conditions:
  Type                 Status  LastHeartbeatTime                 LastTransitionTime                Reason                       Message
  ----                 ------  -----------------                 ------------------                ------                       -------
  NetworkUnavailable   False   Sun, 19 Dec 2021 03:11:42 +0000   Sun, 19 Dec 2021 03:11:42 +0000   CalicoIsUp                   Calico is running on this node
  MemoryPressure       False   Sun, 19 Dec 2021 03:22:08 +0000   Sun, 19 Dec 2021 03:11:03 +0000   KubeletHasSufficientMemory   kubelet has sufficient memory available
  DiskPressure         False   Sun, 19 Dec 2021 03:22:08 +0000   Sun, 19 Dec 2021 03:11:03 +0000   KubeletHasNoDiskPressure     kubelet has no disk pressure
  PIDPressure          False   Sun, 19 Dec 2021 03:22:08 +0000   Sun, 19 Dec 2021 03:11:03 +0000   KubeletHasSufficientPID      kubelet has sufficient PID available
  Ready                True    Sun, 19 Dec 2021 03:22:08 +0000   Sun, 19 Dec 2021 03:11:43 +0000   KubeletReady                 kubelet is posting ready status
Addresses:
  InternalIP:  10.0.0.96
  Hostname:    sandbox-packaging-k8s-odl-depl-gpysc6ygvjci-master-0
Capacity:
  cpu:                8
  ephemeral-storage:  9950188Ki
  hugepages-1Gi:      0
  hugepages-2Mi:      0
  memory:             32876572Ki
  pods:               110
Allocatable:
  cpu:                8
  ephemeral-storage:  9170093246
  hugepages-1Gi:      0
  hugepages-2Mi:      0
  memory:             32774172Ki
  pods:               110
System Info:
  Machine ID:                 dce21e574ff444889e1e7868ad814c1a
  System UUID:                dce21e57-4ff4-4488-9e1e-7868ad814c1a
  Boot ID:                    ec51c161-81f5-4136-b750-59b0f2d9c045
  Kernel Version:             5.11.20-300.fc34.x86_64
  OS Image:                   Fedora CoreOS 34.20210518.3.0
  Operating System:           linux
  Architecture:               amd64
  Container Runtime Version:  docker://20.10.6
  Kubelet Version:            v1.21.1
  Kube-Proxy Version:         v1.21.1
PodCIDR:                      10.100.0.0/24
PodCIDRs:                     10.100.0.0/24
Non-terminated Pods:          (6 in total)
  Namespace                   Name                                        CPU Requests  CPU Limits  Memory Requests  Memory Limits  Age
  ---------                   ----                                        ------------  ----------  ---------------  -------------  ---
  kube-system                 calico-node-bd9l5                           250m (3%)     0 (0%)      0 (0%)           0 (0%)         15m
  kube-system                 coredns-57995474d5-h4pqv                    100m (1%)     0 (0%)      70Mi (0%)        170Mi (0%)     15m
  kube-system                 coredns-57995474d5-lw8f5                    100m (1%)     0 (0%)      70Mi (0%)        170Mi (0%)     15m
  kube-system                 csi-cinder-controllerplugin-0               0 (0%)        0 (0%)      0 (0%)           0 (0%)         15m
  kube-system                 k8s-keystone-auth-qg54x                     200m (2%)     0 (0%)      0 (0%)           0 (0%)         15m
  kube-system                 openstack-cloud-controller-manager-7w9z7    0 (0%)        0 (0%)      0 (0%)           0 (0%)         15m
Allocated resources:
  (Total limits may be over 100 percent, i.e., overcommitted.)
  Resource           Requests    Limits
  --------           --------    ------
  cpu                650m (8%)   0 (0%)
  memory             140Mi (0%)  340Mi (1%)
  ephemeral-storage  0 (0%)      0 (0%)
  hugepages-1Gi      0 (0%)      0 (0%)
  hugepages-2Mi      0 (0%)      0 (0%)
Events:
  Type    Reason     Age   From        Message
  ----    ------     ----  ----        -------
  Normal  Starting   15m   kube-proxy  Starting kube-proxy.
  Normal  NodeReady  15m   kubelet     Node sandbox-packaging-k8s-odl-depl-gpysc6ygvjci-master-0 status is now: NodeReady


Name:               sandbox-packaging-k8s-odl-depl-gpysc6ygvjci-node-0
Roles:              <none>
Labels:             beta.kubernetes.io/arch=amd64
                    beta.kubernetes.io/os=linux
                    kubernetes.io/arch=amd64
                    kubernetes.io/hostname=sandbox-packaging-k8s-odl-depl-gpysc6ygvjci-node-0
                    kubernetes.io/os=linux
                    magnum.openstack.org/nodegroup=default-worker
                    magnum.openstack.org/role=worker
Annotations:        node.alpha.kubernetes.io/ttl: 0
                    projectcalico.org/IPv4Address: 10.0.0.61/24
                    volumes.kubernetes.io/controller-managed-attach-detach: true
CreationTimestamp:  Sun, 19 Dec 2021 03:14:55 +0000
Taints:             node.cloudprovider.kubernetes.io/uninitialized=true:NoSchedule
Unschedulable:      false
Lease:
  HolderIdentity:  sandbox-packaging-k8s-odl-depl-gpysc6ygvjci-node-0
  AcquireTime:     <unset>
  RenewTime:       Sun, 19 Dec 2021 03:26:56 +0000
Conditions:
  Type                 Status  LastHeartbeatTime                 LastTransitionTime                Reason                       Message
  ----                 ------  -----------------                 ------------------                ------                       -------
  NetworkUnavailable   False   Sun, 19 Dec 2021 03:15:13 +0000   Sun, 19 Dec 2021 03:15:13 +0000   CalicoIsUp                   Calico is running on this node
  MemoryPressure       False   Sun, 19 Dec 2021 03:25:27 +0000   Sun, 19 Dec 2021 03:14:55 +0000   KubeletHasSufficientMemory   kubelet has sufficient memory available
  DiskPressure         False   Sun, 19 Dec 2021 03:25:27 +0000   Sun, 19 Dec 2021 03:14:55 +0000   KubeletHasNoDiskPressure     kubelet has no disk pressure
  PIDPressure          False   Sun, 19 Dec 2021 03:25:27 +0000   Sun, 19 Dec 2021 03:14:55 +0000   KubeletHasSufficientPID      kubelet has sufficient PID available
  Ready                True    Sun, 19 Dec 2021 03:25:27 +0000   Sun, 19 Dec 2021 03:15:15 +0000   KubeletReady                 kubelet is posting ready status
Addresses:
  InternalIP:  10.0.0.61
  Hostname:    sandbox-packaging-k8s-odl-depl-gpysc6ygvjci-node-0
Capacity:
  cpu:                8
  ephemeral-storage:  9950188Ki
  hugepages-1Gi:      0
  hugepages-2Mi:      0
  memory:             32876572Ki
  pods:               110
Allocatable:
  cpu:                8
  ephemeral-storage:  9170093246
  hugepages-1Gi:      0
  hugepages-2Mi:      0
  memory:             32774172Ki
  pods:               110
System Info:
  Machine ID:                 859d4fe55e0d416c9f4e8fd82aabac7a
  System UUID:                859d4fe5-5e0d-416c-9f4e-8fd82aabac7a
  Boot ID:                    5fdc6c27-5a25-4b09-beb6-080c36ab8118
  Kernel Version:             5.11.20-300.fc34.x86_64
  OS Image:                   Fedora CoreOS 34.20210518.3.0
  Operating System:           linux
  Architecture:               amd64
  Container Runtime Version:  docker://20.10.6
  Kubelet Version:            v1.21.1
  Kube-Proxy Version:         v1.21.1
PodCIDR:                      10.100.2.0/24
PodCIDRs:                     10.100.2.0/24
Non-terminated Pods:          (1 in total)
  Namespace                   Name                 CPU Requests  CPU Limits  Memory Requests  Memory Limits  Age
  ---------                   ----                 ------------  ----------  ---------------  -------------  ---
  kube-system                 calico-node-mffzq    250m (3%)     0 (0%)      0 (0%)           0 (0%)         12m
Allocated resources:
  (Total limits may be over 100 percent, i.e., overcommitted.)
  Resource           Requests   Limits
  --------           --------   ------
  cpu                250m (3%)  0 (0%)
  memory             0 (0%)     0 (0%)
  ephemeral-storage  0 (0%)     0 (0%)
  hugepages-1Gi      0 (0%)     0 (0%)
  hugepages-2Mi      0 (0%)     0 (0%)
Events:
  Type    Reason                   Age   From        Message
  ----    ------                   ----  ----        -------
  Normal  Starting                 12m   kubelet     Starting kubelet.
  Normal  NodeHasSufficientMemory  12m   kubelet     Node sandbox-packaging-k8s-odl-depl-gpysc6ygvjci-node-0 status is now: NodeHasSufficientMemory
  Normal  NodeHasNoDiskPressure    12m   kubelet     Node sandbox-packaging-k8s-odl-depl-gpysc6ygvjci-node-0 status is now: NodeHasNoDiskPressure
  Normal  NodeHasSufficientPID     12m   kubelet     Node sandbox-packaging-k8s-odl-depl-gpysc6ygvjci-node-0 status is now: NodeHasSufficientPID
  Normal  NodeAllocatableEnforced  12m   kubelet     Updated Node Allocatable limit across pods
  Normal  Starting                 12m   kube-proxy  Starting kube-proxy.
  Normal  NodeReady                11m   kubelet     Node sandbox-packaging-k8s-odl-depl-gpysc6ygvjci-node-0 status is now: NodeReady


Name:               sandbox-packaging-k8s-odl-depl-gpysc6ygvjci-node-1
Roles:              <none>
Labels:             beta.kubernetes.io/arch=amd64
                    beta.kubernetes.io/os=linux
                    kubernetes.io/arch=amd64
                    kubernetes.io/hostname=sandbox-packaging-k8s-odl-depl-gpysc6ygvjci-node-1
                    kubernetes.io/os=linux
                    magnum.openstack.org/nodegroup=default-worker
                    magnum.openstack.org/role=worker
Annotations:        node.alpha.kubernetes.io/ttl: 0
                    projectcalico.org/IPv4Address: 10.0.0.18/24
                    volumes.kubernetes.io/controller-managed-attach-detach: true
CreationTimestamp:  Sun, 19 Dec 2021 03:14:51 +0000
Taints:             node.cloudprovider.kubernetes.io/uninitialized=true:NoSchedule
Unschedulable:      false
Lease:
  HolderIdentity:  sandbox-packaging-k8s-odl-depl-gpysc6ygvjci-node-1
  AcquireTime:     <unset>
  RenewTime:       Sun, 19 Dec 2021 03:26:52 +0000
Conditions:
  Type                 Status  LastHeartbeatTime                 LastTransitionTime                Reason                       Message
  ----                 ------  -----------------                 ------------------                ------                       -------
  NetworkUnavailable   False   Sun, 19 Dec 2021 03:15:09 +0000   Sun, 19 Dec 2021 03:15:09 +0000   CalicoIsUp                   Calico is running on this node
  MemoryPressure       False   Sun, 19 Dec 2021 03:25:24 +0000   Sun, 19 Dec 2021 03:14:51 +0000   KubeletHasSufficientMemory   kubelet has sufficient memory available
  DiskPressure         False   Sun, 19 Dec 2021 03:25:24 +0000   Sun, 19 Dec 2021 03:14:51 +0000   KubeletHasNoDiskPressure     kubelet has no disk pressure
  PIDPressure          False   Sun, 19 Dec 2021 03:25:24 +0000   Sun, 19 Dec 2021 03:14:51 +0000   KubeletHasSufficientPID      kubelet has sufficient PID available
  Ready                True    Sun, 19 Dec 2021 03:25:24 +0000   Sun, 19 Dec 2021 03:15:11 +0000   KubeletReady                 kubelet is posting ready status
Addresses:
  InternalIP:  10.0.0.18
  Hostname:    sandbox-packaging-k8s-odl-depl-gpysc6ygvjci-node-1
Capacity:
  cpu:                8
  ephemeral-storage:  9950188Ki
  hugepages-1Gi:      0
  hugepages-2Mi:      0
  memory:             32876572Ki
  pods:               110
Allocatable:
  cpu:                8
  ephemeral-storage:  9170093246
  hugepages-1Gi:      0
  hugepages-2Mi:      0
  memory:             32774172Ki
  pods:               110
System Info:
  Machine ID:                 6a97ffd3c5b641d1ac5fef5ae364e718
  System UUID:                6a97ffd3-c5b6-41d1-ac5f-ef5ae364e718
  Boot ID:                    343ea2c5-a080-4c8d-9ab5-3a8cbe7b55e0
  Kernel Version:             5.11.20-300.fc34.x86_64
  OS Image:                   Fedora CoreOS 34.20210518.3.0
  Operating System:           linux
  Architecture:               amd64
  Container Runtime Version:  docker://20.10.6
  Kubelet Version:            v1.21.1
  Kube-Proxy Version:         v1.21.1
PodCIDR:                      10.100.1.0/24
PodCIDRs:                     10.100.1.0/24
Non-terminated Pods:          (1 in total)
  Namespace                   Name                 CPU Requests  CPU Limits  Memory Requests  Memory Limits  Age
  ---------                   ----                 ------------  ----------  ---------------  -------------  ---
  kube-system                 calico-node-t99tt    250m (3%)     0 (0%)      0 (0%)           0 (0%)         12m
Allocated resources:
  (Total limits may be over 100 percent, i.e., overcommitted.)
  Resource           Requests   Limits
  --------           --------   ------
  cpu                250m (3%)  0 (0%)
  memory             0 (0%)     0 (0%)
  ephemeral-storage  0 (0%)     0 (0%)
  hugepages-1Gi      0 (0%)     0 (0%)
  hugepages-2Mi      0 (0%)     0 (0%)
Events:
  Type    Reason                   Age                From        Message
  ----    ------                   ----               ----        -------
  Normal  Starting                 12m                kubelet     Starting kubelet.
  Normal  NodeHasSufficientMemory  12m (x2 over 12m)  kubelet     Node sandbox-packaging-k8s-odl-depl-gpysc6ygvjci-node-1 status is now: NodeHasSufficientMemory
  Normal  NodeHasNoDiskPressure    12m (x2 over 12m)  kubelet     Node sandbox-packaging-k8s-odl-depl-gpysc6ygvjci-node-1 status is now: NodeHasNoDiskPressure
  Normal  NodeHasSufficientPID     12m (x2 over 12m)  kubelet     Node sandbox-packaging-k8s-odl-depl-gpysc6ygvjci-node-1 status is now: NodeHasSufficientPID
  Normal  NodeAllocatableEnforced  12m                kubelet     Updated Node Allocatable limit across pods
  Normal  Starting                 12m                kube-proxy  Starting kube-proxy.
  Normal  NodeReady                11m                kubelet     Node sandbox-packaging-k8s-odl-depl-gpysc6ygvjci-node-1 status is now: NodeReady

@askb
Copy link
Author

askb commented Dec 19, 2021

> kubectl -n kube-system describe pod calico-kube-controllers-7b67cb9dd4-h9228
Name:                 calico-kube-controllers-7b67cb9dd4-h9228
Namespace:            kube-system
Priority:             2000000000
Priority Class Name:  system-cluster-critical
Node:                 <none>
Labels:               k8s-app=calico-kube-controllers
                      pod-template-hash=7b67cb9dd4
Annotations:          kubernetes.io/psp: magnum.privileged
                      scheduler.alpha.kubernetes.io/critical-pod: 
Status:               Pending
IP:                   
IPs:                  <none>
Controlled By:        ReplicaSet/calico-kube-controllers-7b67cb9dd4
Containers:
  calico-kube-controllers:
    Image:      registry.public.yul1.vexxhost.net/magnum/kube-controllers:v3.13.1
    Port:       <none>
    Host Port:  <none>
    Readiness:  exec [/usr/bin/check-status -r] delay=0s timeout=1s period=10s #success=1 #failure=3
    Environment:
      ENABLED_CONTROLLERS:  node
      DATASTORE_TYPE:       kubernetes
    Mounts:
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-v5hnc (ro)
Conditions:
  Type           Status
  PodScheduled   False 
Volumes:
  kube-api-access-v5hnc:
    Type:                    Projected (a volume that contains injected data from multiple sources)
    TokenExpirationSeconds:  3607
    ConfigMapName:           kube-root-ca.crt
    ConfigMapOptional:       <nil>
    DownwardAPI:             true
QoS Class:                   BestEffort
Node-Selectors:              kubernetes.io/os=linux
Tolerations:                 CriticalAddonsOnly op=Exists
                             node-role.kubernetes.io/master:NoSchedule
                             node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
                             node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
  Type     Reason            Age    From               Message
  ----     ------            ----   ----               -------
  Warning  FailedScheduling  12m    default-scheduler  0/1 nodes are available: 1 node(s) had taint {node.cloudprovider.kubernetes.io/uninitialized: true}, that the pod didn't tolerate.
  Warning  FailedScheduling  12m    default-scheduler  0/1 nodes are available: 1 node(s) had taint {node.cloudprovider.kubernetes.io/uninitialized: true}, that the pod didn't tolerate.
  Warning  FailedScheduling  8m45s  default-scheduler  0/3 nodes are available: 3 node(s) had taint {node.cloudprovider.kubernetes.io/uninitialized: true}, that the pod didn't tolerate.
  Warning  FailedScheduling  8m34s  default-scheduler  0/3 nodes are available: 3 node(s) had taint {node.cloudprovider.kubernetes.io/uninitialized: true}, that the pod didn't tolerate.
[jenkins@snd-centos7-helm-4c-4g-273 ~]> kubectl -n kube-system logs calico-kube-controllers-7b67cb9dd4-h9228
[jenkins@snd-centos7-helm-4c-4g-273 ~]> kubectl get po -A -o wide
NAMESPACE     NAME                                         READY   STATUS             RESTARTS   AGE     IP              NODE                                                   NOMINATED NODE   READINESS GATES
default       sdnc-opendaylight-0                          0/1     Pending            0          7m18s   <none>          <none>                                                 <none>           <none>
kube-system   calico-kube-controllers-7b67cb9dd4-h9228     0/1     Pending            0          13m     <none>          <none>                                                 <none>           <none>
kube-system   calico-node-2t7d7                            1/1     Running            0          10m     10.0.0.205      sandbox-packaging-k8s-odl-depl-dsm67qpjccu4-node-0     <none>           <none>
kube-system   calico-node-5qlh8                            1/1     Running            0          10m     10.0.0.81       sandbox-packaging-k8s-odl-depl-dsm67qpjccu4-node-1     <none>           <none>
kube-system   calico-node-l4bhq                            1/1     Running            0          13m     10.0.0.38       sandbox-packaging-k8s-odl-depl-dsm67qpjccu4-master-0   <none>           <none>
kube-system   coredns-57995474d5-l2r8l                     1/1     Running            0          13m     10.100.40.130   sandbox-packaging-k8s-odl-depl-dsm67qpjccu4-master-0   <none>           <none>
kube-system   coredns-57995474d5-nsqcb                     1/1     Running            0          13m     10.100.40.129   sandbox-packaging-k8s-odl-depl-dsm67qpjccu4-master-0   <none>           <none>
kube-system   csi-cinder-controllerplugin-0                5/5     Running            0          13m     10.100.40.128   sandbox-packaging-k8s-odl-depl-dsm67qpjccu4-master-0   <none>           <none>
kube-system   dashboard-metrics-scraper-7674b9d54f-k7zbb   0/1     Pending            0          13m     <none>          <none>                                                 <none>           <none>
kube-system   k8s-keystone-auth-mqnd5                      1/1     Running            0          13m     10.0.0.38       sandbox-packaging-k8s-odl-depl-dsm67qpjccu4-master-0   <none>           <none>
kube-system   kube-dns-autoscaler-7967dcdbd7-tkpzg         0/1     Pending            0          13m     <none>          <none>                                                 <none>           <none>
kube-system   kubernetes-dashboard-bfc6ccfdf-gjrqq         0/1     Pending            0          13m     <none>          <none>                                                 <none>           <none>
kube-system   openstack-cloud-controller-manager-5sdxn     0/1     ImagePullBackOff   0          13m     10.0.0.38       sandbox-packaging-k8s-odl-depl-dsm67qpjccu4-master-0   <none>           <none>
[jenkins@snd-centos7-helm-4c-4g-273 ~]> kubectl -n kube-system logs kube-dns-autoscaler-7967dcdbd7-tkpzg
[jenkins@snd-centos7-helm-4c-4g-273 ~]> kubectl -n kube-system logs sdnc-opendaylight-0
Error from server (NotFound): pods "sdnc-opendaylight-0" not found
[jenkins@snd-centos7-helm-4c-4g-273 ~]> kubectl -n kube-system logs openstack-cloud-controller-manager-5sdxn
Error from server (BadRequest): container "openstack-cloud-controller-manager" in pod "openstack-cloud-controller-manager-5sdxn" is waiting to start: trying and failing to pull image
[jenkins@snd-centos7-helm-4c-4g-273 ~]> 

@askb
Copy link
Author

askb commented Dec 20, 2021

[jenkins@snd-centos7-helm-4c-4g-274 ~]> kubectl describe node sandbox-packaging-k8s-odl-depl-kn7p2wdnydfr-master-0 | grep Taints
Taints:             node-role.kubernetes.io/master:NoSchedule
[jenkins@snd-centos7-helm-4c-4g-274 ~]> kubectl taint node master node-role.kubernetes.io/master:NoSchedule-
Error from server (NotFound): nodes "master" not found
[jenkins@snd-centos7-helm-4c-4g-274 ~]> kubectl taint node sandbox-packaging-k8s-odl-depl-kn7p2wdnydfr-master-0 node-role.kubernetes.io/master:NoSchedule-
node/sandbox-packaging-k8s-odl-depl-kn7p2wdnydfr-master-0 untainted
[jenkins@snd-centos7-helm-4c-4g-274 ~]> kubectl describe node sandbox-packaging-k8s-odl-depl-kn7p2wdnydfr-node-0 | grep Taints
Taints:             node.cloudprovider.kubernetes.io/uninitialized=true:NoSchedule
[jenkins@snd-centos7-helm-4c-4g-274 ~]> kubectl describe node sandbox-packaging-k8s-odl-depl-kn7p2wdnydfr-node-1 | grep Taints
Taints:             node.cloudprovider.kubernetes.io/uninitialized=true:NoSchedule
[jenkins@snd-centos7-helm-4c-4g-274 ~]> kubectl taint node sandbox-packaging-^Code-role.kubernetes.io/master:NoSchedule-
[jenkins@snd-centos7-helm-4c-4g-274 ~]> kubectl --namespace default port-forward $POD_NAME 8080:$CONTAINER_PORT
error: unable to forward port because pod is not running. Current status=Pending

@askb
Copy link
Author

askb commented Dec 21, 2021

> kubectl get po -A
NAMESPACE     NAME                                         READY   STATUS             RESTARTS   AGE
default       sdnc-opendaylight-0                          0/1     Pending            0          3m26s
kube-system   calico-kube-controllers-7b67cb9dd4-p2zmv     0/1     Pending            0          10m
kube-system   calico-node-2nsq4                            1/1     Running            0          7m23s
kube-system   calico-node-72zd5                            1/1     Running            0          10m
kube-system   calico-node-89xx6                            1/1     Running            0          7m25s
kube-system   coredns-57995474d5-7bsnr                     1/1     Running            0          10m
kube-system   coredns-57995474d5-p9px9                     1/1     Running            0          10m
kube-system   csi-cinder-controllerplugin-0                5/5     Running            0          10m
kube-system   dashboard-metrics-scraper-7674b9d54f-5mk8h   0/1     Pending            0          10m
kube-system   k8s-keystone-auth-jvtrg                      1/1     Running            0          10m
kube-system   kube-dns-autoscaler-7967dcdbd7-xwvhv         0/1     Pending            0          10m
kube-system   kubernetes-dashboard-bfc6ccfdf-ndq7n         0/1     Pending            0          10m
**kube-system   openstack-cloud-controller-manager-mq6t6     0/1     ImagePullBackOff   0          10m**

@askb
Copy link
Author

askb commented Dec 21, 2021

> kubectl describe pod openstack-cloud-controller-manager-mq6t6 -n kube-system
Name:         openstack-cloud-controller-manager-mq6t6
Namespace:    kube-system
Priority:     0
Node:         sandbox-packaging-k8s-odl-depl-ccxetovfc52m-master-0/10.0.0.74
Start Time:   Tue, 21 Dec 2021 22:47:58 +0000
Labels:       controller-revision-hash=f68cffd6
              k8s-app=openstack-cloud-controller-manager
              pod-template-generation=1
Annotations:  kubernetes.io/psp: magnum.privileged
Status:       Pending
IP:           10.0.0.74
IPs:
  IP:           10.0.0.74
Controlled By:  DaemonSet/openstack-cloud-controller-manager
Containers:
  openstack-cloud-controller-manager:
    Container ID:  
    Image:         registry.public.yul1.vexxhost.net/magnum/openstack-cloud-controller-manager:v.1.18.0
    Image ID:      
    Port:          <none>
    Host Port:     <none>
    Command:
      /bin/openstack-cloud-controller-manager
      --v=2
      --cloud-config=/etc/kubernetes/cloud-config-occm
      --cloud-provider=openstack
      --cluster-name=3f49ab9f-6ef8-49e2-9df4-c69b88008fb5
      --use-service-account-credentials=true
      --bind-address=127.0.0.1
    State:          Waiting
      Reason:       ImagePullBackOff
    Ready:          False
    Restart Count:  0
    Environment:    <none>
    Mounts:
      /etc/kubernetes from cloudconfig (ro)
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-smfq8 (ro)
Conditions:
  Type              Status
  Initialized       True 
  Ready             False 
  ContainersReady   False 
  PodScheduled      True 
Volumes:
  cloudconfig:
    Type:          HostPath (bare host directory volume)
    Path:          /etc/kubernetes
    HostPathType:  
  kube-api-access-smfq8:
    Type:                    Projected (a volume that contains injected data from multiple sources)
    TokenExpirationSeconds:  3607
    ConfigMapName:           kube-root-ca.crt
    ConfigMapOptional:       <nil>
    DownwardAPI:             true
QoS Class:                   BestEffort
Node-Selectors:              node-role.kubernetes.io/master=
Tolerations:                 :NoSchedule op=Exists
                             :NoExecute op=Exists
                             CriticalAddonsOnly op=Exists
                             node.cloudprovider.kubernetes.io/uninitialized=true:NoSchedule
                             node.kubernetes.io/disk-pressure:NoSchedule op=Exists
                             node.kubernetes.io/memory-pressure:NoSchedule op=Exists
                             node.kubernetes.io/network-unavailable:NoSchedule op=Exists
                             node.kubernetes.io/not-ready:NoExecute op=Exists
                             node.kubernetes.io/pid-pressure:NoSchedule op=Exists
                             node.kubernetes.io/unreachable:NoExecute op=Exists
                             node.kubernetes.io/unschedulable:NoSchedule op=Exists
Events:
  Type     Reason     Age                   From               Message
  ----     ------     ----                  ----               -------
  Normal   Scheduled  13m                   default-scheduler  Successfully assigned kube-system/openstack-cloud-controller-manager-mq6t6 to sandbox-packaging-k8s-odl-depl-ccxetovfc52m-master-0
  Normal   Pulling    11m (x4 over 13m)     kubelet            Pulling image "registry.public.yul1.vexxhost.net/magnum/openstack-cloud-controller-manager:v.1.18.0"
  Warning  Failed     11m (x4 over 13m)     kubelet            Failed to pull image "registry.public.yul1.vexxhost.net/magnum/openstack-cloud-controller-manager:v.1.18.0": rpc error: code = Unknown desc = Error response from daemon: manifest for registry.public.yul1.vexxhost.net/magnum/openstack-cloud-controller-manager:v.1.18.0 not found: manifest unknown: manifest unknown
  Warning  Failed     11m (x4 over 13m)     kubelet            Error: ErrImagePull
  Warning  Failed     11m (x6 over 12m)     kubelet            Error: ImagePullBackOff
  Normal   BackOff    2m54s (x43 over 12m)  kubelet            Back-off pulling image "registry.public.yul1.vexxhost.net/magnum/openstack-cloud-controller-manager:v.1.18.0"

@askb
Copy link
Author

askb commented Dec 21, 2021

> docker pull registry.public.yul1.vexxhost.net/magnum/openstack-cloud-controller-manager:v.1.18.0
Error response from daemon: manifest for registry.public.yul1.vexxhost.net/magnum/openstack-cloud-controller-manager:v.1.18.0 not found: manifest unknown: manifest unknown

@askb
Copy link
Author

askb commented Dec 21, 2021

> kubectl get secrets --all-namespaces
NAMESPACE         NAME                                   TYPE                                  DATA   AGE
default           default-token-hw8zt                    kubernetes.io/service-account-token   3      15m
default           sdnc-opendaylight-token-44tnc          kubernetes.io/service-account-token   3      8m23s
default           sh.helm.release.v1.sdnc.v1             helm.sh/release.v1                    1      8m23s
kube-node-lease   default-token-nnbb8                    kubernetes.io/service-account-token   3      15m
kube-public       default-token-gncmj                    kubernetes.io/service-account-token   3      15m
kube-system       admin-token-8zhnd                      kubernetes.io/service-account-token   3      15m
kube-system       calico-kube-controllers-token-xjzt7    kubernetes.io/service-account-token   3      15m
kube-system       calico-node-token-b2rbz                kubernetes.io/service-account-token   3      15m
kube-system       cinder-csi-cloud-config                Opaque                                1      15m
kube-system       cloud-controller-manager-token-k76nk   kubernetes.io/service-account-token   3      15m
kube-system       coredns-token-z9b4h                    kubernetes.io/service-account-token   3      15m
kube-system       csi-cinder-controller-sa-token-kt6p6   kubernetes.io/service-account-token   3      15m
kube-system       csi-cinder-node-sa-token-r4knr         kubernetes.io/service-account-token   3      15m
kube-system       default-token-cs5gf                    kubernetes.io/service-account-token   3      15m
kube-system       k8s-keystone-auth-token-k4t54          kubernetes.io/service-account-token   3      15m
kube-system       kube-dns-autoscaler-token-5knbl        kubernetes.io/service-account-token   3      15m
kube-system       kubernetes-dashboard-certs             Opaque                                0      15m
kube-system       kubernetes-dashboard-csrf              Opaque                                1      15m
kube-system       kubernetes-dashboard-key-holder        Opaque                                0      15m
kube-system       kubernetes-dashboard-token-kv6dp       kubernetes.io/service-account-token   3      15m
kube-system       node-problem-detector-token-fbh2c      kubernetes.io/service-account-token   3      15m
kube-system       os-trustee                             Opaque                                6      15m

@askb
Copy link
Author

askb commented Dec 21, 2021

> kubectl get deployment -n kube-system  
NAME                        READY   UP-TO-DATE   AVAILABLE   AGE
calico-kube-controllers     0/1     1            0           19m
coredns                     2/2     2            2           19m
dashboard-metrics-scraper   0/1     1            0           19m
kube-dns-autoscaler         0/1     1            0           19m
kubernetes-dashboard        0/1     1            0           19m

@askb
Copy link
Author

askb commented Dec 21, 2021

> kubectl get pods -n kube-system
NAME                                         READY   STATUS             RESTARTS   AGE
calico-kube-controllers-7b67cb9dd4-p2zmv     0/1     Pending            0          19m
calico-node-2nsq4                            1/1     Running            0          17m
calico-node-72zd5                            1/1     Running            0          19m
calico-node-89xx6                            1/1     Running            0          17m
coredns-57995474d5-7bsnr                     1/1     Running            0          19m
coredns-57995474d5-p9px9                     1/1     Running            0          19m
csi-cinder-controllerplugin-0                5/5     Running            0          19m
dashboard-metrics-scraper-7674b9d54f-5mk8h   0/1     Pending            0          19m
k8s-keystone-auth-jvtrg                      1/1     Running            0          19m
kube-dns-autoscaler-7967dcdbd7-xwvhv         0/1     Pending            0          19m
kubernetes-dashboard-bfc6ccfdf-ndq7n         0/1     Pending            0          19m
openstack-cloud-controller-manager-mq6t6     0/1     ImagePullBackOff   0          19m

@askb
Copy link
Author

askb commented Dec 21, 2021

> kubectl describe pod openstack-cloud-controller-manager-mq6t6 -n kube-system
Name:         openstack-cloud-controller-manager-mq6t6
Namespace:    kube-system
Priority:     0
Node:         sandbox-packaging-k8s-odl-depl-ccxetovfc52m-master-0/10.0.0.74
Start Time:   Tue, 21 Dec 2021 22:47:58 +0000
Labels:       controller-revision-hash=f68cffd6
              k8s-app=openstack-cloud-controller-manager
              pod-template-generation=1
Annotations:  kubernetes.io/psp: magnum.privileged
Status:       Pending
IP:           10.0.0.74
IPs:
  IP:           10.0.0.74
Controlled By:  DaemonSet/openstack-cloud-controller-manager
Containers:
  openstack-cloud-controller-manager:
    Container ID:  
    Image:         registry.public.yul1.vexxhost.net/magnum/openstack-cloud-controller-manager:v.1.18.0
    Image ID:      
    Port:          <none>
    Host Port:     <none>
    Command:
      /bin/openstack-cloud-controller-manager
      --v=2
      --cloud-config=/etc/kubernetes/cloud-config-occm
      --cloud-provider=openstack
      --cluster-name=3f49ab9f-6ef8-49e2-9df4-c69b88008fb5
      --use-service-account-credentials=true
      --bind-address=127.0.0.1
    State:          Waiting
      Reason:       ImagePullBackOff
    Ready:          False
    Restart Count:  0
    Environment:    <none>
    Mounts:
      /etc/kubernetes from cloudconfig (ro)
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-smfq8 (ro)
Conditions:
  Type              Status
  Initialized       True 
  Ready             False 
  ContainersReady   False 
  PodScheduled      True 
Volumes:
  cloudconfig:
    Type:          HostPath (bare host directory volume)
    Path:          /etc/kubernetes
    HostPathType:  
  kube-api-access-smfq8:
    Type:                    Projected (a volume that contains injected data from multiple sources)
    TokenExpirationSeconds:  3607
    ConfigMapName:           kube-root-ca.crt
    ConfigMapOptional:       <nil>
    DownwardAPI:             true
QoS Class:                   BestEffort
Node-Selectors:              node-role.kubernetes.io/master=
Tolerations:                 :NoSchedule op=Exists
                             :NoExecute op=Exists
                             CriticalAddonsOnly op=Exists
                             node.cloudprovider.kubernetes.io/uninitialized=true:NoSchedule
                             node.kubernetes.io/disk-pressure:NoSchedule op=Exists
                             node.kubernetes.io/memory-pressure:NoSchedule op=Exists
                             node.kubernetes.io/network-unavailable:NoSchedule op=Exists
                             node.kubernetes.io/not-ready:NoExecute op=Exists
                             node.kubernetes.io/pid-pressure:NoSchedule op=Exists
                             node.kubernetes.io/unreachable:NoExecute op=Exists
                             node.kubernetes.io/unschedulable:NoSchedule op=Exists
Events:
  Type     Reason     Age                 From               Message
  ----     ------     ----                ----               -------
  Normal   Scheduled  20m                 default-scheduler  Successfully assigned kube-system/openstack-cloud-controller-manager-mq6t6 to sandbox-packaging-k8s-odl-depl-ccxetovfc52m-master-0
  Normal   Pulling    19m (x4 over 20m)   kubelet            Pulling image "registry.public.yul1.vexxhost.net/magnum/openstack-cloud-controller-manager:v.1.18.0"
  Warning  Failed     19m (x4 over 20m)   kubelet            Failed to pull image "registry.public.yul1.vexxhost.net/magnum/openstack-cloud-controller-manager:v.1.18.0": rpc error: code = Unknown desc = Error response from daemon: manifest for registry.public.yul1.vexxhost.net/magnum/openstack-cloud-controller-manager:v.1.18.0 not found: manifest unknown: manifest unknown
  Warning  Failed     19m (x4 over 20m)   kubelet            Error: ErrImagePull
  Warning  Failed     18m (x6 over 20m)   kubelet            Error: ImagePullBackOff
  Normal   BackOff    19s (x86 over 20m)  kubelet            Back-off pulling image "registry.public.yul1.vexxhost.net/magnum/openstack-cloud-controller-manager:v.1.18.0"

@askb
Copy link
Author

askb commented Dec 21, 2021

// Verify docker pull from the container

> docker pull registry.public.yul1.vexxhost.net/magnum/openstack-cloud-controller-manager:v.1.18.0
Error response from daemon: manifest for registry.public.yul1.vexxhost.net/magnum/openstack-cloud-controller-manager:v.1.18.0 not found: manifest unknown: manifest unknown

// Edit the incorrect version of v.1.8.1 to v1.8.1

> kubectl edit ds openstack-cloud-controller-manager -n kube-system

@askb
Copy link
Author

askb commented Dec 21, 2021

// after fixing this we can see that the pod gets rebooted:

[jenkins@snd-centos7-helm-4c-4g-276 ~]> kubectl get pods -n kube-system
NAME                                         READY   STATUS              RESTARTS   AGE
calico-kube-controllers-7b67cb9dd4-p2zmv     0/1     Pending             0          22m
calico-node-2nsq4                            1/1     Running             0          20m
calico-node-72zd5                            1/1     Running             0          22m
calico-node-89xx6                            1/1     Running             0          20m
coredns-57995474d5-7bsnr                     1/1     Running             0          22m
coredns-57995474d5-p9px9                     1/1     Running             0          22m
csi-cinder-controllerplugin-0                5/5     Running             0          22m
dashboard-metrics-scraper-7674b9d54f-5mk8h   0/1     Pending             0          22m
k8s-keystone-auth-jvtrg                      1/1     Running             0          22m
kube-dns-autoscaler-7967dcdbd7-xwvhv         0/1     Pending             0          22m
kubernetes-dashboard-bfc6ccfdf-ndq7n         0/1     Pending             0          22m
openstack-cloud-controller-manager-pcrhm     0/1     ContainerCreating   0          2s

[jenkins@snd-centos7-helm-4c-4g-276 ~]> kubectl get pods -n kube-system
NAME                                         READY   STATUS    RESTARTS   AGE
calico-kube-controllers-7b67cb9dd4-p2zmv     1/1     Running   0          24m
calico-node-2nsq4                            1/1     Running   0          21m
calico-node-72zd5                            1/1     Running   0          24m
calico-node-89xx6                            1/1     Running   0          21m
coredns-57995474d5-7bsnr                     1/1     Running   0          24m
coredns-57995474d5-p9px9                     1/1     Running   0          24m
csi-cinder-controllerplugin-0                5/5     Running   0          24m
csi-cinder-nodeplugin-kftm2                  2/2     Running   0          97s
csi-cinder-nodeplugin-lw67x                  2/2     Running   0          99s
dashboard-metrics-scraper-7674b9d54f-5mk8h   1/1     Running   0          24m
k8s-keystone-auth-jvtrg                      1/1     Running   0          24m
kube-dns-autoscaler-7967dcdbd7-xwvhv         1/1     Running   0          24m
kubernetes-dashboard-bfc6ccfdf-ndq7n         1/1     Running   0          24m
npd-2mxqn                                    1/1     Running   0          99s
npd-668qs                                    1/1     Running   0          97s
openstack-cloud-controller-manager-pcrhm     1/1     Running   0          107s

> kubectl get pods -A
NAMESPACE     NAME                                         READY   STATUS    RESTARTS   AGE
default       sdnc-opendaylight-0                          1/1     Running   0          18m
kube-system   calico-kube-controllers-7b67cb9dd4-p2zmv     1/1     Running   0          24m
kube-system   calico-node-2nsq4                            1/1     Running   0          22m
kube-system   calico-node-72zd5                            1/1     Running   0          24m
kube-system   calico-node-89xx6                            1/1     Running   0          22m
kube-system   coredns-57995474d5-7bsnr                     1/1     Running   0          24m
kube-system   coredns-57995474d5-p9px9                     1/1     Running   0          24m
kube-system   csi-cinder-controllerplugin-0                5/5     Running   0          24m
kube-system   csi-cinder-nodeplugin-kftm2                  2/2     Running   0          118s
kube-system   csi-cinder-nodeplugin-lw67x                  2/2     Running   0          2m
kube-system   dashboard-metrics-scraper-7674b9d54f-5mk8h   1/1     Running   0          24m
kube-system   k8s-keystone-auth-jvtrg                      1/1     Running   0          24m
kube-system   kube-dns-autoscaler-7967dcdbd7-xwvhv         1/1     Running   0          24m
kube-system   kubernetes-dashboard-bfc6ccfdf-ndq7n         1/1     Running   0          24m
kube-system   npd-2mxqn                                    1/1     Running   0          2m
kube-system   npd-668qs                                    1/1     Running   0          118s
kube-system   openstack-cloud-controller-manager-pcrhm     1/1     Running   0          2m8s


@askb
Copy link
Author

askb commented Dec 21, 2021

// start the sdnc with port forward:

> kubectl get svc
NAME                TYPE        CLUSTER-IP   EXTERNAL-IP   PORT(S)    AGE
kubernetes          ClusterIP   10.254.0.1   <none>        443/TCP    34m
sdnc-opendaylight   ClusterIP   None         <none>        8181/TCP   27m

[jenkins@snd-centos7-helm-4c-4g-276 ~]> export POD_NAME=$(kubectl get pods --namespace default -l "app.kubernetes.io/name=opendaylight,app.kubernetes.io/instance=sdnc" -o jsonpath="{.items[0].metadata.name}")
[jenkins@snd-centos7-helm-4c-4g-276 ~]> export CONTAINER_PORT=$(kubectl get pod --namespace default $POD_NAME -o jsonpath="{.spec.containers[0].ports[0].containerPort}")
[jenkins@snd-centos7-helm-4c-4g-276 ~]> kubectl --namespace default port-forward $POD_NAME 8080:$CONTAINER_PORT
Forwarding from 127.0.0.1:8080 -> 8181
Forwarding from [::1]:8080 -> 8181
Handling connection for 8080


@askb
Copy link
Author

askb commented Dec 21, 2021

// test the setup to connect to the SDNC and list the restconf modules

> curl -p -u admin:admin --request GET 'http://127.0.0.1:8080/restconf/modules' | jq
  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed
100  4259  100  4259    0     0   166k      0 --:--:-- --:--:-- --:--:--  173k
{
  "modules": {
    "module": [
      {
        "name": "prefix-shard-configuration",
        "revision": "2017-01-10",
        "namespace": "urn:opendaylight:params:xml:ns:yang:controller:md:sal:clustering:prefix-shard-configuration"
      },
      {
        "name": "odl-general-entity",
        "revision": "2015-09-30",
        "namespace": "urn:opendaylight:params:xml:ns:yang:mdsal:core:general-entity"
      },
      {
        "name": "aaa-password-service-config",
        "revision": "2017-06-19",
        "namespace": "urn:opendaylight:aaa:password:service:config"
      },
      {
        "name": "cluster-admin",
        "revision": "2015-10-13",
        "namespace": "urn:opendaylight:params:xml:ns:yang:controller:md:sal:cluster:admin"
      },
      {
        "name": "ietf-restconf",
        "revision": "2017-01-26",
        "namespace": "urn:ietf:params:xml:ns:yang:ietf-restconf"
      },
      {
        "name": "ietf-netconf",
        "revision": "2011-06-01",
        "namespace": "urn:ietf:params:xml:ns:netconf:base:1.0",
        "feature": [
          "confirmed-commit",
          "startup",
          "rollback-on-error",
          "validate",
          "url",
          "writable-running",
          "xpath",
          "candidate"
        ]
      },
      {
        "name": "distributed-datastore-provider",
        "revision": "2014-06-12",
        "namespace": "urn:opendaylight:params:xml:ns:yang:controller:config:distributed-datastore-provider"
      },
      {
        "name": "nc-notifications",
        "revision": "2008-07-14",
        "namespace": "urn:ietf:params:xml:ns:netmod:notification"
      },
      {
        "name": "aaa-cert",
        "revision": "2015-11-26",
        "namespace": "urn:opendaylight:yang:aaa:cert"
      },
      {
        "name": "sal-remote-augment",
        "revision": "2014-07-08",
        "namespace": "urn:sal:restconf:event:subscription"
      },
      {
        "name": "ietf-netconf-monitoring",
        "revision": "2010-10-04",
        "namespace": "urn:ietf:params:xml:ns:yang:ietf-netconf-monitoring"
      },
      {
        "name": "ietf-netconf-nmda",
        "revision": "2019-01-07",
        "namespace": "urn:ietf:params:xml:ns:yang:ietf-netconf-nmda",
        "feature": [
          "origin",
          "with-defaults"
        ]
      },
      {
        "name": "ietf-yang-metadata",
        "revision": "2016-08-05",
        "namespace": "urn:ietf:params:xml:ns:yang:ietf-yang-metadata"
      },
      {
        "name": "ietf-netconf-with-defaults",
        "revision": "2011-06-01",
        "namespace": "urn:ietf:params:xml:ns:yang:ietf-netconf-with-defaults"
      },
      {
        "name": "ietf-inet-types",
        "revision": "2013-07-15",
        "namespace": "urn:ietf:params:xml:ns:yang:ietf-inet-types"
      },
      {
        "name": "ietf-yang-library",
        "revision": "2019-01-04",
        "namespace": "urn:ietf:params:xml:ns:yang:ietf-yang-library"
      },
      {
        "name": "yang-ext",
        "revision": "2013-07-09",
        "namespace": "urn:opendaylight:yang:extension:yang-ext"
      },
      {
        "name": "ietf-yang-types",
        "revision": "2013-07-15",
        "namespace": "urn:ietf:params:xml:ns:yang:ietf-yang-types"
      },
      {
        "name": "aaa-encrypt-service-config",
        "revision": "2016-09-15",
        "namespace": "config:aaa:authn:encrypt:service:config"
      },
      {
        "name": "subscribe-to-notification",
        "revision": "2016-10-28",
        "namespace": "subscribe:to:notification"
      },
      {
        "name": "ietf-netconf-notifications",
        "revision": "2012-02-06",
        "namespace": "urn:ietf:params:xml:ns:yang:ietf-netconf-notifications"
      },
      {
        "name": "aaa-app-config",
        "revision": "2017-06-19",
        "namespace": "urn:opendaylight:aaa:app:config"
      },
      {
        "name": "odl-controller-cds-types",
        "revision": "2019-10-24",
        "namespace": "urn:opendaylight:params:xml:ns:yang:controller:cds:types"
      },
      {
        "name": "aaa",
        "revision": "2016-12-14",
        "namespace": "urn:opendaylight:params:xml:ns:yang:aaa"
      },
      {
        "name": "notifications",
        "revision": "2008-07-14",
        "namespace": "urn:ietf:params:xml:ns:netconf:notification:1.0"
      },
      {
        "name": "aaa-cert-mdsal",
        "revision": "2016-03-21",
        "namespace": "urn:opendaylight:yang:aaa:cert:mdsal"
      },
      {
        "name": "aaa-cert-rpc",
        "revision": "2015-12-15",
        "namespace": "urn:opendaylight:yang:aaa:cert:rpc"
      },
      {
        "name": "ietf-netconf-monitoring-extension",
        "revision": "2013-12-10",
        "namespace": "urn:ietf:params:xml:ns:yang:ietf-netconf-monitoring-extension"
      },
      {
        "name": "ietf-origin",
        "revision": "2018-02-14",
        "namespace": "urn:ietf:params:xml:ns:yang:ietf-origin"
      },
      {
        "name": "ietf-restconf",
        "revision": "2013-10-19",
        "namespace": "urn:ietf:params:xml:ns:yang:ietf-restconf"
      },
      {
        "name": "instance-identifier-patch-module",
        "revision": "2015-11-21",
        "namespace": "instance:identifier:patch:module"
      },
      {
        "name": "sal-remote",
        "revision": "2014-01-14",
        "namespace": "urn:opendaylight:params:xml:ns:yang:controller:md:sal:remote"
      },
      {
        "name": "ietf-datastores",
        "revision": "2018-02-14",
        "namespace": "urn:ietf:params:xml:ns:yang:ietf-datastores"
      },
      {
        "name": "entity-owners",
        "revision": "2015-08-04",
        "namespace": "urn:opendaylight:params:xml:ns:yang:controller:md:sal:clustering:entity-owners"
      },
      {
        "name": "ietf-restconf-monitoring",
        "revision": "2017-01-26",
        "namespace": "urn:ietf:params:xml:ns:yang:ietf-restconf-monitoring"
      }
    ]
  }
}

@askb
Copy link
Author

askb commented Dec 23, 2021

kubectl --namespace default port-forward $POD_NAME 8080:$CONTAINER_PORT
# Get all worker nodes
kubectl get node --selector='!node-role.kubernetes.io/master'
# Get all namespaces
kubectl get pods --all-namespaces
# Get all events by timestamps
kubectl get events --sort-by=.metadata.creationTimestamp
JSONPATH='{range .items[*]}{@.metadata.name}:{range @.status.conditions[*]}{@.type}={@.status};{end}{end}' \
>  && kubectl get nodes -o jsonpath="$JSONPATH" | grep "Ready=True"

# kubectl get nodes -o jsonpath='{range .items[*]} {.metadata.name} {" "} {.status.conditions[?(@.type=="Ready")].status} {" "} {.spec.taints} {"\n"} {end}'
#  sandbox-packaging-k8s-odl-depl-laucrtjp4kdp-master-0   True   [{"effect":"NoSchedule","key":"node-role.kubernetes.io/master"},{"effect":"NoSchedule","key":"node.cloudprovider.kubernetes.io/uninitialized","value":"true"}]
#   sandbox-packaging-k8s-odl-depl-laucrtjp4kdp-node-0   True   [{"effect":"NoSchedule","key":"node.cloudprovider.kubernetes.io/uninitialized","value":"true"}]
#   sandbox-packaging-k8s-odl-depl-laucrtjp4kdp-node-1   True   [{"effect":"NoSchedule","key":"node.cloudprovider.kubernetes.io/uninitialized","value":"true"}]

# Wait for pod to get ready before testing
#kubectl wait -n ${POD_NAME} --for=condition=ready pod --all

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment