Skip to content

Instantly share code, notes, and snippets.

@bot11
Created November 15, 2017 13:06
Show Gist options
  • Star 0 You must be signed in to star a gist
  • Fork 0 You must be signed in to fork a gist
  • Save bot11/dd298851daa3c49fe32d038a9fea8394 to your computer and use it in GitHub Desktop.
Save bot11/dd298851daa3c49fe32d038a9fea8394 to your computer and use it in GitHub Desktop.
kubelet logs
This file has been truncated, but you can view the full file.
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.228459 27740 flags.go:52] FLAG: --register-schedulable="true"
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.228465 27740 flags.go:52] FLAG: --register-with-taints="<nil>"
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.228472 27740 flags.go:52] FLAG: --registry-burst="10"
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.228476 27740 flags.go:52] FLAG: --registry-qps="5"
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.228482 27740 flags.go:52] FLAG: --require-kubeconfig="false"
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.228486 27740 flags.go:52] FLAG: --resolv-conf="/etc/resolv.conf"
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.228492 27740 flags.go:52] FLAG: --rkt-api-endpoint="localhost:15441"
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.228497 27740 flags.go:52] FLAG: --rkt-path=""
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.228501 27740 flags.go:52] FLAG: --rkt-stage1-image=""
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.228506 27740 flags.go:52] FLAG: --root-dir="/var/lib/kubelet"
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.228512 27740 flags.go:52] FLAG: --rotate-certificates="true"
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.228516 27740 flags.go:52] FLAG: --runonce="false"
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.228521 27740 flags.go:52] FLAG: --runtime-cgroups=""
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.228525 27740 flags.go:52] FLAG: --runtime-request-timeout="2m0s"
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.228530 27740 flags.go:52] FLAG: --seccomp-profile-root="/var/lib/kubelet/seccomp"a
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.228536 27740 flags.go:52] FLAG: --serialize-image-pulls="true"
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.228541 27740 flags.go:52] FLAG: --stderrthreshold="2"
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.228547 27740 flags.go:52] FLAG: --storage-driver-buffer-duration="1m0s"
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.228552 27740 flags.go:52] FLAG: --storage-driver-db="cadvisor"
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.228556 27740 flags.go:52] FLAG: --storage-driver-host="localhost:8086"
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.228561 27740 flags.go:52] FLAG: --storage-driver-password="root"
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.228565 27740 flags.go:52] FLAG: --storage-driver-secure="false"
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.228570 27740 flags.go:52] FLAG: --storage-driver-table="stats"
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.228575 27740 flags.go:52] FLAG: --storage-driver-user="root"
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.228580 27740 flags.go:52] FLAG: --streaming-connection-idle-timeout="4h0m0s"
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.228585 27740 flags.go:52] FLAG: --sync-frequency="1m0s"
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.228590 27740 flags.go:52] FLAG: --system-cgroups=""
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.228594 27740 flags.go:52] FLAG: --system-reserved=""
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.228599 27740 flags.go:52] FLAG: --system-reserved-cgroup=""
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.228604 27740 flags.go:52] FLAG: --tls-cert-file=""
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.228609 27740 flags.go:52] FLAG: --tls-private-key-file=""
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.228613 27740 flags.go:52] FLAG: --v="4"
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.228618 27740 flags.go:52] FLAG: --version="false"
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.228627 27740 flags.go:52] FLAG: --vmodule=""
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.228632 27740 flags.go:52] FLAG: --volume-plugin-dir="/usr/libexec/kubernetes/kubelet-plugins/volume/exec/"
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.228637 27740 flags.go:52] FLAG: --volume-stats-agg-period="1m0s"
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.228652 27740 feature_gate.go:156] feature gates: map[]
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.228675 27740 controller.go:114] kubelet config controller: starting controller
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.228681 27740 controller.go:118] kubelet config controller: validating combination of defaults and flags
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.968919 27740 server.go:564] Using self-signed cert (/var/lib/kubelet/pki/kubelet.crt, /var/lib/kubelet/pki/kubelet.key)
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.975253 27740 mount_linux.go:168] Detected OS with systemd
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.975273 27740 client.go:75] Connecting to docker on unix:///var/run/docker.sock
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.975292 27740 client.go:95] Start docker client with request timeout=2m0s
Nov 15 01:58:39 af867b kubelet[27740]: W1115 01:58:39.977808 27740 cni.go:196] Unable to update cni config: No networks found in /etc/cni/net.d
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.983386 27740 iptables.go:564] couldn't get iptables-restore version; assuming it doesn't support --wait
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.985167 27740 feature_gate.go:156] feature gates: map[]
Nov 15 01:58:39 af867b kubelet[27740]: W1115 01:58:39.985334 27740 server.go:289] --cloud-provider=auto-detect is deprecated. The desired cloud provider should be set explicitly
Nov 15 01:58:39 af867b kubelet[27740]: I1115 01:58:39.985370 27740 bootstrap.go:57] Using bootstrap kubeconfig to generate TLS client cert, key and kubeconfig file
Nov 15 01:58:39 af867b kubelet[27740]: error: failed to run Kubelet: unable to load bootstrap kubeconfig: stat /etc/kubernetes/bootstrap-kubelet.conf: no such file or directory
Nov 15 01:58:39 af867b systemd[1]: kubelet.service: main process exited, code=exited, status=1/FAILURE
Nov 15 01:58:39 af867b systemd[1]: Unit kubelet.service entered failed state.
Nov 15 01:58:39 af867b systemd[1]: kubelet.service failed.
Nov 15 01:58:50 af867b systemd[1]: kubelet.service holdoff time over, scheduling restart.
Nov 15 01:58:50 af867b systemd[1]: Started kubelet: The Kubernetes Node Agent.
Nov 15 01:58:50 af867b systemd[1]: Starting kubelet: The Kubernetes Node Agent...
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160304 27751 flags.go:52] FLAG: --address="0.0.0.0"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160386 27751 flags.go:52] FLAG: --allow-privileged="true"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160395 27751 flags.go:52] FLAG: --alsologtostderr="false"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160404 27751 flags.go:52] FLAG: --anonymous-auth="true"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160408 27751 flags.go:52] FLAG: --application-metrics-count-limit="100"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160413 27751 flags.go:52] FLAG: --authentication-token-webhook="false"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160418 27751 flags.go:52] FLAG: --authentication-token-webhook-cache-ttl="2m0s"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160426 27751 flags.go:52] FLAG: --authorization-mode="Webhook"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160433 27751 flags.go:52] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160438 27751 flags.go:52] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160442 27751 flags.go:52] FLAG: --azure-container-registry-config=""
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160447 27751 flags.go:52] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160453 27751 flags.go:52] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/bootstrap-kubelet.conf"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160458 27751 flags.go:52] FLAG: --cadvisor-port="0"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160466 27751 flags.go:52] FLAG: --cert-dir="/var/lib/kubelet/pki"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160470 27751 flags.go:52] FLAG: --cgroup-driver="cgroupfs"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160475 27751 flags.go:52] FLAG: --cgroup-root=""
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160479 27751 flags.go:52] FLAG: --cgroups-per-qos="true"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160483 27751 flags.go:52] FLAG: --chaos-chance="0"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160492 27751 flags.go:52] FLAG: --client-ca-file="/etc/kubernetes/pki/ca.crt"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160498 27751 flags.go:52] FLAG: --cloud-config=""
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160502 27751 flags.go:52] FLAG: --cloud-provider="auto-detect"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160507 27751 flags.go:52] FLAG: --cloud-provider-gce-lb-src-cidrs="130.211.0.0/22,35.191.0.0/16,209.85.152.0/22,209.85.204.0/22"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160515 27751 flags.go:52] FLAG: --cluster-dns="[10.96.0.10]"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160529 27751 flags.go:52] FLAG: --cluster-domain="cluster.local"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160535 27751 flags.go:52] FLAG: --cni-bin-dir="/opt/cni/bin"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160540 27751 flags.go:52] FLAG: --cni-conf-dir="/etc/cni/net.d"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160546 27751 flags.go:52] FLAG: --container-hints="/etc/cadvisor/container_hints.json"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160552 27751 flags.go:52] FLAG: --container-runtime="docker"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160557 27751 flags.go:52] FLAG: --container-runtime-endpoint="unix:///var/run/dockershim.sock"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160562 27751 flags.go:52] FLAG: --containerized="false"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160566 27751 flags.go:52] FLAG: --contention-profiling="false"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160570 27751 flags.go:52] FLAG: --cpu-cfs-quota="true"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160575 27751 flags.go:52] FLAG: --cpu-manager-policy="none"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160579 27751 flags.go:52] FLAG: --cpu-manager-reconcile-period="10s"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160584 27751 flags.go:52] FLAG: --docker="unix:///var/run/docker.sock"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160588 27751 flags.go:52] FLAG: --docker-disable-shared-pid="true"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160592 27751 flags.go:52] FLAG: --docker-endpoint="unix:///var/run/docker.sock"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160598 27751 flags.go:52] FLAG: --docker-env-metadata-whitelist=""
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160602 27751 flags.go:52] FLAG: --docker-exec-handler="native"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160606 27751 flags.go:52] FLAG: --docker-only="false"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160610 27751 flags.go:52] FLAG: --docker-root="/var/lib/docker"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160614 27751 flags.go:52] FLAG: --docker-tls="false"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160618 27751 flags.go:52] FLAG: --docker-tls-ca="ca.pem"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160622 27751 flags.go:52] FLAG: --docker-tls-cert="cert.pem"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160626 27751 flags.go:52] FLAG: --docker-tls-key="key.pem"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160630 27751 flags.go:52] FLAG: --dynamic-config-dir=""
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160638 27751 flags.go:52] FLAG: --enable-controller-attach-detach="true"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160642 27751 flags.go:52] FLAG: --enable-custom-metrics="false"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160646 27751 flags.go:52] FLAG: --enable-debugging-handlers="true"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160653 27751 flags.go:52] FLAG: --enable-load-reader="false"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160657 27751 flags.go:52] FLAG: --enable-server="true"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160662 27751 flags.go:52] FLAG: --enforce-node-allocatable="[pods]"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160674 27751 flags.go:52] FLAG: --event-burst="10"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160680 27751 flags.go:52] FLAG: --event-qps="5"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160685 27751 flags.go:52] FLAG: --event-storage-age-limit="default=0"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160690 27751 flags.go:52] FLAG: --event-storage-event-limit="default=0"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160695 27751 flags.go:52] FLAG: --eviction-hard="memory.available<100Mi,nodefs.available<10%,nodefs.inodesFree<5%"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160700 27751 flags.go:52] FLAG: --eviction-max-pod-grace-period="0"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160705 27751 flags.go:52] FLAG: --eviction-minimum-reclaim=""
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160743 27751 flags.go:52] FLAG: --eviction-pressure-transition-period="5m0s"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160750 27751 flags.go:52] FLAG: --eviction-soft=""
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160754 27751 flags.go:52] FLAG: --eviction-soft-grace-period=""
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160758 27751 flags.go:52] FLAG: --exit-on-lock-contention="false"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160763 27751 flags.go:52] FLAG: --experimental-allocatable-ignore-eviction="false"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160767 27751 flags.go:52] FLAG: --experimental-allowed-unsafe-sysctls="[]"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160775 27751 flags.go:52] FLAG: --experimental-bootstrap-kubeconfig="/etc/kubernetes/bootstrap-kubelet.conf"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160781 27751 flags.go:52] FLAG: --experimental-check-node-capabilities-before-mount="false"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160786 27751 flags.go:52] FLAG: --experimental-dockershim="false"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160791 27751 flags.go:52] FLAG: --experimental-dockershim-root-directory="/var/lib/dockershim"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160806 27751 flags.go:52] FLAG: --experimental-fail-swap-on="false"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160811 27751 flags.go:52] FLAG: --experimental-kernel-memcg-notification="false"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160815 27751 flags.go:52] FLAG: --experimental-mounter-path=""
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160819 27751 flags.go:52] FLAG: --experimental-qos-reserved=""
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160828 27751 flags.go:52] FLAG: --fail-swap-on="false"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160833 27751 flags.go:52] FLAG: --feature-gates=""
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160837 27751 flags.go:52] FLAG: --file-check-frequency="20s"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160842 27751 flags.go:52] FLAG: --global-housekeeping-interval="1m0s"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160847 27751 flags.go:52] FLAG: --google-json-key=""
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160851 27751 flags.go:52] FLAG: --hairpin-mode="promiscuous-bridge"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160856 27751 flags.go:52] FLAG: --healthz-bind-address="127.0.0.1"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160860 27751 flags.go:52] FLAG: --healthz-port="10248"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160864 27751 flags.go:52] FLAG: --host-ipc-sources="[*]"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160876 27751 flags.go:52] FLAG: --host-network-sources="[*]"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160886 27751 flags.go:52] FLAG: --host-pid-sources="[*]"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160895 27751 flags.go:52] FLAG: --hostname-override=""
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160899 27751 flags.go:52] FLAG: --housekeeping-interval="10s"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160904 27751 flags.go:52] FLAG: --http-check-frequency="20s"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160909 27751 flags.go:52] FLAG: --image-gc-high-threshold="85"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160918 27751 flags.go:52] FLAG: --image-gc-low-threshold="80"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160922 27751 flags.go:52] FLAG: --image-pull-progress-deadline="1m0s"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160927 27751 flags.go:52] FLAG: --image-service-endpoint=""
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160931 27751 flags.go:52] FLAG: --init-config-dir=""
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160935 27751 flags.go:52] FLAG: --iptables-drop-bit="15"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160939 27751 flags.go:52] FLAG: --iptables-masquerade-bit="14"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160944 27751 flags.go:52] FLAG: --keep-terminated-pod-volumes="false"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160948 27751 flags.go:52] FLAG: --kube-api-burst="10"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160952 27751 flags.go:52] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160957 27751 flags.go:52] FLAG: --kube-api-qps="5"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160961 27751 flags.go:52] FLAG: --kube-reserved=""
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160966 27751 flags.go:52] FLAG: --kube-reserved-cgroup=""
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160970 27751 flags.go:52] FLAG: --kubeconfig="/etc/kubernetes/kubelet.conf"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160976 27751 flags.go:52] FLAG: --kubelet-cgroups=""
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160981 27751 flags.go:52] FLAG: --lock-file=""
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160985 27751 flags.go:52] FLAG: --log-backtrace-at=":0"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160991 27751 flags.go:52] FLAG: --log-cadvisor-usage="false"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.160997 27751 flags.go:52] FLAG: --log-dir=""
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161001 27751 flags.go:52] FLAG: --log-flush-frequency="5s"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161005 27751 flags.go:52] FLAG: --logtostderr="true"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161009 27751 flags.go:52] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161015 27751 flags.go:52] FLAG: --make-iptables-util-chains="true"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161019 27751 flags.go:52] FLAG: --manifest-url=""
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161023 27751 flags.go:52] FLAG: --manifest-url-header=""
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161027 27751 flags.go:52] FLAG: --master-service-namespace="default"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161031 27751 flags.go:52] FLAG: --max-open-files="1000000"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161038 27751 flags.go:52] FLAG: --max-pods="110"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161043 27751 flags.go:52] FLAG: --maximum-dead-containers="-1"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161047 27751 flags.go:52] FLAG: --maximum-dead-containers-per-container="1"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161053 27751 flags.go:52] FLAG: --minimum-container-ttl-duration="0s"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161057 27751 flags.go:52] FLAG: --minimum-image-ttl-duration="2m0s"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161062 27751 flags.go:52] FLAG: --network-plugin="cni"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161066 27751 flags.go:52] FLAG: --network-plugin-dir=""
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161070 27751 flags.go:52] FLAG: --network-plugin-mtu="0"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161075 27751 flags.go:52] FLAG: --node-ip=""
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161079 27751 flags.go:52] FLAG: --node-labels=""
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161087 27751 flags.go:52] FLAG: --node-status-update-frequency="10s"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161092 27751 flags.go:52] FLAG: --non-masquerade-cidr="10.0.0.0/8"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161096 27751 flags.go:52] FLAG: --oom-score-adj="-999"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161100 27751 flags.go:52] FLAG: --pod-cidr=""
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161104 27751 flags.go:52] FLAG: --pod-infra-container-image="gcr.io/google_containers/pause-amd64:3.0"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161109 27751 flags.go:52] FLAG: --pod-manifest-path="/etc/kubernetes/manifests"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161113 27751 flags.go:52] FLAG: --pods-per-core="0"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161118 27751 flags.go:52] FLAG: --port="10250"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161122 27751 flags.go:52] FLAG: --protect-kernel-defaults="false"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161126 27751 flags.go:52] FLAG: --provider-id=""
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161130 27751 flags.go:52] FLAG: --read-only-port="10255"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161134 27751 flags.go:52] FLAG: --really-crash-for-testing="false"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161138 27751 flags.go:52] FLAG: --register-node="true"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161143 27751 flags.go:52] FLAG: --register-schedulable="true"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161148 27751 flags.go:52] FLAG: --register-with-taints="<nil>"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161155 27751 flags.go:52] FLAG: --registry-burst="10"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161160 27751 flags.go:52] FLAG: --registry-qps="5"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161164 27751 flags.go:52] FLAG: --require-kubeconfig="false"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161169 27751 flags.go:52] FLAG: --resolv-conf="/etc/resolv.conf"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161173 27751 flags.go:52] FLAG: --rkt-api-endpoint="localhost:15441"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161177 27751 flags.go:52] FLAG: --rkt-path=""
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161182 27751 flags.go:52] FLAG: --rkt-stage1-image=""
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161187 27751 flags.go:52] FLAG: --root-dir="/var/lib/kubelet"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161192 27751 flags.go:52] FLAG: --rotate-certificates="true"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161197 27751 flags.go:52] FLAG: --runonce="false"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161201 27751 flags.go:52] FLAG: --runtime-cgroups=""
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161205 27751 flags.go:52] FLAG: --runtime-request-timeout="2m0s"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161211 27751 flags.go:52] FLAG: --seccomp-profile-root="/var/lib/kubelet/seccomp"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161217 27751 flags.go:52] FLAG: --serialize-image-pulls="true"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161221 27751 flags.go:52] FLAG: --stderrthreshold="2"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161226 27751 flags.go:52] FLAG: --storage-driver-buffer-duration="1m0s"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161231 27751 flags.go:52] FLAG: --storage-driver-db="cadvisor"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161236 27751 flags.go:52] FLAG: --storage-driver-host="localhost:8086"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161240 27751 flags.go:52] FLAG: --storage-driver-password="root"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161245 27751 flags.go:52] FLAG: --storage-driver-secure="false"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161250 27751 flags.go:52] FLAG: --storage-driver-table="stats"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161254 27751 flags.go:52] FLAG: --storage-driver-user="root"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161259 27751 flags.go:52] FLAG: --streaming-connection-idle-timeout="4h0m0s"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161263 27751 flags.go:52] FLAG: --sync-frequency="1m0s"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161268 27751 flags.go:52] FLAG: --system-cgroups=""
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161272 27751 flags.go:52] FLAG: --system-reserved=""
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161276 27751 flags.go:52] FLAG: --system-reserved-cgroup=""
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161280 27751 flags.go:52] FLAG: --tls-cert-file=""
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161284 27751 flags.go:52] FLAG: --tls-private-key-file=""
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161288 27751 flags.go:52] FLAG: --v="4"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161292 27751 flags.go:52] FLAG: --version="false"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161306 27751 flags.go:52] FLAG: --vmodule=""
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161311 27751 flags.go:52] FLAG: --volume-plugin-dir="/usr/libexec/kubernetes/kubelet-plugins/volume/exec/"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161318 27751 flags.go:52] FLAG: --volume-stats-agg-period="1m0s"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161337 27751 feature_gate.go:156] feature gates: map[]
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161373 27751 controller.go:114] kubelet config controller: starting controller
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.161379 27751 controller.go:118] kubelet config controller: validating combination of defaults and flags
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.201042 27751 mount_linux.go:168] Detected OS with systemd
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.201061 27751 client.go:75] Connecting to docker on unix:///var/run/docker.sock
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.201079 27751 client.go:95] Start docker client with request timeout=2m0s
Nov 15 01:58:50 af867b kubelet[27751]: W1115 01:58:50.202238 27751 cni.go:196] Unable to update cni config: No networks found in /etc/cni/net.d
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.204753 27751 iptables.go:564] couldn't get iptables-restore version; assuming it doesn't support --wait
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.208194 27751 feature_gate.go:156] feature gates: map[]
Nov 15 01:58:50 af867b kubelet[27751]: W1115 01:58:50.208336 27751 server.go:289] --cloud-provider=auto-detect is deprecated. The desired cloud provider should be set explicitly
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.208379 27751 bootstrap.go:49] Kubeconfig /etc/kubernetes/kubelet.conf exists, skipping bootstrap
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.211728 27751 server.go:350] Starting client certificate rotation.
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.211751 27751 certificate_manager.go:192] Certificate rotation is enabled.
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.211762 27751 certificate_manager.go:322] Certificate rotation deadline is 2018-09-10 17:20:34.010209172 +0000 UTC
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.211779 27751 certificate_manager.go:200] shouldRotate() is true, forcing immediate rotation
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.211783 27751 certificate_manager.go:272] Rotating certificates
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.238095 27751 certificate_manager.go:361] Requesting new certificate.
Nov 15 01:58:50 af867b kubelet[27751]: E1115 01:58:50.238957 27751 certificate_manager.go:284] Failed while requesting a signed certificate from the master: cannot create certificate signing request: Post https://10.241.226.117:6443/apis/certificates.k8s.io/v1beta1/certificatesigningrequests: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.239616 27751 manager.go:149] cAdvisor running in container: "/sys/fs/cgroup/cpu,cpuacct/system.slice/kubelet.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.239812 27751 certificate_manager.go:214] Waiting 7191h21m43.770412212s for next certificate rotation
Nov 15 01:58:50 af867b kubelet[27751]: W1115 01:58:50.291243 27751 manager.go:157] unable to connect to Rkt api service: rkt: cannot tcp Dial rkt api service: dial tcp [::1]:15441: getsockopt: connection refused
Nov 15 01:58:50 af867b kubelet[27751]: W1115 01:58:50.291366 27751 manager.go:166] unable to connect to CRI-O api service: Get http://%2Fvar%2Frun%2Fcrio.sock/info: dial unix /var/run/crio.sock: connect: no such file or directory
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.353207 27751 fs.go:139] Filesystem UUIDs: map[54b776b7-9cad-4499-83eb-44a283cbe533:/dev/dm-2 76116fc0-ac1e-4350-a196-8b2a40745a21:/dev/dm-0 938071eb-ffd8-471a-883b-a569092d96df:/dev/xvda1 bd4931fb-de32-4ef5-9b28-80c248c5732b:/dev/dm-1]
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.353247 27751 fs.go:140] Filesystem partitions: map[tmpfs:{mountpoint:/dev/shm major:0 minor:19 fsType:tmpfs blockSize:0} /dev/mapper/vg_main-lv_root:{mountpoint:/ major:251 minor:0 fsType:xfs blockSize:0} /dev/mapper/vg_main-lv_appVolume:{mountpoint:/u01/applicationSpace major:251 minor:2 fsType:ext4 blockSize:0} /dev/xvda1:{mountpoint:/boot major:202 minor:1 fsType:xfs blockSize:0} vg_main-lv_docker:{mountpoint: major:251 minor:4 fsType:devicemapper blockSize:1024}]
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.357147 27751 manager.go:216] Machine: {NumCores:2 CpuFrequency:2693566 MemoryCapacity:7570800640 HugePages:[{PageSize:2048 NumPages:0}] MachineID:971c640a83aa4477b9ced5d696a8368d SystemUUID:E17D452D-518E-4CA0-86A9-E10C39526A0E BootID:622e4334-9a44-481c-9fec-23e22f9c6fa7 Filesystems:[{Device:/dev/xvda1 DeviceMajor:202 DeviceMinor:1 Capacity:520794112 Type:vfs Inodes:512000 HasInodes:true} {Device:vg_main-lv_docker DeviceMajor:251 DeviceMinor:4 Capacity:48318382080 Type:devicemapper Inodes:0 HasInodes:false} {Device:tmpfs DeviceMajor:0 DeviceMinor:19 Capacity:3785400320 Type:vfs Inodes:924170 HasInodes:true} {Device:/dev/mapper/vg_main-lv_root DeviceMajor:251 DeviceMinor:0 Capacity:10693378048 Type:vfs Inodes:10452992 HasInodes:true} {Device:/dev/mapper/vg_main-lv_appVolume DeviceMajor:251 DeviceMinor:2 Capacity:21003628544 Type:vfs Inodes:1310720 HasInodes:true}] DiskMap:map[251:0:{Name:dm-0 Major:251 Minor:0 Size:10703863808 Scheduler:none} 251:1:{Name:dm-1 Major:251 Minor:1 Size:21218983936 Scheduler:none} 251:2:{Name:dm-2 Major:251 Minor:2 Size:21474836480 Scheduler:none} 251:3:{Name:dm-3 Major:251 Minor:3 Size:2147483648 Scheduler:none} 251:4:{Name:dm-4 Major:251 Minor:4 Size:48318382080 Scheduler:none} 251:5:{Name:dm-5 Major:251 Minor:5 Size:48318382080 Scheduler:none} 202:0:{Name:xvda Major:202 Minor:0 Size:107374182400 Scheduler:none}] NetworkDevices:[{Name:eth0 MacAddress:c6:b0:53:ea:c2:42 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:7570800640 Cores:[{Id:0 Threads:[0] Caches:[{Size:32768 Type:Data Level:1} {Size:32768 Type:Instruction Level:1} {Size:262144 Type:Unified Level:2} {Size:31457280 Type:Unified Level:3}]} {Id:1 Threads:[1] Caches:[{Size:32768 Type:Data Level:1} {Size:32768 Type:Instruction Level:1} {Size:262144 Type:Unified Level:2} {Size:31457280 Type:Unified Level:3}]}] Caches:[]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.359394 27751 manager.go:222] Version: {KernelVersion:4.1.12-61.1.33.el7uek.x86_64 ContainerOsVersion:Oracle Linux Server 7.2 DockerVersion:17.03.1-ce DockerAPIVersion:1.27 CadvisorVersion: CadvisorRevision:}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.359935 27751 server.go:229] Sending events to api server.
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.360010 27751 server.go:422] --cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.361549 27751 container_manager_linux.go:252] container manager verified user specified cgroup-root exists: /
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.361570 27751 container_manager_linux.go:257] Creating Container Manager object based on Node Config: {RuntimeCgroupsName: SystemCgroupsName: KubeletCgroupsName: ContainerRuntime:docker CgroupsPerQOS:true CgroupRoot:/ CgroupDriver:cgroupfs ProtectKernelDefaults:false NodeAllocatableConfig:{KubeReservedCgroupName: SystemReservedCgroupName: EnforceNodeAllocatable:map[pods:{}] KubeReserved:map[] SystemReserved:map[] HardEvictionThresholds:[{Signal:memory.available Operator:LessThan Value:{Quantity:100Mi Percentage:0} GracePeriod:0s MinReclaim:<nil>} {Signal:nodefs.available Operator:LessThan Value:{Quantity:<nil> Percentage:0.1} GracePeriod:0s MinReclaim:<nil>} {Signal:nodefs.inodesFree Operator:LessThan Value:{Quantity:<nil> Percentage:0.05} GracePeriod:0s MinReclaim:<nil>}]} ExperimentalQOSReserved:map[] ExperimentalCPUManagerPolicy:none ExperimentalCPUManagerReconcilePeriod:10s}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.361681 27751 container_manager_linux.go:288] Creating device plugin handler: false
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.361744 27751 oom_linux.go:65] attempting to set "/proc/self/oom_score_adj" to "-999"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.361792 27751 server.go:686] Using root directory: /var/lib/kubelet
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.361827 27751 kubelet.go:273] Adding manifest file: /etc/kubernetes/manifests
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.361863 27751 file.go:52] Watching path "/etc/kubernetes/manifests"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.361884 27751 kubelet.go:283] Watching apiserver
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.365063 27751 reflector.go:202] Starting reflector *v1.Node (0s) from k8s.io/kubernetes/pkg/kubelet/kubelet.go:422
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.365115 27751 reflector.go:240] Listing and watching *v1.Node from k8s.io/kubernetes/pkg/kubelet/kubelet.go:422
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.365530 27751 reflector.go:202] Starting reflector *v1.Pod (0s) from k8s.io/kubernetes/pkg/kubelet/config/apiserver.go:47
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.365543 27751 reflector.go:240] Listing and watching *v1.Pod from k8s.io/kubernetes/pkg/kubelet/config/apiserver.go:47
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.365766 27751 reflector.go:202] Starting reflector *v1.Service (0s) from k8s.io/kubernetes/pkg/kubelet/kubelet.go:413
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.365778 27751 reflector.go:240] Listing and watching *v1.Service from k8s.io/kubernetes/pkg/kubelet/kubelet.go:413
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.362695 27751 file.go:161] Reading manifest file "/etc/kubernetes/manifests/etcd.yaml"
Nov 15 01:58:50 af867b kubelet[27751]: E1115 01:58:50.372898 27751 reflector.go:205] k8s.io/kubernetes/pkg/kubelet/config/apiserver.go:47: Failed to list *v1.Pod: Get https://10.241.226.117:6443/api/v1/pods?fieldSelector=spec.nodeName%3Daf867b&resourceVersion=0: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:50 af867b kubelet[27751]: E1115 01:58:50.372981 27751 reflector.go:205] k8s.io/kubernetes/pkg/kubelet/kubelet.go:422: Failed to list *v1.Node: Get https://10.241.226.117:6443/api/v1/nodes?fieldSelector=metadata.name%3Daf867b&resourceVersion=0: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:50 af867b kubelet[27751]: E1115 01:58:50.374346 27751 reflector.go:205] k8s.io/kubernetes/pkg/kubelet/kubelet.go:413: Failed to list *v1.Service: Get https://10.241.226.117:6443/api/v1/services?resourceVersion=0: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.381809 27751 iptables.go:564] couldn't get iptables-restore version; assuming it doesn't support --wait
Nov 15 01:58:50 af867b kubelet[27751]: W1115 01:58:50.387510 27751 kubelet_network.go:69] Hairpin mode set to "promiscuous-bridge" but kubenet is not enabled, falling back to "hairpin-veth"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.387548 27751 kubelet.go:517] Hairpin mode set to "hairpin-veth"
Nov 15 01:58:50 af867b kubelet[27751]: W1115 01:58:50.387654 27751 cni.go:196] Unable to update cni config: No networks found in /etc/cni/net.d
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.387674 27751 plugins.go:187] Loaded network plugin "cni"
Nov 15 01:58:50 af867b kubelet[27751]: W1115 01:58:50.390982 27751 cni.go:196] Unable to update cni config: No networks found in /etc/cni/net.d
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.398376 27751 iptables.go:564] couldn't get iptables-restore version; assuming it doesn't support --wait
Nov 15 01:58:50 af867b kubelet[27751]: W1115 01:58:50.399259 27751 cni.go:196] Unable to update cni config: No networks found in /etc/cni/net.d
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.399273 27751 plugins.go:187] Loaded network plugin "cni"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.399283 27751 docker_service.go:207] Docker cri networking managed by cni
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.430353 27751 file.go:161] Reading manifest file "/etc/kubernetes/manifests/kube-apiserver.yaml"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.431369 27751 file.go:161] Reading manifest file "/etc/kubernetes/manifests/kube-controller-manager.yaml"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.432377 27751 file.go:161] Reading manifest file "/etc/kubernetes/manifests/kube-scheduler.yaml"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.433839 27751 config.go:282] Setting pods for source file
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.434076 27751 config.go:404] Receiving a new pod "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.434098 27751 config.go:404] Receiving a new pod "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.434108 27751 config.go:404] Receiving a new pod "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.434117 27751 config.go:404] Receiving a new pod "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.449163 27751 docker_service.go:224] Setting cgroupDriver to cgroupfs
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.450199 27751 docker_legacy.go:151] No legacy containers found, stop performing legacy cleanup.
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.450243 27751 kubelet.go:606] Starting the GRPC server for the docker CRI shim.
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.450262 27751 docker_server.go:51] Start dockershim grpc server
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.452075 27751 oom_linux.go:65] attempting to set "/proc/1401/oom_score_adj" to "-999"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.452120 27751 oom_linux.go:65] attempting to set "/proc/1443/oom_score_adj" to "-999"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.505521 27751 remote_runtime.go:43] Connecting to runtime service unix:///var/run/dockershim.sock
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.505737 27751 remote_image.go:40] Connecting to image service unix:///var/run/dockershim.sock
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.505987 27751 plugins.go:56] Registering credential provider: .dockercfg
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.507418 27751 kuberuntime_manager.go:177] Container runtime docker initialized, version: 17.03.1-ce, apiVersion: 1.27.0
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.507676 27751 plugins.go:420] Loaded volume plugin "kubernetes.io/aws-ebs"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.507691 27751 plugins.go:420] Loaded volume plugin "kubernetes.io/empty-dir"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.507701 27751 plugins.go:420] Loaded volume plugin "kubernetes.io/gce-pd"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.507727 27751 plugins.go:420] Loaded volume plugin "kubernetes.io/git-repo"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.507738 27751 plugins.go:420] Loaded volume plugin "kubernetes.io/host-path"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.507747 27751 plugins.go:420] Loaded volume plugin "kubernetes.io/nfs"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.507756 27751 plugins.go:420] Loaded volume plugin "kubernetes.io/secret"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.507766 27751 plugins.go:420] Loaded volume plugin "kubernetes.io/iscsi"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.507776 27751 plugins.go:420] Loaded volume plugin "kubernetes.io/glusterfs"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.507785 27751 plugins.go:420] Loaded volume plugin "kubernetes.io/rbd"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.507794 27751 plugins.go:420] Loaded volume plugin "kubernetes.io/cinder"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.507803 27751 plugins.go:420] Loaded volume plugin "kubernetes.io/quobyte"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.507812 27751 plugins.go:420] Loaded volume plugin "kubernetes.io/cephfs"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.507823 27751 plugins.go:420] Loaded volume plugin "kubernetes.io/downward-api"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.507831 27751 plugins.go:420] Loaded volume plugin "kubernetes.io/fc"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.507840 27751 plugins.go:420] Loaded volume plugin "kubernetes.io/flocker"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.507849 27751 plugins.go:420] Loaded volume plugin "kubernetes.io/azure-file"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.507859 27751 plugins.go:420] Loaded volume plugin "kubernetes.io/configmap"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.507868 27751 plugins.go:420] Loaded volume plugin "kubernetes.io/vsphere-volume"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.507878 27751 plugins.go:420] Loaded volume plugin "kubernetes.io/azure-disk"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.507888 27751 plugins.go:420] Loaded volume plugin "kubernetes.io/photon-pd"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.507897 27751 plugins.go:420] Loaded volume plugin "kubernetes.io/projected"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.507906 27751 plugins.go:420] Loaded volume plugin "kubernetes.io/portworx-volume"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.507915 27751 plugins.go:420] Loaded volume plugin "kubernetes.io/scaleio"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.507959 27751 plugins.go:420] Loaded volume plugin "kubernetes.io/local-volume"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.507974 27751 plugins.go:420] Loaded volume plugin "kubernetes.io/storageos"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.509091 27751 server.go:718] Started kubelet v1.8.2
Nov 15 01:58:50 af867b kubelet[27751]: E1115 01:58:50.509665 27751 kubelet.go:1234] Image garbage collection failed once. Stats initialization may not have completed yet: failed to get imageFs info: unable to find data for container /
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.511538 27751 mount_linux.go:535] Directory /var/lib/kubelet is already on a shared mount
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.511660 27751 kubelet_node_status.go:280] Setting node annotation to enable volume controller attach/detach
Nov 15 01:58:50 af867b kubelet[27751]: E1115 01:58:50.511981 27751 event.go:209] Unable to write event: 'Post https://10.241.226.117:6443/api/v1/namespaces/default/events: dial tcp 10.241.226.117:6443: getsockopt: connection refused' (may retry after sleeping)
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.512496 27751 server.go:128] Starting to listen on 0.0.0.0:10250
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.512961 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'Starting' Starting kubelet.
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.512993 27751 server.go:148] Starting to listen read-only on 0.0.0.0:10255
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.514002 27751 server.go:296] Adding debug handlers to kubelet server.
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.522216 27751 kubelet.go:1222] Container garbage collection succeeded
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.522623 27751 kubelet_node_status.go:443] Recording NodeHasSufficientDisk event message for node af867b
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.522646 27751 kubelet_node_status.go:443] Recording NodeHasSufficientMemory event message for node af867b
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.522656 27751 kubelet_node_status.go:443] Recording NodeHasNoDiskPressure event message for node af867b
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.523136 27751 node_container_manager.go:70] Attempting to enforce Node Allocatable with config: {KubeReservedCgroupName: SystemReservedCgroupName: EnforceNodeAllocatable:map[pods:{}] KubeReserved:map[] SystemReserved:map[] HardEvictionThresholds:[{Signal:memory.available Operator:LessThan Value:{Quantity:100Mi Percentage:0} GracePeriod:0s MinReclaim:<nil>} {Signal:nodefs.available Operator:LessThan Value:{Quantity:<nil> Percentage:0.1} GracePeriod:0s MinReclaim:<nil>} {Signal:nodefs.inodesFree Operator:LessThan Value:{Quantity:<nil> Percentage:0.05} GracePeriod:0s MinReclaim:<nil>}]}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.523207 27751 fs_resource_analyzer.go:66] Starting FS ResourceAnalyzer
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.523226 27751 status_manager.go:140] Starting to sync pod status with apiserver
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.523242 27751 kubelet.go:1768] Starting kubelet main sync loop.
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.523253 27751 kubelet.go:1779] skipping pod synchronization - [container runtime is down PLEG is not healthy: pleg was last seen active 2562047h47m16.854775807s ago; threshold is 3m0s]
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.523328 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientDisk' Node af867b status is now: NodeHasSufficientDisk
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.523343 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientMemory' Node af867b status is now: NodeHasSufficientMemory
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.523359 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasNoDiskPressure' Node af867b status is now: NodeHasNoDiskPressure
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.523454 27751 qos_container_manager_linux.go:320] [ContainerManager]: Updated QoS cgroup configuration
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.523528 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeAllocatableEnforced' Updated Node Allocatable limit across pods
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.523550 27751 container_manager_linux.go:440] [ContainerManager]: Adding periodic tasks for docker CRI integration
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.523607 27751 container_manager_linux.go:446] [ContainerManager]: Discovered runtime cgroups name: /system.slice/docker.service
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.523636 27751 oom_linux.go:65] attempting to set "/proc/27751/oom_score_adj" to "-999"
Nov 15 01:58:50 af867b kubelet[27751]: E1115 01:58:50.523724 27751 container_manager_linux.go:603] [ContainerManager]: Fail to get rootfs information unable to find data for container /
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.523755 27751 volume_manager.go:244] The desired_state_of_world populator starts
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.523760 27751 volume_manager.go:246] Starting Kubelet Volume Manager
Nov 15 01:58:50 af867b kubelet[27751]: W1115 01:58:50.525614 27751 cni.go:196] Unable to update cni config: No networks found in /etc/cni/net.d
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.525906 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 01:58:50 af867b kubelet[27751]: E1115 01:58:50.525928 27751 kubelet.go:2095] Container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.627912 27751 kubelet_node_status.go:280] Setting node annotation to enable volume controller attach/detach
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.631375 27751 kubelet_node_status.go:443] Recording NodeHasSufficientDisk event message for node af867b
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.631402 27751 kubelet_node_status.go:443] Recording NodeHasSufficientMemory event message for node af867b
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.631415 27751 kubelet_node_status.go:443] Recording NodeHasNoDiskPressure event message for node af867b
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.631439 27751 kubelet_node_status.go:83] Attempting to register node af867b
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.631747 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientDisk' Node af867b status is now: NodeHasSufficientDisk
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.631770 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientMemory' Node af867b status is now: NodeHasSufficientMemory
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.631784 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasNoDiskPressure' Node af867b status is now: NodeHasNoDiskPressure
Nov 15 01:58:50 af867b kubelet[27751]: E1115 01:58:50.634356 27751 kubelet_node_status.go:107] Unable to register node "af867b" with API server: Post https://10.241.226.117:6443/api/v1/nodes: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:50 af867b kubelet[27751]: E1115 01:58:50.656590 27751 factory.go:340] devicemapper filesystem stats will not be reported: unable to find thin_ls binary
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.657816 27751 factory.go:355] Registering Docker factory
Nov 15 01:58:50 af867b kubelet[27751]: W1115 01:58:50.657844 27751 manager.go:265] Registration of the rkt container factory failed: unable to communicate with Rkt api service: rkt: cannot tcp Dial rkt api service: dial tcp [::1]:15441: getsockopt: connection refused
Nov 15 01:58:50 af867b kubelet[27751]: W1115 01:58:50.657951 27751 manager.go:276] Registration of the crio container factory failed: Get http://%2Fvar%2Frun%2Fcrio.sock/info: dial unix /var/run/crio.sock: connect: no such file or directory
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.657963 27751 factory.go:54] Registering systemd factory
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.658096 27751 factory.go:86] Registering Raw factory
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.658225 27751 manager.go:1140] Started watching for new ooms in manager
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.658286 27751 factory.go:116] Factory "docker" was unable to handle container "/"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.658305 27751 factory.go:105] Error trying to work out if we can handle /: / not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.658310 27751 factory.go:116] Factory "systemd" was unable to handle container "/"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.658318 27751 factory.go:112] Using factory "raw" for container "/"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.658989 27751 manager.go:932] Added container: "/" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.659274 27751 handler.go:325] Added event &{/ 2017-11-14 17:24:13.606 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.659306 27751 manager.go:311] Starting recovery of all containers
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.668163 27751 container.go:409] Start housekeeping for container "/"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.688233 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/lvm2-lvmetad.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.688279 27751 factory.go:105] Error trying to work out if we can handle /system.slice/lvm2-lvmetad.service: /system.slice/lvm2-lvmetad.service not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.688287 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/lvm2-lvmetad.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.688295 27751 factory.go:112] Using factory "raw" for container "/system.slice/lvm2-lvmetad.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.688444 27751 manager.go:932] Added container: "/system.slice/lvm2-lvmetad.service" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.688534 27751 handler.go:325] Added event &{/system.slice/lvm2-lvmetad.service 2017-11-14 17:38:36.664194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.688570 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/network.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.688583 27751 factory.go:105] Error trying to work out if we can handle /system.slice/network.service: /system.slice/network.service not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.688588 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/network.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.688594 27751 factory.go:112] Using factory "raw" for container "/system.slice/network.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.688700 27751 manager.go:932] Added container: "/system.slice/network.service" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.688885 27751 handler.go:325] Added event &{/system.slice/network.service 2017-11-14 17:38:36.665194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.688907 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/system-serial\\x2dgetty.slice"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.688916 27751 factory.go:105] Error trying to work out if we can handle /system.slice/system-serial\x2dgetty.slice: /system.slice/system-serial\x2dgetty.slice not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.688932 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/system-serial\\x2dgetty.slice"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.688940 27751 factory.go:112] Using factory "raw" for container "/system.slice/system-serial\\x2dgetty.slice"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.689055 27751 manager.go:932] Added container: "/system.slice/system-serial\\x2dgetty.slice" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.689146 27751 handler.go:325] Added event &{/system.slice/system-serial\x2dgetty.slice 2017-11-14 17:38:36.667194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.689171 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/systemd-sysctl.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.689179 27751 factory.go:105] Error trying to work out if we can handle /system.slice/systemd-sysctl.service: /system.slice/systemd-sysctl.service not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.689184 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/systemd-sysctl.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.689190 27751 factory.go:112] Using factory "raw" for container "/system.slice/systemd-sysctl.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.689319 27751 manager.go:932] Added container: "/system.slice/systemd-sysctl.service" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.689408 27751 handler.go:325] Added event &{/system.slice/systemd-sysctl.service 2017-11-14 17:38:36.668194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.689444 27751 factory.go:116] Factory "docker" was unable to handle container "/kubepods/burstable/pod9d6dd5e700f66143c0b1a919b27a8a33"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.689455 27751 factory.go:105] Error trying to work out if we can handle /kubepods/burstable/pod9d6dd5e700f66143c0b1a919b27a8a33: /kubepods/burstable/pod9d6dd5e700f66143c0b1a919b27a8a33 not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.689460 27751 factory.go:116] Factory "systemd" was unable to handle container "/kubepods/burstable/pod9d6dd5e700f66143c0b1a919b27a8a33"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.689467 27751 factory.go:112] Using factory "raw" for container "/kubepods/burstable/pod9d6dd5e700f66143c0b1a919b27a8a33"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.689606 27751 manager.go:932] Added container: "/kubepods/burstable/pod9d6dd5e700f66143c0b1a919b27a8a33" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.689733 27751 handler.go:325] Added event &{/kubepods/burstable/pod9d6dd5e700f66143c0b1a919b27a8a33 2017-11-14 17:38:41.589194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.689758 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/dev-disk-by\\x2did-dm\\x2duuid\\x2dLVM\\x2d5qQyVBIei1sAiW92atVQlKpHgr5hO0wRiOcalnY9G5qZcIpq1wnIC3VtjIEfyLcn.swap"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.689771 27751 factory.go:105] Error trying to work out if we can handle /system.slice/dev-disk-by\x2did-dm\x2duuid\x2dLVM\x2d5qQyVBIei1sAiW92atVQlKpHgr5hO0wRiOcalnY9G5qZcIpq1wnIC3VtjIEfyLcn.swap: /system.slice/dev-disk-by\x2did-dm\x2duuid\x2dLVM\x2d5qQyVBIei1sAiW92atVQlKpHgr5hO0wRiOcalnY9G5qZcIpq1wnIC3VtjIEfyLcn.swap not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.689782 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/dev-disk-by\\x2did-dm\\x2duuid\\x2dLVM\\x2d5qQyVBIei1sAiW92atVQlKpHgr5hO0wRiOcalnY9G5qZcIpq1wnIC3VtjIEfyLcn.swap"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.689791 27751 factory.go:112] Using factory "raw" for container "/system.slice/dev-disk-by\\x2did-dm\\x2duuid\\x2dLVM\\x2d5qQyVBIei1sAiW92atVQlKpHgr5hO0wRiOcalnY9G5qZcIpq1wnIC3VtjIEfyLcn.swap"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.689922 27751 manager.go:932] Added container: "/system.slice/dev-disk-by\\x2did-dm\\x2duuid\\x2dLVM\\x2d5qQyVBIei1sAiW92atVQlKpHgr5hO0wRiOcalnY9G5qZcIpq1wnIC3VtjIEfyLcn.swap" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.690030 27751 handler.go:325] Added event &{/system.slice/dev-disk-by\x2did-dm\x2duuid\x2dLVM\x2d5qQyVBIei1sAiW92atVQlKpHgr5hO0wRiOcalnY9G5qZcIpq1wnIC3VtjIEfyLcn.swap 2017-11-14 17:38:36.661194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.690047 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/proc-xen.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.690053 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/proc-xen.mount", but ignoring.
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.690061 27751 manager.go:901] ignoring container "/system.slice/proc-xen.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.690076 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/system-selinux\\x2dpolicy\\x2dmigrate\\x2dlocal\\x2dchanges.slice"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.690085 27751 factory.go:105] Error trying to work out if we can handle /system.slice/system-selinux\x2dpolicy\x2dmigrate\x2dlocal\x2dchanges.slice: /system.slice/system-selinux\x2dpolicy\x2dmigrate\x2dlocal\x2dchanges.slice not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.690090 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/system-selinux\\x2dpolicy\\x2dmigrate\\x2dlocal\\x2dchanges.slice"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.690097 27751 factory.go:112] Using factory "raw" for container "/system.slice/system-selinux\\x2dpolicy\\x2dmigrate\\x2dlocal\\x2dchanges.slice"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.690214 27751 manager.go:932] Added container: "/system.slice/system-selinux\\x2dpolicy\\x2dmigrate\\x2dlocal\\x2dchanges.slice" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.690328 27751 handler.go:325] Added event &{/system.slice/system-selinux\x2dpolicy\x2dmigrate\x2dlocal\x2dchanges.slice 2017-11-14 17:38:36.667194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.690348 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/systemd-tmpfiles-setup.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.690356 27751 factory.go:105] Error trying to work out if we can handle /system.slice/systemd-tmpfiles-setup.service: /system.slice/systemd-tmpfiles-setup.service not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.690360 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/systemd-tmpfiles-setup.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.690366 27751 factory.go:112] Using factory "raw" for container "/system.slice/systemd-tmpfiles-setup.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.690492 27751 manager.go:932] Added container: "/system.slice/systemd-tmpfiles-setup.service" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.690591 27751 handler.go:325] Added event &{/system.slice/systemd-tmpfiles-setup.service 2017-11-14 17:38:36.668194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.690616 27751 factory.go:116] Factory "docker" was unable to handle container "/kubepods/besteffort/podb69bc062-c962-11e7-83ed-c6b053eac242"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.690624 27751 factory.go:105] Error trying to work out if we can handle /kubepods/besteffort/podb69bc062-c962-11e7-83ed-c6b053eac242: /kubepods/besteffort/podb69bc062-c962-11e7-83ed-c6b053eac242 not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.690629 27751 factory.go:116] Factory "systemd" was unable to handle container "/kubepods/besteffort/podb69bc062-c962-11e7-83ed-c6b053eac242"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.690634 27751 factory.go:112] Using factory "raw" for container "/kubepods/besteffort/podb69bc062-c962-11e7-83ed-c6b053eac242"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.690785 27751 manager.go:932] Added container: "/kubepods/besteffort/podb69bc062-c962-11e7-83ed-c6b053eac242" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.690891 27751 handler.go:325] Added event &{/kubepods/besteffort/podb69bc062-c962-11e7-83ed-c6b053eac242 2017-11-14 17:39:06.579194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.690911 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/system-getty.slice"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.690920 27751 factory.go:105] Error trying to work out if we can handle /system.slice/system-getty.slice: /system.slice/system-getty.slice not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.690925 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/system-getty.slice"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.690931 27751 factory.go:112] Using factory "raw" for container "/system.slice/system-getty.slice"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.691043 27751 manager.go:932] Added container: "/system.slice/system-getty.slice" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.691134 27751 handler.go:325] Added event &{/system.slice/system-getty.slice 2017-11-14 17:38:36.666194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.691152 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/systemd-udevd.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.691160 27751 factory.go:105] Error trying to work out if we can handle /system.slice/systemd-udevd.service: /system.slice/systemd-udevd.service not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.691164 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/systemd-udevd.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.691170 27751 factory.go:112] Using factory "raw" for container "/system.slice/systemd-udevd.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.691287 27751 manager.go:932] Added container: "/system.slice/systemd-udevd.service" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.691399 27751 handler.go:325] Added event &{/system.slice/systemd-udevd.service 2017-11-14 17:38:36.668194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.691414 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.691421 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper.mount", but ignoring.
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.691429 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.691451 27751 factory.go:116] Factory "docker" was unable to handle container "/kubepods/besteffort/podd76e26fba3bf2bfd215eb29011d55250"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.691458 27751 factory.go:105] Error trying to work out if we can handle /kubepods/besteffort/podd76e26fba3bf2bfd215eb29011d55250: /kubepods/besteffort/podd76e26fba3bf2bfd215eb29011d55250 not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.691463 27751 factory.go:116] Factory "systemd" was unable to handle container "/kubepods/besteffort/podd76e26fba3bf2bfd215eb29011d55250"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.691468 27751 factory.go:112] Using factory "raw" for container "/kubepods/besteffort/podd76e26fba3bf2bfd215eb29011d55250"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.691592 27751 manager.go:932] Added container: "/kubepods/besteffort/podd76e26fba3bf2bfd215eb29011d55250" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.691700 27751 handler.go:325] Added event &{/kubepods/besteffort/podd76e26fba3bf2bfd215eb29011d55250 2017-11-14 17:38:41.583194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.691735 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/boot.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.691743 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/boot.mount", but ignoring.
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.691750 27751 manager.go:901] ignoring container "/system.slice/boot.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.691758 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/chronyd.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.691765 27751 factory.go:105] Error trying to work out if we can handle /system.slice/chronyd.service: /system.slice/chronyd.service not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.691770 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/chronyd.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.691776 27751 factory.go:112] Using factory "raw" for container "/system.slice/chronyd.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.691905 27751 manager.go:932] Added container: "/system.slice/chronyd.service" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.692003 27751 handler.go:325] Added event &{/system.slice/chronyd.service 2017-11-14 17:38:36.660194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.692030 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/dev-disk-by\\x2duuid-bd4931fb\\x2dde32\\x2d4ef5\\x2d9b28\\x2d80c248c5732b.swap"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.692042 27751 factory.go:105] Error trying to work out if we can handle /system.slice/dev-disk-by\x2duuid-bd4931fb\x2dde32\x2d4ef5\x2d9b28\x2d80c248c5732b.swap: /system.slice/dev-disk-by\x2duuid-bd4931fb\x2dde32\x2d4ef5\x2d9b28\x2d80c248c5732b.swap not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.692049 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/dev-disk-by\\x2duuid-bd4931fb\\x2dde32\\x2d4ef5\\x2d9b28\\x2d80c248c5732b.swap"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.692057 27751 factory.go:112] Using factory "raw" for container "/system.slice/dev-disk-by\\x2duuid-bd4931fb\\x2dde32\\x2d4ef5\\x2d9b28\\x2d80c248c5732b.swap"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.692178 27751 manager.go:932] Added container: "/system.slice/dev-disk-by\\x2duuid-bd4931fb\\x2dde32\\x2d4ef5\\x2d9b28\\x2d80c248c5732b.swap" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.692276 27751 handler.go:325] Added event &{/system.slice/dev-disk-by\x2duuid-bd4931fb\x2dde32\x2d4ef5\x2d9b28\x2d80c248c5732b.swap 2017-11-14 17:38:36.661194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.692312 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/systemd-logind.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.692321 27751 factory.go:105] Error trying to work out if we can handle /system.slice/systemd-logind.service: /system.slice/systemd-logind.service not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.692326 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/systemd-logind.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.692332 27751 factory.go:112] Using factory "raw" for container "/system.slice/systemd-logind.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.692443 27751 manager.go:932] Added container: "/system.slice/systemd-logind.service" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.692546 27751 handler.go:325] Added event &{/system.slice/systemd-logind.service 2017-11-14 17:38:36.667194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.692565 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/systemd-user-sessions.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.692573 27751 factory.go:105] Error trying to work out if we can handle /system.slice/systemd-user-sessions.service: /system.slice/systemd-user-sessions.service not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.692578 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/systemd-user-sessions.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.692584 27751 factory.go:112] Using factory "raw" for container "/system.slice/systemd-user-sessions.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.692697 27751 manager.go:932] Added container: "/system.slice/systemd-user-sessions.service" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.692765 27751 container.go:409] Start housekeeping for container "/system.slice/system-selinux\\x2dpolicy\\x2dmigrate\\x2dlocal\\x2dchanges.slice"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.693195 27751 container.go:409] Start housekeeping for container "/system.slice/lvm2-lvmetad.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.693595 27751 container.go:409] Start housekeeping for container "/system.slice/network.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.694016 27751 container.go:409] Start housekeeping for container "/system.slice/system-serial\\x2dgetty.slice"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.694356 27751 container.go:409] Start housekeeping for container "/system.slice/systemd-sysctl.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.694722 27751 container.go:409] Start housekeeping for container "/kubepods/burstable/pod9d6dd5e700f66143c0b1a919b27a8a33"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.695126 27751 container.go:409] Start housekeeping for container "/system.slice/dev-disk-by\\x2did-dm\\x2duuid\\x2dLVM\\x2d5qQyVBIei1sAiW92atVQlKpHgr5hO0wRiOcalnY9G5qZcIpq1wnIC3VtjIEfyLcn.swap"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.695488 27751 container.go:409] Start housekeeping for container "/system.slice/systemd-udevd.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.695948 27751 container.go:409] Start housekeeping for container "/system.slice/systemd-tmpfiles-setup.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.696329 27751 container.go:409] Start housekeeping for container "/kubepods/besteffort/podb69bc062-c962-11e7-83ed-c6b053eac242"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.696852 27751 container.go:409] Start housekeeping for container "/system.slice/system-getty.slice"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.697217 27751 container.go:409] Start housekeeping for container "/system.slice/chronyd.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.697591 27751 container.go:409] Start housekeeping for container "/kubepods/besteffort/podd76e26fba3bf2bfd215eb29011d55250"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.698064 27751 container.go:409] Start housekeeping for container "/system.slice/dev-disk-by\\x2duuid-bd4931fb\\x2dde32\\x2d4ef5\\x2d9b28\\x2d80c248c5732b.swap"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.698566 27751 handler.go:325] Added event &{/system.slice/systemd-user-sessions.service 2017-11-14 17:38:36.669194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.698587 27751 factory.go:116] Factory "docker" was unable to handle container "/kubepods"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.698595 27751 factory.go:105] Error trying to work out if we can handle /kubepods: /kubepods not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.698600 27751 factory.go:116] Factory "systemd" was unable to handle container "/kubepods"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.698605 27751 factory.go:112] Using factory "raw" for container "/kubepods"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.698735 27751 manager.go:932] Added container: "/kubepods" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.698848 27751 handler.go:325] Added event &{/kubepods 2017-11-14 17:38:36.658194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.698867 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.698875 27751 factory.go:105] Error trying to work out if we can handle /system.slice: /system.slice not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.698880 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.698884 27751 factory.go:112] Using factory "raw" for container "/system.slice"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.698992 27751 manager.go:932] Added container: "/system.slice" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.699077 27751 handler.go:325] Added event &{/system.slice 2017-11-14 17:38:36.659194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.699094 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/dev-dm\\x2d1.swap"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.699101 27751 factory.go:105] Error trying to work out if we can handle /system.slice/dev-dm\x2d1.swap: /system.slice/dev-dm\x2d1.swap not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.699106 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/dev-dm\\x2d1.swap"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.699111 27751 factory.go:112] Using factory "raw" for container "/system.slice/dev-dm\\x2d1.swap"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.699229 27751 manager.go:932] Added container: "/system.slice/dev-dm\\x2d1.swap" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.699314 27751 handler.go:325] Added event &{/system.slice/dev-dm\x2d1.swap 2017-11-14 17:38:36.662194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.699331 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/docker.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.699338 27751 factory.go:105] Error trying to work out if we can handle /system.slice/docker.service: /system.slice/docker.service not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.699343 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/docker.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.699349 27751 factory.go:112] Using factory "raw" for container "/system.slice/docker.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.699457 27751 manager.go:932] Added container: "/system.slice/docker.service" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.699553 27751 handler.go:325] Added event &{/system.slice/docker.service 2017-11-14 17:38:36.663194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.699568 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/sys-kernel-debug.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.699574 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/sys-kernel-debug.mount", but ignoring.
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.699581 27751 manager.go:901] ignoring container "/system.slice/sys-kernel-debug.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.699587 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/u01-applicationSpace.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.699592 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/u01-applicationSpace.mount", but ignoring.
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.699600 27751 manager.go:901] ignoring container "/system.slice/u01-applicationSpace.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.699605 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/run-docker-netns-default.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.699610 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/run-docker-netns-default.mount", but ignoring.
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.699617 27751 manager.go:901] ignoring container "/system.slice/run-docker-netns-default.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.699627 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/systemd-update-utmp.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.699634 27751 factory.go:105] Error trying to work out if we can handle /system.slice/systemd-update-utmp.service: /system.slice/systemd-update-utmp.service not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.699652 27751 container.go:409] Start housekeeping for container "/system.slice/docker.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.700108 27751 container.go:409] Start housekeeping for container "/system.slice/systemd-logind.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.700489 27751 container.go:409] Start housekeeping for container "/system.slice/systemd-user-sessions.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.701250 27751 container.go:409] Start housekeeping for container "/system.slice"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.701936 27751 container.go:409] Start housekeeping for container "/kubepods"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.702372 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/systemd-update-utmp.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.702383 27751 factory.go:112] Using factory "raw" for container "/system.slice/systemd-update-utmp.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.702507 27751 manager.go:932] Added container: "/system.slice/systemd-update-utmp.service" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.702601 27751 handler.go:325] Added event &{/system.slice/systemd-update-utmp.service 2017-11-14 17:38:36.669194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.702617 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/-.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.702624 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/-.mount", but ignoring.
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.702631 27751 manager.go:901] ignoring container "/system.slice/-.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.702639 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/crond.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.702646 27751 factory.go:105] Error trying to work out if we can handle /system.slice/crond.service: /system.slice/crond.service not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.702651 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/crond.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.702656 27751 factory.go:112] Using factory "raw" for container "/system.slice/crond.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.702791 27751 manager.go:932] Added container: "/system.slice/crond.service" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.702877 27751 handler.go:325] Added event &{/system.slice/crond.service 2017-11-14 17:38:36.660194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.702896 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/kmod-static-nodes.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.702904 27751 factory.go:105] Error trying to work out if we can handle /system.slice/kmod-static-nodes.service: /system.slice/kmod-static-nodes.service not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.702909 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/kmod-static-nodes.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.702915 27751 factory.go:112] Using factory "raw" for container "/system.slice/kmod-static-nodes.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.703036 27751 manager.go:932] Added container: "/system.slice/kmod-static-nodes.service" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.703120 27751 handler.go:325] Added event &{/system.slice/kmod-static-nodes.service 2017-11-14 17:38:36.663194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.703144 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/rhel-readonly.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.703152 27751 factory.go:105] Error trying to work out if we can handle /system.slice/rhel-readonly.service: /system.slice/rhel-readonly.service not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.703157 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/rhel-readonly.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.703162 27751 factory.go:112] Using factory "raw" for container "/system.slice/rhel-readonly.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.703277 27751 manager.go:932] Added container: "/system.slice/rhel-readonly.service" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.703363 27751 handler.go:325] Added event &{/system.slice/rhel-readonly.service 2017-11-14 17:38:36.665194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.703384 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/dev-disk-by\\x2did-dm\\x2dname\\x2dvg_main\\x2dlv_swap.swap"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.703393 27751 factory.go:105] Error trying to work out if we can handle /system.slice/dev-disk-by\x2did-dm\x2dname\x2dvg_main\x2dlv_swap.swap: /system.slice/dev-disk-by\x2did-dm\x2dname\x2dvg_main\x2dlv_swap.swap not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.703397 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/dev-disk-by\\x2did-dm\\x2dname\\x2dvg_main\\x2dlv_swap.swap"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.703405 27751 factory.go:112] Using factory "raw" for container "/system.slice/dev-disk-by\\x2did-dm\\x2dname\\x2dvg_main\\x2dlv_swap.swap"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.703528 27751 manager.go:932] Added container: "/system.slice/dev-disk-by\\x2did-dm\\x2dname\\x2dvg_main\\x2dlv_swap.swap" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.703621 27751 handler.go:325] Added event &{/system.slice/dev-disk-by\x2did-dm\x2dname\x2dvg_main\x2dlv_swap.swap 2017-11-14 17:38:36.661194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.703635 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/dev-hugepages.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.703641 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/dev-hugepages.mount", but ignoring.
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.703649 27751 manager.go:901] ignoring container "/system.slice/dev-hugepages.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.703657 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/kubelet.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.703664 27751 factory.go:105] Error trying to work out if we can handle /system.slice/kubelet.service: /system.slice/kubelet.service not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.703669 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/kubelet.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.703674 27751 factory.go:112] Using factory "raw" for container "/system.slice/kubelet.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.703810 27751 manager.go:932] Added container: "/system.slice/kubelet.service" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.703901 27751 handler.go:325] Added event &{/system.slice/kubelet.service 2017-11-15 01:58:50.660194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.703918 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/sshd.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.703925 27751 factory.go:105] Error trying to work out if we can handle /system.slice/sshd.service: /system.slice/sshd.service not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.703930 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/sshd.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.703935 27751 factory.go:112] Using factory "raw" for container "/system.slice/sshd.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.704052 27751 manager.go:932] Added container: "/system.slice/sshd.service" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.704133 27751 handler.go:325] Added event &{/system.slice/sshd.service 2017-11-14 17:38:36.666194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.704150 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/systemd-fsck-root.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.704158 27751 factory.go:105] Error trying to work out if we can handle /system.slice/systemd-fsck-root.service: /system.slice/systemd-fsck-root.service not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.704163 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/systemd-fsck-root.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.704168 27751 factory.go:112] Using factory "raw" for container "/system.slice/systemd-fsck-root.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.704294 27751 manager.go:932] Added container: "/system.slice/systemd-fsck-root.service" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.704380 27751 handler.go:325] Added event &{/system.slice/systemd-fsck-root.service 2017-11-14 17:38:36.667194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.704397 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/systemd-random-seed.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.704406 27751 factory.go:105] Error trying to work out if we can handle /system.slice/systemd-random-seed.service: /system.slice/systemd-random-seed.service not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.704410 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/systemd-random-seed.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.704416 27751 factory.go:112] Using factory "raw" for container "/system.slice/systemd-random-seed.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.704527 27751 manager.go:932] Added container: "/system.slice/systemd-random-seed.service" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.704615 27751 handler.go:325] Added event &{/system.slice/systemd-random-seed.service 2017-11-14 17:38:36.668194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.704632 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/tuned.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.704639 27751 factory.go:105] Error trying to work out if we can handle /system.slice/tuned.service: /system.slice/tuned.service not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.704644 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/tuned.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.704650 27751 factory.go:112] Using factory "raw" for container "/system.slice/tuned.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.704890 27751 manager.go:932] Added container: "/system.slice/tuned.service" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.704984 27751 handler.go:325] Added event &{/system.slice/tuned.service 2017-11-14 17:38:36.669194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.705001 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/acpid.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.705009 27751 factory.go:105] Error trying to work out if we can handle /system.slice/acpid.service: /system.slice/acpid.service not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.705013 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/acpid.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.705018 27751 factory.go:112] Using factory "raw" for container "/system.slice/acpid.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.705138 27751 manager.go:932] Added container: "/system.slice/acpid.service" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.705226 27751 handler.go:325] Added event &{/system.slice/acpid.service 2017-11-14 17:38:36.660194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.705243 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/dev-vg_main-lv_swap.swap"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.705251 27751 factory.go:105] Error trying to work out if we can handle /system.slice/dev-vg_main-lv_swap.swap: /system.slice/dev-vg_main-lv_swap.swap not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.705257 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/dev-vg_main-lv_swap.swap"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.705263 27751 factory.go:112] Using factory "raw" for container "/system.slice/dev-vg_main-lv_swap.swap"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.705374 27751 manager.go:932] Added container: "/system.slice/dev-vg_main-lv_swap.swap" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.705459 27751 handler.go:325] Added event &{/system.slice/dev-vg_main-lv_swap.swap 2017-11-14 17:38:36.662194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.705478 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/systemd-journal-flush.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.705490 27751 factory.go:105] Error trying to work out if we can handle /system.slice/systemd-journal-flush.service: /system.slice/systemd-journal-flush.service not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.705496 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/systemd-journal-flush.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.705501 27751 factory.go:112] Using factory "raw" for container "/system.slice/systemd-journal-flush.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.705618 27751 manager.go:932] Added container: "/system.slice/systemd-journal-flush.service" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.705703 27751 handler.go:325] Added event &{/system.slice/systemd-journal-flush.service 2017-11-14 17:38:36.667194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.705754 27751 factory.go:116] Factory "docker" was unable to handle container "/kubepods/burstable/pod42253414d7c5f285b756a2243a4df250"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.705766 27751 factory.go:105] Error trying to work out if we can handle /kubepods/burstable/pod42253414d7c5f285b756a2243a4df250: /kubepods/burstable/pod42253414d7c5f285b756a2243a4df250 not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.705771 27751 factory.go:116] Factory "systemd" was unable to handle container "/kubepods/burstable/pod42253414d7c5f285b756a2243a4df250"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.705777 27751 factory.go:112] Using factory "raw" for container "/kubepods/burstable/pod42253414d7c5f285b756a2243a4df250"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.705911 27751 manager.go:932] Added container: "/kubepods/burstable/pod42253414d7c5f285b756a2243a4df250" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.706021 27751 handler.go:325] Added event &{/kubepods/burstable/pod42253414d7c5f285b756a2243a4df250 2017-11-14 17:38:41.571194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.706039 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/dm-event.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.706046 27751 factory.go:105] Error trying to work out if we can handle /system.slice/dm-event.service: /system.slice/dm-event.service not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.706050 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/dm-event.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.706056 27751 factory.go:112] Using factory "raw" for container "/system.slice/dm-event.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.706170 27751 manager.go:932] Added container: "/system.slice/dm-event.service" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.706256 27751 handler.go:325] Added event &{/system.slice/dm-event.service 2017-11-14 17:38:36.663194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.706278 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/opc-guest-agent.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.706285 27751 factory.go:105] Error trying to work out if we can handle /system.slice/opc-guest-agent.service: /system.slice/opc-guest-agent.service not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.706290 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/opc-guest-agent.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.706295 27751 factory.go:112] Using factory "raw" for container "/system.slice/opc-guest-agent.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.706402 27751 manager.go:932] Added container: "/system.slice/opc-guest-agent.service" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.706488 27751 handler.go:325] Added event &{/system.slice/opc-guest-agent.service 2017-11-14 17:38:36.665194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.706505 27751 factory.go:116] Factory "docker" was unable to handle container "/kubepods/burstable"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.706512 27751 factory.go:105] Error trying to work out if we can handle /kubepods/burstable: /kubepods/burstable not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.706517 27751 factory.go:116] Factory "systemd" was unable to handle container "/kubepods/burstable"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.706522 27751 factory.go:112] Using factory "raw" for container "/kubepods/burstable"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.706650 27751 manager.go:932] Added container: "/kubepods/burstable" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.706764 27751 handler.go:325] Added event &{/kubepods/burstable 2017-11-14 17:38:36.658194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.706785 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/dbus.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.706792 27751 factory.go:105] Error trying to work out if we can handle /system.slice/dbus.service: /system.slice/dbus.service not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.706797 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/dbus.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.706802 27751 factory.go:112] Using factory "raw" for container "/system.slice/dbus.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.706916 27751 manager.go:932] Added container: "/system.slice/dbus.service" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.707009 27751 handler.go:325] Added event &{/system.slice/dbus.service 2017-11-14 17:38:36.660194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.707033 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/irqbalance.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.707041 27751 factory.go:105] Error trying to work out if we can handle /system.slice/irqbalance.service: /system.slice/irqbalance.service not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.707046 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/irqbalance.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.707052 27751 factory.go:112] Using factory "raw" for container "/system.slice/irqbalance.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.707158 27751 manager.go:932] Added container: "/system.slice/irqbalance.service" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.707251 27751 handler.go:325] Added event &{/system.slice/irqbalance.service 2017-11-14 17:38:36.663194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.707267 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/lvm2-monitor.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.707275 27751 factory.go:105] Error trying to work out if we can handle /system.slice/lvm2-monitor.service: /system.slice/lvm2-monitor.service not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.707280 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/lvm2-monitor.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.707285 27751 factory.go:112] Using factory "raw" for container "/system.slice/lvm2-monitor.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.707403 27751 manager.go:932] Added container: "/system.slice/lvm2-monitor.service" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.707487 27751 handler.go:325] Added event &{/system.slice/lvm2-monitor.service 2017-11-14 17:38:36.665194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.707504 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/serial_console.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.707512 27751 factory.go:105] Error trying to work out if we can handle /system.slice/serial_console.service: /system.slice/serial_console.service not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.707517 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/serial_console.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.707522 27751 factory.go:112] Using factory "raw" for container "/system.slice/serial_console.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.707647 27751 manager.go:932] Added container: "/system.slice/serial_console.service" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.707756 27751 handler.go:325] Added event &{/system.slice/serial_console.service 2017-11-14 17:38:36.666194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.707776 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/systemd-remount-fs.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.707784 27751 factory.go:105] Error trying to work out if we can handle /system.slice/systemd-remount-fs.service: /system.slice/systemd-remount-fs.service not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.707789 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/systemd-remount-fs.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.707794 27751 factory.go:112] Using factory "raw" for container "/system.slice/systemd-remount-fs.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.707914 27751 manager.go:932] Added container: "/system.slice/systemd-remount-fs.service" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.708000 27751 handler.go:325] Added event &{/system.slice/systemd-remount-fs.service 2017-11-14 17:38:36.668194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.708022 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/systemd-vconsole-setup.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.708031 27751 factory.go:105] Error trying to work out if we can handle /system.slice/systemd-vconsole-setup.service: /system.slice/systemd-vconsole-setup.service not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.708035 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/systemd-vconsole-setup.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.708042 27751 factory.go:112] Using factory "raw" for container "/system.slice/systemd-vconsole-setup.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.708160 27751 manager.go:932] Added container: "/system.slice/systemd-vconsole-setup.service" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.708253 27751 handler.go:325] Added event &{/system.slice/systemd-vconsole-setup.service 2017-11-14 17:38:36.669194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.708269 27751 factory.go:116] Factory "docker" was unable to handle container "/user.slice"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.708276 27751 factory.go:105] Error trying to work out if we can handle /user.slice: /user.slice not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.708281 27751 factory.go:116] Factory "systemd" was unable to handle container "/user.slice"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.708286 27751 factory.go:112] Using factory "raw" for container "/user.slice"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.708393 27751 manager.go:932] Added container: "/user.slice" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.708475 27751 handler.go:325] Added event &{/user.slice 2017-11-14 17:38:36.670194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.708494 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/dev-mapper-vg_main\\x2dlv_swap.swap"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.708504 27751 factory.go:105] Error trying to work out if we can handle /system.slice/dev-mapper-vg_main\x2dlv_swap.swap: /system.slice/dev-mapper-vg_main\x2dlv_swap.swap not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.708509 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/dev-mapper-vg_main\\x2dlv_swap.swap"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.708515 27751 factory.go:112] Using factory "raw" for container "/system.slice/dev-mapper-vg_main\\x2dlv_swap.swap"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.708632 27751 manager.go:932] Added container: "/system.slice/dev-mapper-vg_main\\x2dlv_swap.swap" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.708926 27751 container.go:409] Start housekeeping for container "/system.slice/dev-dm\\x2d1.swap"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.709277 27751 container.go:409] Start housekeeping for container "/system.slice/dev-vg_main-lv_swap.swap"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.709645 27751 container.go:409] Start housekeeping for container "/system.slice/systemd-update-utmp.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.710030 27751 container.go:409] Start housekeeping for container "/system.slice/crond.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.710410 27751 container.go:409] Start housekeeping for container "/system.slice/kmod-static-nodes.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.710785 27751 container.go:409] Start housekeeping for container "/system.slice/rhel-readonly.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.711116 27751 container.go:409] Start housekeeping for container "/system.slice/dev-disk-by\\x2did-dm\\x2dname\\x2dvg_main\\x2dlv_swap.swap"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.711470 27751 container.go:409] Start housekeeping for container "/system.slice/kubelet.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.711871 27751 container.go:409] Start housekeeping for container "/system.slice/sshd.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.712251 27751 container.go:409] Start housekeeping for container "/system.slice/systemd-fsck-root.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.712609 27751 container.go:409] Start housekeeping for container "/system.slice/systemd-random-seed.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.712787 27751 handler.go:325] Added event &{/system.slice/dev-mapper-vg_main\x2dlv_swap.swap 2017-11-14 17:38:36.662194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.712809 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/rsyslog.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.712817 27751 factory.go:105] Error trying to work out if we can handle /system.slice/rsyslog.service: /system.slice/rsyslog.service not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.712822 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/rsyslog.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.712827 27751 factory.go:112] Using factory "raw" for container "/system.slice/rsyslog.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.712944 27751 manager.go:932] Added container: "/system.slice/rsyslog.service" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.713029 27751 handler.go:325] Added event &{/system.slice/rsyslog.service 2017-11-14 17:38:36.665194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.713045 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/run-user-1000.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.713051 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/run-user-1000.mount", but ignoring.
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.713058 27751 manager.go:901] ignoring container "/system.slice/run-user-1000.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.713068 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/systemd-udev-trigger.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.713075 27751 factory.go:105] Error trying to work out if we can handle /system.slice/systemd-udev-trigger.service: /system.slice/systemd-udev-trigger.service not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.713080 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/systemd-udev-trigger.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.713085 27751 factory.go:112] Using factory "raw" for container "/system.slice/systemd-udev-trigger.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.713202 27751 manager.go:932] Added container: "/system.slice/systemd-udev-trigger.service" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.713296 27751 handler.go:325] Added event &{/system.slice/systemd-udev-trigger.service 2017-11-14 17:38:36.668194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.713314 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/systemd-journald.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.713322 27751 factory.go:105] Error trying to work out if we can handle /system.slice/systemd-journald.service: /system.slice/systemd-journald.service not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.713326 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/systemd-journald.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.713332 27751 factory.go:112] Using factory "raw" for container "/system.slice/systemd-journald.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.713446 27751 manager.go:932] Added container: "/system.slice/systemd-journald.service" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.713539 27751 handler.go:325] Added event &{/system.slice/systemd-journald.service 2017-11-14 17:38:36.667194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.713558 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/systemd-tmpfiles-setup-dev.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.713566 27751 factory.go:105] Error trying to work out if we can handle /system.slice/systemd-tmpfiles-setup-dev.service: /system.slice/systemd-tmpfiles-setup-dev.service not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.713570 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/systemd-tmpfiles-setup-dev.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.713577 27751 factory.go:112] Using factory "raw" for container "/system.slice/systemd-tmpfiles-setup-dev.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.713690 27751 manager.go:932] Added container: "/system.slice/systemd-tmpfiles-setup-dev.service" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.713830 27751 handler.go:325] Added event &{/system.slice/systemd-tmpfiles-setup-dev.service 2017-11-14 17:38:36.668194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.713848 27751 factory.go:116] Factory "docker" was unable to handle container "/kubepods/besteffort"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.713856 27751 factory.go:105] Error trying to work out if we can handle /kubepods/besteffort: /kubepods/besteffort not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.713860 27751 factory.go:116] Factory "systemd" was unable to handle container "/kubepods/besteffort"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.713866 27751 factory.go:112] Using factory "raw" for container "/kubepods/besteffort"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.713976 27751 manager.go:932] Added container: "/kubepods/besteffort" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.714077 27751 handler.go:325] Added event &{/kubepods/besteffort 2017-11-14 17:38:36.658194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.714091 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/dev-mqueue.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.714097 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/dev-mqueue.mount", but ignoring.
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.714104 27751 manager.go:901] ignoring container "/system.slice/dev-mqueue.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.714113 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/rhel-dmesg.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.714120 27751 factory.go:105] Error trying to work out if we can handle /system.slice/rhel-dmesg.service: /system.slice/rhel-dmesg.service not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.714124 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/rhel-dmesg.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.714129 27751 factory.go:112] Using factory "raw" for container "/system.slice/rhel-dmesg.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.714240 27751 manager.go:932] Added container: "/system.slice/rhel-dmesg.service" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.714331 27751 handler.go:325] Added event &{/system.slice/rhel-dmesg.service 2017-11-14 17:38:36.665194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.714346 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/sys-kernel-config.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.714352 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/sys-kernel-config.mount", but ignoring.
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.714359 27751 manager.go:901] ignoring container "/system.slice/sys-kernel-config.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.714369 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/system-lvm2\\x2dpvscan.slice"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.714376 27751 factory.go:105] Error trying to work out if we can handle /system.slice/system-lvm2\x2dpvscan.slice: /system.slice/system-lvm2\x2dpvscan.slice not handled by systemd handler
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.714381 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/system-lvm2\\x2dpvscan.slice"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.714386 27751 factory.go:112] Using factory "raw" for container "/system.slice/system-lvm2\\x2dpvscan.slice"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.714505 27751 manager.go:932] Added container: "/system.slice/system-lvm2\\x2dpvscan.slice" (aliases: [], namespace: "")
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.714603 27751 handler.go:325] Added event &{/system.slice/system-lvm2\x2dpvscan.slice 2017-11-14 17:38:36.666194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.714617 27751 manager.go:316] Recovery completed
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.720792 27751 container.go:409] Start housekeeping for container "/system.slice/system-lvm2\\x2dpvscan.slice"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.721340 27751 container.go:409] Start housekeeping for container "/system.slice/systemd-journal-flush.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.721800 27751 container.go:409] Start housekeeping for container "/kubepods/burstable/pod42253414d7c5f285b756a2243a4df250"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.722333 27751 container.go:409] Start housekeeping for container "/system.slice/dm-event.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.722893 27751 container.go:409] Start housekeeping for container "/system.slice/opc-guest-agent.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.723385 27751 container.go:409] Start housekeeping for container "/kubepods/burstable"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.723986 27751 container.go:409] Start housekeeping for container "/system.slice/dbus.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.726465 27751 container.go:409] Start housekeeping for container "/system.slice/irqbalance.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.727009 27751 container.go:409] Start housekeeping for container "/system.slice/lvm2-monitor.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.727473 27751 container.go:409] Start housekeeping for container "/system.slice/serial_console.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.727992 27751 container.go:409] Start housekeeping for container "/system.slice/systemd-remount-fs.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.728558 27751 container.go:409] Start housekeeping for container "/system.slice/systemd-vconsole-setup.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.728848 27751 container.go:409] Start housekeeping for container "/system.slice/tuned.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.729248 27751 container.go:409] Start housekeeping for container "/system.slice/acpid.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.732742 27751 container.go:409] Start housekeeping for container "/user.slice"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.733135 27751 container.go:409] Start housekeeping for container "/system.slice/dev-mapper-vg_main\\x2dlv_swap.swap"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.733460 27751 container.go:409] Start housekeeping for container "/system.slice/rsyslog.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.733835 27751 container.go:409] Start housekeeping for container "/system.slice/systemd-udev-trigger.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.734167 27751 container.go:409] Start housekeeping for container "/system.slice/systemd-journald.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.734531 27751 container.go:409] Start housekeeping for container "/system.slice/systemd-tmpfiles-setup-dev.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.734885 27751 container.go:409] Start housekeeping for container "/kubepods/besteffort"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.735288 27751 container.go:409] Start housekeeping for container "/system.slice/rhel-dmesg.service"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.791991 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/boot.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.792033 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/boot.mount", but ignoring.
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.792045 27751 manager.go:901] ignoring container "/system.slice/boot.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.792053 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.792062 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper.mount", but ignoring.
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.792071 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.792079 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/sys-kernel-debug.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.792086 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/sys-kernel-debug.mount", but ignoring.
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.792095 27751 manager.go:901] ignoring container "/system.slice/sys-kernel-debug.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.792102 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/u01-applicationSpace.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.792109 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/u01-applicationSpace.mount", but ignoring.
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.792118 27751 manager.go:901] ignoring container "/system.slice/u01-applicationSpace.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.792125 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/-.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.792132 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/-.mount", but ignoring.
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.792141 27751 manager.go:901] ignoring container "/system.slice/-.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.792147 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/dev-hugepages.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.792155 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/dev-hugepages.mount", but ignoring.
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.792163 27751 manager.go:901] ignoring container "/system.slice/dev-hugepages.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.792171 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/proc-xen.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.792179 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/proc-xen.mount", but ignoring.
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.792188 27751 manager.go:901] ignoring container "/system.slice/proc-xen.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.792194 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/run-docker-netns-default.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.792202 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/run-docker-netns-default.mount", but ignoring.
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.792210 27751 manager.go:901] ignoring container "/system.slice/run-docker-netns-default.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.792217 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/dev-mqueue.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.792224 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/dev-mqueue.mount", but ignoring.
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.792233 27751 manager.go:901] ignoring container "/system.slice/dev-mqueue.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.792240 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/run-user-1000.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.792248 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/run-user-1000.mount", but ignoring.
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.792256 27751 manager.go:901] ignoring container "/system.slice/run-user-1000.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.792263 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/sys-kernel-config.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.792270 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/sys-kernel-config.mount", but ignoring.
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.792280 27751 manager.go:901] ignoring container "/system.slice/sys-kernel-config.mount"
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.792338 27751 eviction_manager.go:221] eviction manager: synchronize housekeeping
Nov 15 01:58:50 af867b kubelet[27751]: E1115 01:58:50.792446 27751 eviction_manager.go:238] eviction manager: unexpected err: failed to get node info: node 'af867b' not found
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.834549 27751 kubelet_node_status.go:280] Setting node annotation to enable volume controller attach/detach
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.837086 27751 kubelet_node_status.go:443] Recording NodeHasSufficientDisk event message for node af867b
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.837113 27751 kubelet_node_status.go:443] Recording NodeHasSufficientMemory event message for node af867b
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.837125 27751 kubelet_node_status.go:443] Recording NodeHasNoDiskPressure event message for node af867b
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.837141 27751 kubelet_node_status.go:83] Attempting to register node af867b
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.837408 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientDisk' Node af867b status is now: NodeHasSufficientDisk
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.837434 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientMemory' Node af867b status is now: NodeHasSufficientMemory
Nov 15 01:58:50 af867b kubelet[27751]: I1115 01:58:50.837448 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasNoDiskPressure' Node af867b status is now: NodeHasNoDiskPressure
Nov 15 01:58:50 af867b kubelet[27751]: E1115 01:58:50.839121 27751 kubelet_node_status.go:107] Unable to register node "af867b" with API server: Post https://10.241.226.117:6443/api/v1/nodes: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:51 af867b kubelet[27751]: I1115 01:58:51.239355 27751 kubelet_node_status.go:280] Setting node annotation to enable volume controller attach/detach
Nov 15 01:58:51 af867b kubelet[27751]: I1115 01:58:51.244096 27751 kubelet_node_status.go:443] Recording NodeHasSufficientDisk event message for node af867b
Nov 15 01:58:51 af867b kubelet[27751]: I1115 01:58:51.244147 27751 kubelet_node_status.go:443] Recording NodeHasSufficientMemory event message for node af867b
Nov 15 01:58:51 af867b kubelet[27751]: I1115 01:58:51.244160 27751 kubelet_node_status.go:443] Recording NodeHasNoDiskPressure event message for node af867b
Nov 15 01:58:51 af867b kubelet[27751]: I1115 01:58:51.244180 27751 kubelet_node_status.go:83] Attempting to register node af867b
Nov 15 01:58:51 af867b kubelet[27751]: I1115 01:58:51.244501 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientDisk' Node af867b status is now: NodeHasSufficientDisk
Nov 15 01:58:51 af867b kubelet[27751]: I1115 01:58:51.244525 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientMemory' Node af867b status is now: NodeHasSufficientMemory
Nov 15 01:58:51 af867b kubelet[27751]: I1115 01:58:51.244539 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasNoDiskPressure' Node af867b status is now: NodeHasNoDiskPressure
Nov 15 01:58:51 af867b kubelet[27751]: E1115 01:58:51.244816 27751 kubelet_node_status.go:107] Unable to register node "af867b" with API server: Post https://10.241.226.117:6443/api/v1/nodes: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:51 af867b kubelet[27751]: I1115 01:58:51.373119 27751 reflector.go:240] Listing and watching *v1.Pod from k8s.io/kubernetes/pkg/kubelet/config/apiserver.go:47
Nov 15 01:58:51 af867b kubelet[27751]: E1115 01:58:51.374210 27751 reflector.go:205] k8s.io/kubernetes/pkg/kubelet/config/apiserver.go:47: Failed to list *v1.Pod: Get https://10.241.226.117:6443/api/v1/pods?fieldSelector=spec.nodeName%3Daf867b&resourceVersion=0: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:51 af867b kubelet[27751]: I1115 01:58:51.374699 27751 reflector.go:240] Listing and watching *v1.Node from k8s.io/kubernetes/pkg/kubelet/kubelet.go:422
Nov 15 01:58:51 af867b kubelet[27751]: E1115 01:58:51.375297 27751 reflector.go:205] k8s.io/kubernetes/pkg/kubelet/kubelet.go:422: Failed to list *v1.Node: Get https://10.241.226.117:6443/api/v1/nodes?fieldSelector=metadata.name%3Daf867b&resourceVersion=0: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:51 af867b kubelet[27751]: I1115 01:58:51.376957 27751 reflector.go:240] Listing and watching *v1.Service from k8s.io/kubernetes/pkg/kubelet/kubelet.go:413
Nov 15 01:58:51 af867b kubelet[27751]: E1115 01:58:51.377422 27751 reflector.go:205] k8s.io/kubernetes/pkg/kubelet/kubelet.go:413: Failed to list *v1.Service: Get https://10.241.226.117:6443/api/v1/services?resourceVersion=0: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:52 af867b kubelet[27751]: I1115 01:58:52.045079 27751 kubelet_node_status.go:280] Setting node annotation to enable volume controller attach/detach
Nov 15 01:58:52 af867b kubelet[27751]: I1115 01:58:52.048064 27751 kubelet_node_status.go:443] Recording NodeHasSufficientDisk event message for node af867b
Nov 15 01:58:52 af867b kubelet[27751]: I1115 01:58:52.048095 27751 kubelet_node_status.go:443] Recording NodeHasSufficientMemory event message for node af867b
Nov 15 01:58:52 af867b kubelet[27751]: I1115 01:58:52.048105 27751 kubelet_node_status.go:443] Recording NodeHasNoDiskPressure event message for node af867b
Nov 15 01:58:52 af867b kubelet[27751]: I1115 01:58:52.048123 27751 kubelet_node_status.go:83] Attempting to register node af867b
Nov 15 01:58:52 af867b kubelet[27751]: I1115 01:58:52.048440 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientDisk' Node af867b status is now: NodeHasSufficientDisk
Nov 15 01:58:52 af867b kubelet[27751]: I1115 01:58:52.048464 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientMemory' Node af867b status is now: NodeHasSufficientMemory
Nov 15 01:58:52 af867b kubelet[27751]: I1115 01:58:52.048479 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasNoDiskPressure' Node af867b status is now: NodeHasNoDiskPressure
Nov 15 01:58:52 af867b kubelet[27751]: E1115 01:58:52.051022 27751 kubelet_node_status.go:107] Unable to register node "af867b" with API server: Post https://10.241.226.117:6443/api/v1/nodes: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:52 af867b kubelet[27751]: I1115 01:58:52.374455 27751 reflector.go:240] Listing and watching *v1.Pod from k8s.io/kubernetes/pkg/kubelet/config/apiserver.go:47
Nov 15 01:58:52 af867b kubelet[27751]: I1115 01:58:52.375621 27751 reflector.go:240] Listing and watching *v1.Node from k8s.io/kubernetes/pkg/kubelet/kubelet.go:422
Nov 15 01:58:52 af867b kubelet[27751]: E1115 01:58:52.377170 27751 reflector.go:205] k8s.io/kubernetes/pkg/kubelet/kubelet.go:422: Failed to list *v1.Node: Get https://10.241.226.117:6443/api/v1/nodes?fieldSelector=metadata.name%3Daf867b&resourceVersion=0: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:52 af867b kubelet[27751]: E1115 01:58:52.377231 27751 reflector.go:205] k8s.io/kubernetes/pkg/kubelet/config/apiserver.go:47: Failed to list *v1.Pod: Get https://10.241.226.117:6443/api/v1/pods?fieldSelector=spec.nodeName%3Daf867b&resourceVersion=0: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:52 af867b kubelet[27751]: I1115 01:58:52.377944 27751 reflector.go:240] Listing and watching *v1.Service from k8s.io/kubernetes/pkg/kubelet/kubelet.go:413
Nov 15 01:58:52 af867b kubelet[27751]: E1115 01:58:52.378771 27751 reflector.go:205] k8s.io/kubernetes/pkg/kubelet/kubelet.go:413: Failed to list *v1.Service: Get https://10.241.226.117:6443/api/v1/services?resourceVersion=0: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:53 af867b kubelet[27751]: I1115 01:58:53.377441 27751 reflector.go:240] Listing and watching *v1.Node from k8s.io/kubernetes/pkg/kubelet/kubelet.go:422
Nov 15 01:58:53 af867b kubelet[27751]: E1115 01:58:53.378677 27751 reflector.go:205] k8s.io/kubernetes/pkg/kubelet/kubelet.go:422: Failed to list *v1.Node: Get https://10.241.226.117:6443/api/v1/nodes?fieldSelector=metadata.name%3Daf867b&resourceVersion=0: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:53 af867b kubelet[27751]: I1115 01:58:53.379207 27751 reflector.go:240] Listing and watching *v1.Pod from k8s.io/kubernetes/pkg/kubelet/config/apiserver.go:47
Nov 15 01:58:53 af867b kubelet[27751]: E1115 01:58:53.379755 27751 reflector.go:205] k8s.io/kubernetes/pkg/kubelet/config/apiserver.go:47: Failed to list *v1.Pod: Get https://10.241.226.117:6443/api/v1/pods?fieldSelector=spec.nodeName%3Daf867b&resourceVersion=0: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:53 af867b kubelet[27751]: I1115 01:58:53.380654 27751 reflector.go:240] Listing and watching *v1.Service from k8s.io/kubernetes/pkg/kubelet/kubelet.go:413
Nov 15 01:58:53 af867b kubelet[27751]: E1115 01:58:53.381163 27751 reflector.go:205] k8s.io/kubernetes/pkg/kubelet/kubelet.go:413: Failed to list *v1.Service: Get https://10.241.226.117:6443/api/v1/services?resourceVersion=0: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:53 af867b kubelet[27751]: I1115 01:58:53.651300 27751 kubelet_node_status.go:280] Setting node annotation to enable volume controller attach/detach
Nov 15 01:58:53 af867b kubelet[27751]: I1115 01:58:53.654073 27751 kubelet_node_status.go:443] Recording NodeHasSufficientDisk event message for node af867b
Nov 15 01:58:53 af867b kubelet[27751]: I1115 01:58:53.654107 27751 kubelet_node_status.go:443] Recording NodeHasSufficientMemory event message for node af867b
Nov 15 01:58:53 af867b kubelet[27751]: I1115 01:58:53.654119 27751 kubelet_node_status.go:443] Recording NodeHasNoDiskPressure event message for node af867b
Nov 15 01:58:53 af867b kubelet[27751]: I1115 01:58:53.654136 27751 kubelet_node_status.go:83] Attempting to register node af867b
Nov 15 01:58:53 af867b kubelet[27751]: I1115 01:58:53.654401 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientDisk' Node af867b status is now: NodeHasSufficientDisk
Nov 15 01:58:53 af867b kubelet[27751]: I1115 01:58:53.654425 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientMemory' Node af867b status is now: NodeHasSufficientMemory
Nov 15 01:58:53 af867b kubelet[27751]: I1115 01:58:53.654446 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasNoDiskPressure' Node af867b status is now: NodeHasNoDiskPressure
Nov 15 01:58:53 af867b kubelet[27751]: E1115 01:58:53.656627 27751 kubelet_node_status.go:107] Unable to register node "af867b" with API server: Post https://10.241.226.117:6443/api/v1/nodes: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:54 af867b kubelet[27751]: I1115 01:58:54.378951 27751 reflector.go:240] Listing and watching *v1.Node from k8s.io/kubernetes/pkg/kubelet/kubelet.go:422
Nov 15 01:58:54 af867b kubelet[27751]: E1115 01:58:54.380045 27751 reflector.go:205] k8s.io/kubernetes/pkg/kubelet/kubelet.go:422: Failed to list *v1.Node: Get https://10.241.226.117:6443/api/v1/nodes?fieldSelector=metadata.name%3Daf867b&resourceVersion=0: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:54 af867b kubelet[27751]: I1115 01:58:54.380337 27751 reflector.go:240] Listing and watching *v1.Pod from k8s.io/kubernetes/pkg/kubelet/config/apiserver.go:47
Nov 15 01:58:54 af867b kubelet[27751]: E1115 01:58:54.380834 27751 reflector.go:205] k8s.io/kubernetes/pkg/kubelet/config/apiserver.go:47: Failed to list *v1.Pod: Get https://10.241.226.117:6443/api/v1/pods?fieldSelector=spec.nodeName%3Daf867b&resourceVersion=0: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:54 af867b kubelet[27751]: I1115 01:58:54.381390 27751 reflector.go:240] Listing and watching *v1.Service from k8s.io/kubernetes/pkg/kubelet/kubelet.go:413
Nov 15 01:58:54 af867b kubelet[27751]: E1115 01:58:54.382064 27751 reflector.go:205] k8s.io/kubernetes/pkg/kubelet/kubelet.go:413: Failed to list *v1.Service: Get https://10.241.226.117:6443/api/v1/services?resourceVersion=0: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.380290 27751 reflector.go:240] Listing and watching *v1.Node from k8s.io/kubernetes/pkg/kubelet/kubelet.go:422
Nov 15 01:58:55 af867b kubelet[27751]: E1115 01:58:55.381601 27751 reflector.go:205] k8s.io/kubernetes/pkg/kubelet/kubelet.go:422: Failed to list *v1.Node: Get https://10.241.226.117:6443/api/v1/nodes?fieldSelector=metadata.name%3Daf867b&resourceVersion=0: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.382213 27751 reflector.go:240] Listing and watching *v1.Pod from k8s.io/kubernetes/pkg/kubelet/config/apiserver.go:47
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.382899 27751 reflector.go:240] Listing and watching *v1.Service from k8s.io/kubernetes/pkg/kubelet/kubelet.go:413
Nov 15 01:58:55 af867b kubelet[27751]: E1115 01:58:55.383112 27751 reflector.go:205] k8s.io/kubernetes/pkg/kubelet/config/apiserver.go:47: Failed to list *v1.Pod: Get https://10.241.226.117:6443/api/v1/pods?fieldSelector=spec.nodeName%3Daf867b&resourceVersion=0: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:55 af867b kubelet[27751]: E1115 01:58:55.383963 27751 reflector.go:205] k8s.io/kubernetes/pkg/kubelet/kubelet.go:413: Failed to list *v1.Service: Get https://10.241.226.117:6443/api/v1/services?resourceVersion=0: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.523483 27751 kubelet.go:1837] SyncLoop (ADD, "file"): "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9), kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373), etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250), kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.523580 27751 kubelet_node_status.go:280] Setting node annotation to enable volume controller attach/detach
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.526532 27751 kubelet_node_status.go:443] Recording NodeHasSufficientDisk event message for node af867b
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.526565 27751 kubelet_node_status.go:443] Recording NodeHasSufficientMemory event message for node af867b
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.526577 27751 kubelet_node_status.go:443] Recording NodeHasNoDiskPressure event message for node af867b
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.526697 27751 kubelet_node_status.go:280] Setting node annotation to enable volume controller attach/detach
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.526935 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasNoDiskPressure' Node af867b status is now: NodeHasNoDiskPressure
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.526959 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientDisk' Node af867b status is now: NodeHasSufficientDisk
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.526983 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientMemory' Node af867b status is now: NodeHasSufficientMemory
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.527047 27751 kubelet_pods.go:1284] Generating status for "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.527138 27751 kubelet_node_status.go:280] Setting node annotation to enable volume controller attach/detach
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.530596 27751 kubelet_node_status.go:443] Recording NodeHasSufficientDisk event message for node af867b
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.530625 27751 kubelet_node_status.go:443] Recording NodeHasSufficientMemory event message for node af867b
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.530638 27751 kubelet_node_status.go:443] Recording NodeHasNoDiskPressure event message for node af867b
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.530692 27751 kubelet_node_status.go:280] Setting node annotation to enable volume controller attach/detach
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.531034 27751 kubelet_node_status.go:443] Recording NodeHasSufficientDisk event message for node af867b
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.531048 27751 kubelet_node_status.go:443] Recording NodeHasSufficientMemory event message for node af867b
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.531058 27751 kubelet_node_status.go:443] Recording NodeHasNoDiskPressure event message for node af867b
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.531298 27751 qos_container_manager_linux.go:320] [ContainerManager]: Updated QoS cgroup configuration
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.532667 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientDisk' Node af867b status is now: NodeHasSufficientDisk
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.532694 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientMemory' Node af867b status is now: NodeHasSufficientMemory
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.532708 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasNoDiskPressure' Node af867b status is now: NodeHasNoDiskPressure
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.532750 27751 kubelet_pods.go:1284] Generating status for "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.532821 27751 kubelet_node_status.go:280] Setting node annotation to enable volume controller attach/detach
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.532847 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientMemory' Node af867b status is now: NodeHasSufficientMemory
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.532864 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientDisk' Node af867b status is now: NodeHasSufficientDisk
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.532877 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasNoDiskPressure' Node af867b status is now: NodeHasNoDiskPressure
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.534922 27751 factory.go:116] Factory "docker" was unable to handle container "/kubepods/burstable/podf49ee4da5c66af63a0b4bcea4f69baf9"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.534944 27751 factory.go:105] Error trying to work out if we can handle /kubepods/burstable/podf49ee4da5c66af63a0b4bcea4f69baf9: /kubepods/burstable/podf49ee4da5c66af63a0b4bcea4f69baf9 not handled by systemd handler
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.534951 27751 factory.go:116] Factory "systemd" was unable to handle container "/kubepods/burstable/podf49ee4da5c66af63a0b4bcea4f69baf9"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.534961 27751 factory.go:112] Using factory "raw" for container "/kubepods/burstable/podf49ee4da5c66af63a0b4bcea4f69baf9"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.535284 27751 manager.go:932] Added container: "/kubepods/burstable/podf49ee4da5c66af63a0b4bcea4f69baf9" (aliases: [], namespace: "")
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.535403 27751 handler.go:325] Added event &{/kubepods/burstable/podf49ee4da5c66af63a0b4bcea4f69baf9 2017-11-15 01:58:55.533194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.535438 27751 container.go:409] Start housekeeping for container "/kubepods/burstable/podf49ee4da5c66af63a0b4bcea4f69baf9"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.536806 27751 kubelet.go:1610] Creating a mirror pod for static pod "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)"
Nov 15 01:58:55 af867b kubelet[27751]: W1115 01:58:55.537114 27751 status_manager.go:431] Failed to get status for pod "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)": Get https://10.241.226.117:6443/api/v1/namespaces/kube-system/pods/kube-controller-manager-af867b: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:55 af867b kubelet[27751]: E1115 01:58:55.537690 27751 kubelet.go:1612] Failed creating a mirror pod for "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)": Post https://10.241.226.117:6443/api/v1/namespaces/kube-system/pods: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.540049 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.540637 27751 kubelet_node_status.go:443] Recording NodeHasSufficientDisk event message for node af867b
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.540651 27751 kubelet_node_status.go:443] Recording NodeHasSufficientMemory event message for node af867b
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.540660 27751 kubelet_node_status.go:443] Recording NodeHasNoDiskPressure event message for node af867b
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.540703 27751 kubelet_node_status.go:280] Setting node annotation to enable volume controller attach/detach
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.541049 27751 kubelet_node_status.go:443] Recording NodeHasSufficientDisk event message for node af867b
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.541063 27751 kubelet_node_status.go:443] Recording NodeHasSufficientMemory event message for node af867b
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.541071 27751 kubelet_node_status.go:443] Recording NodeHasNoDiskPressure event message for node af867b
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.541235 27751 qos_container_manager_linux.go:320] [ContainerManager]: Updated QoS cgroup configuration
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.542938 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientDisk' Node af867b status is now: NodeHasSufficientDisk
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.542966 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientMemory' Node af867b status is now: NodeHasSufficientMemory
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.542981 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasNoDiskPressure' Node af867b status is now: NodeHasNoDiskPressure
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.543023 27751 kubelet_pods.go:1284] Generating status for "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.543087 27751 kubelet_node_status.go:280] Setting node annotation to enable volume controller attach/detach
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.543110 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientMemory' Node af867b status is now: NodeHasSufficientMemory
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.543132 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientDisk' Node af867b status is now: NodeHasSufficientDisk
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.543147 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasNoDiskPressure' Node af867b status is now: NodeHasNoDiskPressure
Nov 15 01:58:55 af867b kubelet[27751]: W1115 01:58:55.545413 27751 status_manager.go:431] Failed to get status for pod "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)": Get https://10.241.226.117:6443/api/v1/namespaces/kube-system/pods/kube-scheduler-af867b: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.546776 27751 factory.go:116] Factory "docker" was unable to handle container "/kubepods/burstable/podbc22704d9f4dc5d62a8217cfd5c14373"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.546802 27751 factory.go:105] Error trying to work out if we can handle /kubepods/burstable/podbc22704d9f4dc5d62a8217cfd5c14373: /kubepods/burstable/podbc22704d9f4dc5d62a8217cfd5c14373 not handled by systemd handler
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.546809 27751 factory.go:116] Factory "systemd" was unable to handle container "/kubepods/burstable/podbc22704d9f4dc5d62a8217cfd5c14373"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.546817 27751 factory.go:112] Using factory "raw" for container "/kubepods/burstable/podbc22704d9f4dc5d62a8217cfd5c14373"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.546961 27751 manager.go:932] Added container: "/kubepods/burstable/podbc22704d9f4dc5d62a8217cfd5c14373" (aliases: [], namespace: "")
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.547065 27751 handler.go:325] Added event &{/kubepods/burstable/podbc22704d9f4dc5d62a8217cfd5c14373 2017-11-15 01:58:55.545194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.547096 27751 container.go:409] Start housekeeping for container "/kubepods/burstable/podbc22704d9f4dc5d62a8217cfd5c14373"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.549252 27751 kubelet.go:1610] Creating a mirror pod for static pod "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)"
Nov 15 01:58:55 af867b kubelet[27751]: E1115 01:58:55.549687 27751 kubelet.go:1612] Failed creating a mirror pod for "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)": Post https://10.241.226.117:6443/api/v1/namespaces/kube-system/pods: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.549813 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.550088 27751 kubelet_node_status.go:443] Recording NodeHasSufficientDisk event message for node af867b
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.550102 27751 kubelet_node_status.go:443] Recording NodeHasSufficientMemory event message for node af867b
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.550110 27751 kubelet_node_status.go:443] Recording NodeHasNoDiskPressure event message for node af867b
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.550147 27751 kubelet.go:1911] SyncLoop (housekeeping, skipped): sources aren't ready yet.
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.550203 27751 kubelet_node_status.go:443] Recording NodeHasSufficientDisk event message for node af867b
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.550214 27751 kubelet_node_status.go:443] Recording NodeHasSufficientMemory event message for node af867b
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.550224 27751 kubelet_node_status.go:443] Recording NodeHasNoDiskPressure event message for node af867b
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.550348 27751 kubelet.go:1610] Creating a mirror pod for static pod "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.550504 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientDisk' Node af867b status is now: NodeHasSufficientDisk
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.550526 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientMemory' Node af867b status is now: NodeHasSufficientMemory
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.550538 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasNoDiskPressure' Node af867b status is now: NodeHasNoDiskPressure
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.550581 27751 kubelet_pods.go:1284] Generating status for "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.550676 27751 kubelet_node_status.go:280] Setting node annotation to enable volume controller attach/detach
Nov 15 01:58:55 af867b kubelet[27751]: E1115 01:58:55.551632 27751 kubelet.go:1612] Failed creating a mirror pod for "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)": Post https://10.241.226.117:6443/api/v1/namespaces/kube-system/pods: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.551754 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.551779 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientMemory' Node af867b status is now: NodeHasSufficientMemory
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.551794 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientDisk' Node af867b status is now: NodeHasSufficientDisk
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.551807 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasNoDiskPressure' Node af867b status is now: NodeHasNoDiskPressure
Nov 15 01:58:55 af867b kubelet[27751]: W1115 01:58:55.552671 27751 status_manager.go:431] Failed to get status for pod "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)": Get https://10.241.226.117:6443/api/v1/namespaces/kube-system/pods/etcd-af867b: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.556070 27751 kubelet_node_status.go:443] Recording NodeHasSufficientDisk event message for node af867b
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.556093 27751 kubelet_node_status.go:443] Recording NodeHasSufficientMemory event message for node af867b
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.556104 27751 kubelet_node_status.go:443] Recording NodeHasNoDiskPressure event message for node af867b
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.556271 27751 qos_container_manager_linux.go:320] [ContainerManager]: Updated QoS cgroup configuration
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.557117 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientDisk' Node af867b status is now: NodeHasSufficientDisk
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.557140 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientMemory' Node af867b status is now: NodeHasSufficientMemory
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.557154 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasNoDiskPressure' Node af867b status is now: NodeHasNoDiskPressure
Nov 15 01:58:55 af867b kubelet[27751]: W1115 01:58:55.558331 27751 status_manager.go:431] Failed to get status for pod "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)": Get https://10.241.226.117:6443/api/v1/namespaces/kube-system/pods/kube-apiserver-af867b: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.560327 27751 factory.go:116] Factory "docker" was unable to handle container "/kubepods/burstable/pod4e0fac5dee63099d647b4d031a37ad7d"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.560353 27751 factory.go:105] Error trying to work out if we can handle /kubepods/burstable/pod4e0fac5dee63099d647b4d031a37ad7d: /kubepods/burstable/pod4e0fac5dee63099d647b4d031a37ad7d not handled by systemd handler
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.560359 27751 factory.go:116] Factory "systemd" was unable to handle container "/kubepods/burstable/pod4e0fac5dee63099d647b4d031a37ad7d"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.560367 27751 factory.go:112] Using factory "raw" for container "/kubepods/burstable/pod4e0fac5dee63099d647b4d031a37ad7d"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.560491 27751 manager.go:932] Added container: "/kubepods/burstable/pod4e0fac5dee63099d647b4d031a37ad7d" (aliases: [], namespace: "")
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.560588 27751 handler.go:325] Added event &{/kubepods/burstable/pod4e0fac5dee63099d647b4d031a37ad7d 2017-11-15 01:58:55.559194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.560623 27751 container.go:409] Start housekeeping for container "/kubepods/burstable/pod4e0fac5dee63099d647b4d031a37ad7d"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.563010 27751 kubelet.go:1610] Creating a mirror pod for static pod "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)"
Nov 15 01:58:55 af867b kubelet[27751]: E1115 01:58:55.563455 27751 kubelet.go:1612] Failed creating a mirror pod for "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)": Post https://10.241.226.117:6443/api/v1/namespaces/kube-system/pods: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.563649 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.624024 27751 reconciler.go:212] operationExecutor.VerifyControllerAttachedVolume started for volume "k8s-certs" (UniqueName: "kubernetes.io/host-path/f49ee4da5c66af63a0b4bcea4f69baf9-k8s-certs") pod "kube-controller-manager-af867b" (UID: "f49ee4da5c66af63a0b4bcea4f69baf9")
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.624065 27751 reconciler.go:212] operationExecutor.VerifyControllerAttachedVolume started for volume "ca-certs" (UniqueName: "kubernetes.io/host-path/f49ee4da5c66af63a0b4bcea4f69baf9-ca-certs") pod "kube-controller-manager-af867b" (UID: "f49ee4da5c66af63a0b4bcea4f69baf9")
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.624122 27751 reconciler.go:212] operationExecutor.VerifyControllerAttachedVolume started for volume "kubeconfig" (UniqueName: "kubernetes.io/host-path/f49ee4da5c66af63a0b4bcea4f69baf9-kubeconfig") pod "kube-controller-manager-af867b" (UID: "f49ee4da5c66af63a0b4bcea4f69baf9")
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.624146 27751 reconciler.go:212] operationExecutor.VerifyControllerAttachedVolume started for volume "flexvolume-dir" (UniqueName: "kubernetes.io/host-path/f49ee4da5c66af63a0b4bcea4f69baf9-flexvolume-dir") pod "kube-controller-manager-af867b" (UID: "f49ee4da5c66af63a0b4bcea4f69baf9")
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.624172 27751 reconciler.go:212] operationExecutor.VerifyControllerAttachedVolume started for volume "ca-certs-etc-pki" (UniqueName: "kubernetes.io/host-path/f49ee4da5c66af63a0b4bcea4f69baf9-ca-certs-etc-pki") pod "kube-controller-manager-af867b" (UID: "f49ee4da5c66af63a0b4bcea4f69baf9")
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.724088 27751 reconciler.go:257] operationExecutor.MountVolume started for volume "flexvolume-dir" (UniqueName: "kubernetes.io/host-path/f49ee4da5c66af63a0b4bcea4f69baf9-flexvolume-dir") pod "kube-controller-manager-af867b" (UID: "f49ee4da5c66af63a0b4bcea4f69baf9")
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.724154 27751 reconciler.go:212] operationExecutor.VerifyControllerAttachedVolume started for volume "etcd" (UniqueName: "kubernetes.io/host-path/d76e26fba3bf2bfd215eb29011d55250-etcd") pod "etcd-af867b" (UID: "d76e26fba3bf2bfd215eb29011d55250")
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.724185 27751 reconciler.go:212] operationExecutor.VerifyControllerAttachedVolume started for volume "ca-certs" (UniqueName: "kubernetes.io/host-path/4e0fac5dee63099d647b4d031a37ad7d-ca-certs") pod "kube-apiserver-af867b" (UID: "4e0fac5dee63099d647b4d031a37ad7d")
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.724210 27751 reconciler.go:212] operationExecutor.VerifyControllerAttachedVolume started for volume "ca-certs-etc-pki" (UniqueName: "kubernetes.io/host-path/4e0fac5dee63099d647b4d031a37ad7d-ca-certs-etc-pki") pod "kube-apiserver-af867b" (UID: "4e0fac5dee63099d647b4d031a37ad7d")
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.724236 27751 reconciler.go:212] operationExecutor.VerifyControllerAttachedVolume started for volume "kubeconfig" (UniqueName: "kubernetes.io/host-path/bc22704d9f4dc5d62a8217cfd5c14373-kubeconfig") pod "kube-scheduler-af867b" (UID: "bc22704d9f4dc5d62a8217cfd5c14373")
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.724270 27751 reconciler.go:257] operationExecutor.MountVolume started for volume "k8s-certs" (UniqueName: "kubernetes.io/host-path/f49ee4da5c66af63a0b4bcea4f69baf9-k8s-certs") pod "kube-controller-manager-af867b" (UID: "f49ee4da5c66af63a0b4bcea4f69baf9")
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.724297 27751 reconciler.go:257] operationExecutor.MountVolume started for volume "ca-certs" (UniqueName: "kubernetes.io/host-path/f49ee4da5c66af63a0b4bcea4f69baf9-ca-certs") pod "kube-controller-manager-af867b" (UID: "f49ee4da5c66af63a0b4bcea4f69baf9")
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.724325 27751 reconciler.go:257] operationExecutor.MountVolume started for volume "kubeconfig" (UniqueName: "kubernetes.io/host-path/f49ee4da5c66af63a0b4bcea4f69baf9-kubeconfig") pod "kube-controller-manager-af867b" (UID: "f49ee4da5c66af63a0b4bcea4f69baf9")
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.724363 27751 reconciler.go:257] operationExecutor.MountVolume started for volume "ca-certs-etc-pki" (UniqueName: "kubernetes.io/host-path/f49ee4da5c66af63a0b4bcea4f69baf9-ca-certs-etc-pki") pod "kube-controller-manager-af867b" (UID: "f49ee4da5c66af63a0b4bcea4f69baf9")
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.724394 27751 reconciler.go:212] operationExecutor.VerifyControllerAttachedVolume started for volume "k8s-certs" (UniqueName: "kubernetes.io/host-path/4e0fac5dee63099d647b4d031a37ad7d-k8s-certs") pod "kube-apiserver-af867b" (UID: "4e0fac5dee63099d647b4d031a37ad7d")
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.724538 27751 operation_generator.go:484] MountVolume.SetUp succeeded for volume "flexvolume-dir" (UniqueName: "kubernetes.io/host-path/f49ee4da5c66af63a0b4bcea4f69baf9-flexvolume-dir") pod "kube-controller-manager-af867b" (UID: "f49ee4da5c66af63a0b4bcea4f69baf9")
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.724573 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-controller-manager-af867b", UID:"f49ee4da5c66af63a0b4bcea4f69baf9", APIVersion:"v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "flexvolume-dir"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.724677 27751 operation_generator.go:484] MountVolume.SetUp succeeded for volume "k8s-certs" (UniqueName: "kubernetes.io/host-path/f49ee4da5c66af63a0b4bcea4f69baf9-k8s-certs") pod "kube-controller-manager-af867b" (UID: "f49ee4da5c66af63a0b4bcea4f69baf9")
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.724695 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-controller-manager-af867b", UID:"f49ee4da5c66af63a0b4bcea4f69baf9", APIVersion:"v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "k8s-certs"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.724769 27751 operation_generator.go:484] MountVolume.SetUp succeeded for volume "ca-certs" (UniqueName: "kubernetes.io/host-path/f49ee4da5c66af63a0b4bcea4f69baf9-ca-certs") pod "kube-controller-manager-af867b" (UID: "f49ee4da5c66af63a0b4bcea4f69baf9")
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.724789 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-controller-manager-af867b", UID:"f49ee4da5c66af63a0b4bcea4f69baf9", APIVersion:"v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "ca-certs"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.724834 27751 operation_generator.go:484] MountVolume.SetUp succeeded for volume "kubeconfig" (UniqueName: "kubernetes.io/host-path/f49ee4da5c66af63a0b4bcea4f69baf9-kubeconfig") pod "kube-controller-manager-af867b" (UID: "f49ee4da5c66af63a0b4bcea4f69baf9")
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.724853 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-controller-manager-af867b", UID:"f49ee4da5c66af63a0b4bcea4f69baf9", APIVersion:"v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "kubeconfig"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.724891 27751 operation_generator.go:484] MountVolume.SetUp succeeded for volume "ca-certs-etc-pki" (UniqueName: "kubernetes.io/host-path/f49ee4da5c66af63a0b4bcea4f69baf9-ca-certs-etc-pki") pod "kube-controller-manager-af867b" (UID: "f49ee4da5c66af63a0b4bcea4f69baf9")
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.724906 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-controller-manager-af867b", UID:"f49ee4da5c66af63a0b4bcea4f69baf9", APIVersion:"v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "ca-certs-etc-pki"
Nov 15 01:58:55 af867b kubelet[27751]: W1115 01:58:55.793547 27751 cni.go:196] Unable to update cni config: No networks found in /etc/cni/net.d
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.793700 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 01:58:55 af867b kubelet[27751]: E1115 01:58:55.793746 27751 kubelet.go:2095] Container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.824008 27751 reconciler.go:257] operationExecutor.MountVolume started for volume "k8s-certs" (UniqueName: "kubernetes.io/host-path/4e0fac5dee63099d647b4d031a37ad7d-k8s-certs") pod "kube-apiserver-af867b" (UID: "4e0fac5dee63099d647b4d031a37ad7d")
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.824048 27751 reconciler.go:257] operationExecutor.MountVolume started for volume "etcd" (UniqueName: "kubernetes.io/host-path/d76e26fba3bf2bfd215eb29011d55250-etcd") pod "etcd-af867b" (UID: "d76e26fba3bf2bfd215eb29011d55250")
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.824101 27751 reconciler.go:257] operationExecutor.MountVolume started for volume "ca-certs" (UniqueName: "kubernetes.io/host-path/4e0fac5dee63099d647b4d031a37ad7d-ca-certs") pod "kube-apiserver-af867b" (UID: "4e0fac5dee63099d647b4d031a37ad7d")
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.824133 27751 reconciler.go:257] operationExecutor.MountVolume started for volume "ca-certs-etc-pki" (UniqueName: "kubernetes.io/host-path/4e0fac5dee63099d647b4d031a37ad7d-ca-certs-etc-pki") pod "kube-apiserver-af867b" (UID: "4e0fac5dee63099d647b4d031a37ad7d")
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.824162 27751 reconciler.go:257] operationExecutor.MountVolume started for volume "kubeconfig" (UniqueName: "kubernetes.io/host-path/bc22704d9f4dc5d62a8217cfd5c14373-kubeconfig") pod "kube-scheduler-af867b" (UID: "bc22704d9f4dc5d62a8217cfd5c14373")
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.824238 27751 operation_generator.go:484] MountVolume.SetUp succeeded for volume "kubeconfig" (UniqueName: "kubernetes.io/host-path/bc22704d9f4dc5d62a8217cfd5c14373-kubeconfig") pod "kube-scheduler-af867b" (UID: "bc22704d9f4dc5d62a8217cfd5c14373")
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.824269 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-scheduler-af867b", UID:"bc22704d9f4dc5d62a8217cfd5c14373", APIVersion:"v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "kubeconfig"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.824315 27751 operation_generator.go:484] MountVolume.SetUp succeeded for volume "k8s-certs" (UniqueName: "kubernetes.io/host-path/4e0fac5dee63099d647b4d031a37ad7d-k8s-certs") pod "kube-apiserver-af867b" (UID: "4e0fac5dee63099d647b4d031a37ad7d")
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.824330 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-apiserver-af867b", UID:"4e0fac5dee63099d647b4d031a37ad7d", APIVersion:"v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "k8s-certs"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.824377 27751 operation_generator.go:484] MountVolume.SetUp succeeded for volume "etcd" (UniqueName: "kubernetes.io/host-path/d76e26fba3bf2bfd215eb29011d55250-etcd") pod "etcd-af867b" (UID: "d76e26fba3bf2bfd215eb29011d55250")
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.824394 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"etcd-af867b", UID:"d76e26fba3bf2bfd215eb29011d55250", APIVersion:"v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "etcd"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.824438 27751 operation_generator.go:484] MountVolume.SetUp succeeded for volume "ca-certs" (UniqueName: "kubernetes.io/host-path/4e0fac5dee63099d647b4d031a37ad7d-ca-certs") pod "kube-apiserver-af867b" (UID: "4e0fac5dee63099d647b4d031a37ad7d")
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.824453 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-apiserver-af867b", UID:"4e0fac5dee63099d647b4d031a37ad7d", APIVersion:"v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "ca-certs"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.824489 27751 operation_generator.go:484] MountVolume.SetUp succeeded for volume "ca-certs-etc-pki" (UniqueName: "kubernetes.io/host-path/4e0fac5dee63099d647b4d031a37ad7d-ca-certs-etc-pki") pod "kube-apiserver-af867b" (UID: "4e0fac5dee63099d647b4d031a37ad7d")
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.824502 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-apiserver-af867b", UID:"4e0fac5dee63099d647b4d031a37ad7d", APIVersion:"v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "ca-certs-etc-pki"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.840263 27751 volume_manager.go:366] All volumes are attached and mounted for pod "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.840291 27751 kuberuntime_manager.go:370] No sandbox for pod "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)" can be found. Need to start a new one
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.840305 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:true CreateSandbox:true SandboxID: Attempt:0 NextInitContainerToStart:nil ContainersToStart:[0] ContainersToKill:map[]} for pod "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.840362 27751 kuberuntime_manager.go:565] SyncPod received new pod "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)", will create a sandbox for it
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.840371 27751 kuberuntime_manager.go:574] Stopping PodSandbox for "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)", will start new one
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.840402 27751 kuberuntime_manager.go:626] Creating sandbox for pod "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.842967 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/burstable/podf49ee4da5c66af63a0b4bcea4f69baf9"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.842990 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/burstable/podf49ee4da5c66af63a0b4bcea4f69baf9"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.851861 27751 volume_manager.go:366] All volumes are attached and mounted for pod "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.851882 27751 kuberuntime_manager.go:370] No sandbox for pod "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)" can be found. Need to start a new one
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.851893 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:true CreateSandbox:true SandboxID: Attempt:0 NextInitContainerToStart:nil ContainersToStart:[0] ContainersToKill:map[]} for pod "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.851927 27751 kuberuntime_manager.go:565] SyncPod received new pod "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)", will create a sandbox for it
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.851936 27751 kuberuntime_manager.go:574] Stopping PodSandbox for "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)", will start new one
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.851952 27751 kuberuntime_manager.go:626] Creating sandbox for pod "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.852148 27751 volume_manager.go:366] All volumes are attached and mounted for pod "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.852162 27751 kuberuntime_manager.go:370] No sandbox for pod "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)" can be found. Need to start a new one
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.852170 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:true CreateSandbox:true SandboxID: Attempt:0 NextInitContainerToStart:nil ContainersToStart:[0] ContainersToKill:map[]} for pod "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.852205 27751 kuberuntime_manager.go:565] SyncPod received new pod "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)", will create a sandbox for it
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.852213 27751 kuberuntime_manager.go:574] Stopping PodSandbox for "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)", will start new one
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.852227 27751 kuberuntime_manager.go:626] Creating sandbox for pod "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.863878 27751 volume_manager.go:366] All volumes are attached and mounted for pod "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.863898 27751 kuberuntime_manager.go:370] No sandbox for pod "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)" can be found. Need to start a new one
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.863907 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:true CreateSandbox:true SandboxID: Attempt:0 NextInitContainerToStart:nil ContainersToStart:[0] ContainersToKill:map[]} for pod "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)"
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.863938 27751 kuberuntime_manager.go:565] SyncPod received new pod "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)", will create a sandbox for it
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.863947 27751 kuberuntime_manager.go:574] Stopping PodSandbox for "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)", will start new one
Nov 15 01:58:55 af867b kubelet[27751]: I1115 01:58:55.863961 27751 kuberuntime_manager.go:626] Creating sandbox for pod "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)"
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.019261 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/burstable/pod4e0fac5dee63099d647b4d031a37ad7d"
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.019307 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/burstable/pod4e0fac5dee63099d647b4d031a37ad7d"
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.019680 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/besteffort/podd76e26fba3bf2bfd215eb29011d55250"
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.019692 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/besteffort/podd76e26fba3bf2bfd215eb29011d55250"
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.019966 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/burstable/podbc22704d9f4dc5d62a8217cfd5c14373"
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.019977 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/burstable/podbc22704d9f4dc5d62a8217cfd5c14373"
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.381837 27751 reflector.go:240] Listing and watching *v1.Node from k8s.io/kubernetes/pkg/kubelet/kubelet.go:422
Nov 15 01:58:56 af867b kubelet[27751]: E1115 01:58:56.382646 27751 reflector.go:205] k8s.io/kubernetes/pkg/kubelet/kubelet.go:422: Failed to list *v1.Node: Get https://10.241.226.117:6443/api/v1/nodes?fieldSelector=metadata.name%3Daf867b&resourceVersion=0: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.383318 27751 reflector.go:240] Listing and watching *v1.Pod from k8s.io/kubernetes/pkg/kubelet/config/apiserver.go:47
Nov 15 01:58:56 af867b kubelet[27751]: E1115 01:58:56.383848 27751 reflector.go:205] k8s.io/kubernetes/pkg/kubelet/config/apiserver.go:47: Failed to list *v1.Pod: Get https://10.241.226.117:6443/api/v1/pods?fieldSelector=spec.nodeName%3Daf867b&resourceVersion=0: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.384608 27751 reflector.go:240] Listing and watching *v1.Service from k8s.io/kubernetes/pkg/kubelet/kubelet.go:413
Nov 15 01:58:56 af867b kubelet[27751]: E1115 01:58:56.385126 27751 reflector.go:205] k8s.io/kubernetes/pkg/kubelet/kubelet.go:413: Failed to list *v1.Service: Get https://10.241.226.117:6443/api/v1/services?resourceVersion=0: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.530696 27751 kubelet.go:1911] SyncLoop (housekeeping, skipped): sources aren't ready yet.
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.742544 27751 docker_sandbox.go:691] Will attempt to re-write config file /var/lib/docker/containers/439f518d46ec0d596518b7c17c503c52a85dd4da21d8a56ec174d5b6b3b98e03/resolv.conf with:
Nov 15 01:58:56 af867b kubelet[27751]: [nameserver 10.196.65.209 search opcwlaas.oraclecloud.internal. opcwlaas.oraclecloud.internal.]
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.743421 27751 factory.go:112] Using factory "docker" for container "/kubepods/burstable/podf49ee4da5c66af63a0b4bcea4f69baf9/439f518d46ec0d596518b7c17c503c52a85dd4da21d8a56ec174d5b6b3b98e03"
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.743703 27751 kuberuntime_manager.go:640] Created PodSandbox "439f518d46ec0d596518b7c17c503c52a85dd4da21d8a56ec174d5b6b3b98e03" for pod "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)"
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.791640 27751 manager.go:932] Added container: "/kubepods/burstable/podf49ee4da5c66af63a0b4bcea4f69baf9/439f518d46ec0d596518b7c17c503c52a85dd4da21d8a56ec174d5b6b3b98e03" (aliases: [k8s_POD_kube-controller-manager-af867b_kube-system_f49ee4da5c66af63a0b4bcea4f69baf9_0 439f518d46ec0d596518b7c17c503c52a85dd4da21d8a56ec174d5b6b3b98e03], namespace: "docker")
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.791801 27751 handler.go:325] Added event &{/kubepods/burstable/podf49ee4da5c66af63a0b4bcea4f69baf9/439f518d46ec0d596518b7c17c503c52a85dd4da21d8a56ec174d5b6b3b98e03 2017-11-15 01:58:56.323194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.791930 27751 container.go:409] Start housekeeping for container "/kubepods/burstable/podf49ee4da5c66af63a0b4bcea4f69baf9/439f518d46ec0d596518b7c17c503c52a85dd4da21d8a56ec174d5b6b3b98e03"
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.794801 27751 docker_sandbox.go:691] Will attempt to re-write config file /var/lib/docker/containers/fe18d337ce42bd3a4d2aacb1349be028274681d8073fbe5d6377946e815fa810/resolv.conf with:
Nov 15 01:58:56 af867b kubelet[27751]: [nameserver 10.196.65.209 search opcwlaas.oraclecloud.internal. opcwlaas.oraclecloud.internal.]
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.795300 27751 kuberuntime_manager.go:640] Created PodSandbox "fe18d337ce42bd3a4d2aacb1349be028274681d8073fbe5d6377946e815fa810" for pod "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)"
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.795444 27751 generic.go:146] GenericPLEG: bc22704d9f4dc5d62a8217cfd5c14373/fe18d337ce42bd3a4d2aacb1349be028274681d8073fbe5d6377946e815fa810: non-existent -> running
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.795475 27751 generic.go:146] GenericPLEG: f49ee4da5c66af63a0b4bcea4f69baf9/439f518d46ec0d596518b7c17c503c52a85dd4da21d8a56ec174d5b6b3b98e03: non-existent -> running
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.795830 27751 kuberuntime_manager.go:705] Creating container &Container{Name:kube-controller-manager,Image:gcr.io/google_containers/kube-controller-manager-amd64:v1.8.3,Command:[kube-controller-manager --leader-elect=true --use-service-account-credentials=true --controllers=*,bootstrapsigner,tokencleaner --root-ca-file=/etc/kubernetes/pki/ca.crt --cluster-signing-cert-file=/etc/kubernetes/pki/ca.crt --address=127.0.0.1 --kubeconfig=/etc/kubernetes/controller-manager.conf --service-account-private-key-file=/etc/kubernetes/pki/sa.key --cluster-signing-key-file=/etc/kubernetes/pki/ca.key],Args:[],WorkingDir:,Ports:[],Env:[{http_proxy http://www-proxy.us.oracle.com:80 nil} {https_proxy https://www-proxy.us.oracle.com:80 nil} {no_proxy 10.241.226.117 nil}],Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{200 -3} {<nil>} 200m DecimalSI},},},VolumeMounts:[{k8s-certs true /etc/kubernetes/pki <nil>} {ca-certs true /etc/ssl/certs <nil>} {kubeconfig true /etc/kubernetes/controller-manager.conf <nil>} {flexvolume-dir false /usr/libexec/kubernetes/kubelet-plugins/volume/exec <nil>} {ca-certs-etc-pki true /etc/pki <nil>}],LivenessProbe:&Probe{Handler:Handler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:10252,Host:127.0.0.1,Scheme:HTTP,HTTPHeaders:[],},TCPSocket:nil,},InitialDelaySeconds:15,TimeoutSeconds:15,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:8,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[],TerminationMessagePolicy:File,} in pod kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.835144 27751 factory.go:112] Using factory "docker" for container "/kubepods/burstable/podbc22704d9f4dc5d62a8217cfd5c14373/fe18d337ce42bd3a4d2aacb1349be028274681d8073fbe5d6377946e815fa810"
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.836202 27751 kuberuntime_container.go:100] Generating ref for container kube-controller-manager: &v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-controller-manager-af867b", UID:"f49ee4da5c66af63a0b4bcea4f69baf9", APIVersion:"v1", ResourceVersion:"", FieldPath:"spec.containers{kube-controller-manager}"}
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.836240 27751 container_manager_linux.go:634] Calling devicePluginHandler AllocateDevices
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.836279 27751 kubelet_pods.go:123] container: kube-system/kube-controller-manager-af867b/kube-controller-manager podIP: "10.196.65.210" creating hosts mount: true
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.836531 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-controller-manager-af867b", UID:"f49ee4da5c66af63a0b4bcea4f69baf9", APIVersion:"v1", ResourceVersion:"", FieldPath:"spec.containers{kube-controller-manager}"}): type: 'Normal' reason: 'Pulled' Container image "gcr.io/google_containers/kube-controller-manager-amd64:v1.8.3" already present on machine
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.837154 27751 kuberuntime_manager.go:705] Creating container &Container{Name:kube-scheduler,Image:gcr.io/google_containers/kube-scheduler-amd64:v1.8.3,Command:[kube-scheduler --kubeconfig=/etc/kubernetes/scheduler.conf --address=127.0.0.1 --leader-elect=true],Args:[],WorkingDir:,Ports:[],Env:[{http_proxy http://www-proxy.us.oracle.com:80 nil} {https_proxy https://www-proxy.us.oracle.com:80 nil} {no_proxy 10.241.226.117 nil}],Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{100 -3} {<nil>} 100m DecimalSI},},},VolumeMounts:[{kubeconfig true /etc/kubernetes/scheduler.conf <nil>}],LivenessProbe:&Probe{Handler:Handler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:10251,Host:127.0.0.1,Scheme:HTTP,HTTPHeaders:[],},TCPSocket:nil,},InitialDelaySeconds:15,TimeoutSeconds:15,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:8,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[],TerminationMessagePolicy:File,} in pod kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.857871 27751 kubelet_node_status.go:280] Setting node annotation to enable volume controller attach/detach
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.860814 27751 kubelet_node_status.go:443] Recording NodeHasSufficientDisk event message for node af867b
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.860839 27751 kubelet_node_status.go:443] Recording NodeHasSufficientMemory event message for node af867b
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.860848 27751 kubelet_node_status.go:443] Recording NodeHasNoDiskPressure event message for node af867b
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.860870 27751 kubelet_node_status.go:83] Attempting to register node af867b
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.861111 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientDisk' Node af867b status is now: NodeHasSufficientDisk
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.861133 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientMemory' Node af867b status is now: NodeHasSufficientMemory
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.861145 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasNoDiskPressure' Node af867b status is now: NodeHasNoDiskPressure
Nov 15 01:58:56 af867b kubelet[27751]: E1115 01:58:56.862725 27751 kubelet_node_status.go:107] Unable to register node "af867b" with API server: Post https://10.241.226.117:6443/api/v1/nodes: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.907098 27751 manager.go:932] Added container: "/kubepods/burstable/podbc22704d9f4dc5d62a8217cfd5c14373/fe18d337ce42bd3a4d2aacb1349be028274681d8073fbe5d6377946e815fa810" (aliases: [k8s_POD_kube-scheduler-af867b_kube-system_bc22704d9f4dc5d62a8217cfd5c14373_0 fe18d337ce42bd3a4d2aacb1349be028274681d8073fbe5d6377946e815fa810], namespace: "docker")
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.907268 27751 handler.go:325] Added event &{/kubepods/burstable/podbc22704d9f4dc5d62a8217cfd5c14373/fe18d337ce42bd3a4d2aacb1349be028274681d8073fbe5d6377946e815fa810 2017-11-15 01:58:56.650194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.907484 27751 container.go:409] Start housekeeping for container "/kubepods/burstable/podbc22704d9f4dc5d62a8217cfd5c14373/fe18d337ce42bd3a4d2aacb1349be028274681d8073fbe5d6377946e815fa810"
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.908891 27751 kuberuntime_container.go:100] Generating ref for container kube-scheduler: &v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-scheduler-af867b", UID:"bc22704d9f4dc5d62a8217cfd5c14373", APIVersion:"v1", ResourceVersion:"", FieldPath:"spec.containers{kube-scheduler}"}
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.908937 27751 container_manager_linux.go:634] Calling devicePluginHandler AllocateDevices
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.908980 27751 kubelet_pods.go:123] container: kube-system/kube-scheduler-af867b/kube-scheduler podIP: "10.196.65.210" creating hosts mount: true
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.909247 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-scheduler-af867b", UID:"bc22704d9f4dc5d62a8217cfd5c14373", APIVersion:"v1", ResourceVersion:"", FieldPath:"spec.containers{kube-scheduler}"}): type: 'Normal' reason: 'Pulled' Container image "gcr.io/google_containers/kube-scheduler-amd64:v1.8.3" already present on machine
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.910257 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/burstable/podf49ee4da5c66af63a0b4bcea4f69baf9"
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.923653 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/burstable/podbc22704d9f4dc5d62a8217cfd5c14373"
Nov 15 01:58:56 af867b kubelet[27751]: I1115 01:58:56.930830 27751 kuberuntime_manager.go:833] getSandboxIDByPodUID got sandbox IDs ["fe18d337ce42bd3a4d2aacb1349be028274681d8073fbe5d6377946e815fa810"] for pod "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)"
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.170072 27751 factory.go:112] Using factory "docker" for container "/kubepods/burstable/pod4e0fac5dee63099d647b4d031a37ad7d/19b06ca14052c17f92bf03d5316fd7642fe4fee6bf193755575355be3e87a13d"
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.171018 27751 docker_sandbox.go:691] Will attempt to re-write config file /var/lib/docker/containers/19b06ca14052c17f92bf03d5316fd7642fe4fee6bf193755575355be3e87a13d/resolv.conf with:
Nov 15 01:58:57 af867b kubelet[27751]: [nameserver 10.196.65.209 search opcwlaas.oraclecloud.internal. opcwlaas.oraclecloud.internal.]
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.171442 27751 kuberuntime_manager.go:640] Created PodSandbox "19b06ca14052c17f92bf03d5316fd7642fe4fee6bf193755575355be3e87a13d" for pod "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)"
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.171561 27751 generic.go:345] PLEG: Write status for kube-scheduler-af867b/kube-system: &container.PodStatus{ID:"bc22704d9f4dc5d62a8217cfd5c14373", Name:"kube-scheduler-af867b", Namespace:"kube-system", IP:"", ContainerStatuses:[]*container.ContainerStatus{}, SandboxStatuses:[]*runtime.PodSandboxStatus{(*runtime.PodSandboxStatus)(0xc42123c280)}} (err: <nil>)
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.171652 27751 kubelet.go:1871] SyncLoop (PLEG): "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)", event: &pleg.PodLifecycleEvent{ID:"bc22704d9f4dc5d62a8217cfd5c14373", Type:"ContainerStarted", Data:"fe18d337ce42bd3a4d2aacb1349be028274681d8073fbe5d6377946e815fa810"}
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.173323 27751 manager.go:932] Added container: "/kubepods/burstable/pod4e0fac5dee63099d647b4d031a37ad7d/19b06ca14052c17f92bf03d5316fd7642fe4fee6bf193755575355be3e87a13d" (aliases: [k8s_POD_kube-apiserver-af867b_kube-system_4e0fac5dee63099d647b4d031a37ad7d_0 19b06ca14052c17f92bf03d5316fd7642fe4fee6bf193755575355be3e87a13d], namespace: "docker")
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.173461 27751 handler.go:325] Added event &{/kubepods/burstable/pod4e0fac5dee63099d647b4d031a37ad7d/19b06ca14052c17f92bf03d5316fd7642fe4fee6bf193755575355be3e87a13d 2017-11-15 01:58:56.869194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.173508 27751 container.go:409] Start housekeeping for container "/kubepods/burstable/pod4e0fac5dee63099d647b4d031a37ad7d/19b06ca14052c17f92bf03d5316fd7642fe4fee6bf193755575355be3e87a13d"
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.176294 27751 kuberuntime_manager.go:833] getSandboxIDByPodUID got sandbox IDs ["439f518d46ec0d596518b7c17c503c52a85dd4da21d8a56ec174d5b6b3b98e03"] for pod "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)"
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.176465 27751 kuberuntime_manager.go:705] Creating container &Container{Name:kube-apiserver,Image:gcr.io/google_containers/kube-apiserver-amd64:v1.8.3,Command:[kube-apiserver --proxy-client-cert-file=/etc/kubernetes/pki/front-proxy-client.crt --enable-bootstrap-token-auth=true --tls-cert-file=/etc/kubernetes/pki/apiserver.crt --secure-port=6443 --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --proxy-client-key-file=/etc/kubernetes/pki/front-proxy-client.key --allow-privileged=true --kubelet-preferred-address-types=InternalIP,ExternalIP,Hostname --kubelet-client-certificate=/etc/kubernetes/pki/apiserver-kubelet-client.crt --tls-private-key-file=/etc/kubernetes/pki/apiserver.key --requestheader-username-headers=X-Remote-User --requestheader-extra-headers-prefix=X-Remote-Extra- --requestheader-allowed-names=front-proxy-client --advertise-address=10.241.226.117 --service-cluster-ip-range=10.96.0.0/12 --service-account-key-file=/etc/kubernetes/pki/sa.pub --client-ca-file=/etc/kubernetes/pki/ca.crt --kubelet-client-key=/etc/kubernetes/pki/apiserver-kubelet-client.key --insecure-port=0 --admission-control=Initializers,NamespaceLifecycle,LimitRanger,ServiceAccount,PersistentVolumeLabel,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,ResourceQuota --requestheader-group-headers=X-Remote-Group --authorization-mode=Node,RBAC --etcd-servers=http://127.0.0.1:2379],Args:[],WorkingDir:,Ports:[],Env:[{http_proxy http://www-proxy.us.oracle.com:80 nil} {https_proxy https://www-proxy.us.oracle.com:80 nil} {no_proxy 10.241.226.117 nil}],Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{250 -3} {<nil>} 250m DecimalSI},},},VolumeMounts:[{k8s-certs true /etc/kubernetes/pki <nil>} {ca-certs true /etc/ssl/certs <nil>} {ca-certs-etc-pki true /etc/pki <nil>}],LivenessProbe:&Probe{Handler:Handler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:6443,Host:127.0.0.1,Scheme:HTTPS,HTTPHeaders:[],},TCPSocket:nil,},InitialDelaySeconds:15,TimeoutSeconds:15,Perio
Nov 15 01:58:57 af867b kubelet[27751]: dSeconds:10,SuccessThreshold:1,FailureThreshold:8,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[],TerminationMessagePolicy:File,} in pod kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.178352 27751 kuberuntime_container.go:100] Generating ref for container kube-apiserver: &v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-apiserver-af867b", UID:"4e0fac5dee63099d647b4d031a37ad7d", APIVersion:"v1", ResourceVersion:"", FieldPath:"spec.containers{kube-apiserver}"}
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.178391 27751 container_manager_linux.go:634] Calling devicePluginHandler AllocateDevices
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.178422 27751 kubelet_pods.go:123] container: kube-system/kube-apiserver-af867b/kube-apiserver podIP: "10.196.65.210" creating hosts mount: true
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.178643 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-apiserver-af867b", UID:"4e0fac5dee63099d647b4d031a37ad7d", APIVersion:"v1", ResourceVersion:"", FieldPath:"spec.containers{kube-apiserver}"}): type: 'Normal' reason: 'Pulled' Container image "gcr.io/google_containers/kube-apiserver-amd64:v1.8.3" already present on machine
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.179854 27751 generic.go:345] PLEG: Write status for kube-controller-manager-af867b/kube-system: &container.PodStatus{ID:"f49ee4da5c66af63a0b4bcea4f69baf9", Name:"kube-controller-manager-af867b", Namespace:"kube-system", IP:"", ContainerStatuses:[]*container.ContainerStatus{}, SandboxStatuses:[]*runtime.PodSandboxStatus{(*runtime.PodSandboxStatus)(0xc4211e5900)}} (err: <nil>)
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.179904 27751 kubelet.go:1871] SyncLoop (PLEG): "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)", event: &pleg.PodLifecycleEvent{ID:"f49ee4da5c66af63a0b4bcea4f69baf9", Type:"ContainerStarted", Data:"439f518d46ec0d596518b7c17c503c52a85dd4da21d8a56ec174d5b6b3b98e03"}
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.180382 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/burstable/pod4e0fac5dee63099d647b4d031a37ad7d"
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.382888 27751 reflector.go:240] Listing and watching *v1.Node from k8s.io/kubernetes/pkg/kubelet/kubelet.go:422
Nov 15 01:58:57 af867b kubelet[27751]: E1115 01:58:57.384130 27751 reflector.go:205] k8s.io/kubernetes/pkg/kubelet/kubelet.go:422: Failed to list *v1.Node: Get https://10.241.226.117:6443/api/v1/nodes?fieldSelector=metadata.name%3Daf867b&resourceVersion=0: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.384180 27751 reflector.go:240] Listing and watching *v1.Pod from k8s.io/kubernetes/pkg/kubelet/config/apiserver.go:47
Nov 15 01:58:57 af867b kubelet[27751]: E1115 01:58:57.384811 27751 reflector.go:205] k8s.io/kubernetes/pkg/kubelet/config/apiserver.go:47: Failed to list *v1.Pod: Get https://10.241.226.117:6443/api/v1/pods?fieldSelector=spec.nodeName%3Daf867b&resourceVersion=0: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.385288 27751 reflector.go:240] Listing and watching *v1.Service from k8s.io/kubernetes/pkg/kubelet/kubelet.go:413
Nov 15 01:58:57 af867b kubelet[27751]: E1115 01:58:57.385978 27751 reflector.go:205] k8s.io/kubernetes/pkg/kubelet/kubelet.go:413: Failed to list *v1.Service: Get https://10.241.226.117:6443/api/v1/services?resourceVersion=0: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.474546 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-controller-manager-af867b", UID:"f49ee4da5c66af63a0b4bcea4f69baf9", APIVersion:"v1", ResourceVersion:"", FieldPath:"spec.containers{kube-controller-manager}"}): type: 'Normal' reason: 'Created' Created container
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.505945 27751 worker.go:164] Probe target container not found: kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9) - kube-controller-manager
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.630310 27751 docker_sandbox.go:691] Will attempt to re-write config file /var/lib/docker/containers/d5076ad0e9fb270d1b8c4ff7cdadbf32d6b1e30dc42ae24dfbc5cb01bb5aa934/resolv.conf with:
Nov 15 01:58:57 af867b kubelet[27751]: [nameserver 10.196.65.209 search opcwlaas.oraclecloud.internal. opcwlaas.oraclecloud.internal.]
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.631917 27751 factory.go:112] Using factory "docker" for container "/kubepods/besteffort/podd76e26fba3bf2bfd215eb29011d55250/d5076ad0e9fb270d1b8c4ff7cdadbf32d6b1e30dc42ae24dfbc5cb01bb5aa934"
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.632232 27751 kuberuntime_manager.go:640] Created PodSandbox "d5076ad0e9fb270d1b8c4ff7cdadbf32d6b1e30dc42ae24dfbc5cb01bb5aa934" for pod "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)"
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.681507 27751 worker.go:164] Probe target container not found: kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373) - kube-scheduler
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.746631 27751 manager.go:932] Added container: "/kubepods/besteffort/podd76e26fba3bf2bfd215eb29011d55250/d5076ad0e9fb270d1b8c4ff7cdadbf32d6b1e30dc42ae24dfbc5cb01bb5aa934" (aliases: [k8s_POD_etcd-af867b_kube-system_d76e26fba3bf2bfd215eb29011d55250_0 d5076ad0e9fb270d1b8c4ff7cdadbf32d6b1e30dc42ae24dfbc5cb01bb5aa934], namespace: "docker")
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.746792 27751 handler.go:325] Added event &{/kubepods/besteffort/podd76e26fba3bf2bfd215eb29011d55250/d5076ad0e9fb270d1b8c4ff7cdadbf32d6b1e30dc42ae24dfbc5cb01bb5aa934 2017-11-15 01:58:57.291194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.746916 27751 container.go:409] Start housekeeping for container "/kubepods/besteffort/podd76e26fba3bf2bfd215eb29011d55250/d5076ad0e9fb270d1b8c4ff7cdadbf32d6b1e30dc42ae24dfbc5cb01bb5aa934"
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.752138 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-scheduler-af867b", UID:"bc22704d9f4dc5d62a8217cfd5c14373", APIVersion:"v1", ResourceVersion:"", FieldPath:"spec.containers{kube-scheduler}"}): type: 'Normal' reason: 'Created' Created container
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.752304 27751 kuberuntime_manager.go:705] Creating container &Container{Name:etcd,Image:gcr.io/google_containers/etcd-amd64:3.0.17,Command:[etcd --listen-client-urls=http://127.0.0.1:2379 --advertise-client-urls=http://127.0.0.1:2379 --data-dir=/var/lib/etcd],Args:[],WorkingDir:,Ports:[],Env:[],Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[{etcd false /var/lib/etcd <nil>}],LivenessProbe:&Probe{Handler:Handler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/health,Port:2379,Host:127.0.0.1,Scheme:HTTP,HTTPHeaders:[],},TCPSocket:nil,},InitialDelaySeconds:15,TimeoutSeconds:15,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:8,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[],TerminationMessagePolicy:File,} in pod etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.853969 27751 factory.go:112] Using factory "docker" for container "/kubepods/burstable/podf49ee4da5c66af63a0b4bcea4f69baf9/272af7e3b2b1a203250d154349fdf77f296d7b7f65ce2c77b6b3a94e53dba356"
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.857617 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-controller-manager-af867b", UID:"f49ee4da5c66af63a0b4bcea4f69baf9", APIVersion:"v1", ResourceVersion:"", FieldPath:"spec.containers{kube-controller-manager}"}): type: 'Normal' reason: 'Started' Started container
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.984603 27751 manager.go:932] Added container: "/kubepods/burstable/podf49ee4da5c66af63a0b4bcea4f69baf9/272af7e3b2b1a203250d154349fdf77f296d7b7f65ce2c77b6b3a94e53dba356" (aliases: [k8s_kube-controller-manager_kube-controller-manager-af867b_kube-system_f49ee4da5c66af63a0b4bcea4f69baf9_0 272af7e3b2b1a203250d154349fdf77f296d7b7f65ce2c77b6b3a94e53dba356], namespace: "docker")
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.985367 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-scheduler-af867b", UID:"bc22704d9f4dc5d62a8217cfd5c14373", APIVersion:"v1", ResourceVersion:"", FieldPath:"spec.containers{kube-scheduler}"}): type: 'Normal' reason: 'Started' Started container
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.985869 27751 kuberuntime_container.go:100] Generating ref for container etcd: &v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"etcd-af867b", UID:"d76e26fba3bf2bfd215eb29011d55250", APIVersion:"v1", ResourceVersion:"", FieldPath:"spec.containers{etcd}"}
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.985897 27751 container_manager_linux.go:634] Calling devicePluginHandler AllocateDevices
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.985947 27751 kubelet_pods.go:123] container: kube-system/etcd-af867b/etcd podIP: "10.196.65.210" creating hosts mount: true
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.986201 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"etcd-af867b", UID:"d76e26fba3bf2bfd215eb29011d55250", APIVersion:"v1", ResourceVersion:"", FieldPath:"spec.containers{etcd}"}): type: 'Normal' reason: 'Pulled' Container image "gcr.io/google_containers/etcd-amd64:3.0.17" already present on machine
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.986458 27751 handler.go:325] Added event &{/kubepods/burstable/podf49ee4da5c66af63a0b4bcea4f69baf9/272af7e3b2b1a203250d154349fdf77f296d7b7f65ce2c77b6b3a94e53dba356 2017-11-15 01:58:57.645194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.986585 27751 container.go:409] Start housekeeping for container "/kubepods/burstable/podf49ee4da5c66af63a0b4bcea4f69baf9/272af7e3b2b1a203250d154349fdf77f296d7b7f65ce2c77b6b3a94e53dba356"
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.990975 27751 factory.go:112] Using factory "docker" for container "/kubepods/burstable/podbc22704d9f4dc5d62a8217cfd5c14373/413566b22305750f9a9aa46fbe256c11e75293e80f6e0d4afb7ec9e6afcdee05"
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.994167 27751 manager.go:932] Added container: "/kubepods/burstable/podbc22704d9f4dc5d62a8217cfd5c14373/413566b22305750f9a9aa46fbe256c11e75293e80f6e0d4afb7ec9e6afcdee05" (aliases: [k8s_kube-scheduler_kube-scheduler-af867b_kube-system_bc22704d9f4dc5d62a8217cfd5c14373_0 413566b22305750f9a9aa46fbe256c11e75293e80f6e0d4afb7ec9e6afcdee05], namespace: "docker")
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.994298 27751 handler.go:325] Added event &{/kubepods/burstable/podbc22704d9f4dc5d62a8217cfd5c14373/413566b22305750f9a9aa46fbe256c11e75293e80f6e0d4afb7ec9e6afcdee05 2017-11-15 01:58:57.871194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.994331 27751 container.go:409] Start housekeeping for container "/kubepods/burstable/podbc22704d9f4dc5d62a8217cfd5c14373/413566b22305750f9a9aa46fbe256c11e75293e80f6e0d4afb7ec9e6afcdee05"
Nov 15 01:58:57 af867b kubelet[27751]: I1115 01:58:57.996809 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/besteffort/podd76e26fba3bf2bfd215eb29011d55250"
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.005581 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-apiserver-af867b", UID:"4e0fac5dee63099d647b4d031a37ad7d", APIVersion:"v1", ResourceVersion:"", FieldPath:"spec.containers{kube-apiserver}"}): type: 'Normal' reason: 'Created' Created container
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.308559 27751 factory.go:112] Using factory "docker" for container "/kubepods/burstable/pod4e0fac5dee63099d647b4d031a37ad7d/8da2c70a27f08a2f062af80b5708e01ac34ce76b42ab4a6eaa0288e2daf8a043"
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.313617 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-apiserver-af867b", UID:"4e0fac5dee63099d647b4d031a37ad7d", APIVersion:"v1", ResourceVersion:"", FieldPath:"spec.containers{kube-apiserver}"}): type: 'Normal' reason: 'Started' Started container
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.318645 27751 manager.go:932] Added container: "/kubepods/burstable/pod4e0fac5dee63099d647b4d031a37ad7d/8da2c70a27f08a2f062af80b5708e01ac34ce76b42ab4a6eaa0288e2daf8a043" (aliases: [k8s_kube-apiserver_kube-apiserver-af867b_kube-system_4e0fac5dee63099d647b4d031a37ad7d_0 8da2c70a27f08a2f062af80b5708e01ac34ce76b42ab4a6eaa0288e2daf8a043], namespace: "docker")
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.319491 27751 handler.go:325] Added event &{/kubepods/burstable/pod4e0fac5dee63099d647b4d031a37ad7d/8da2c70a27f08a2f062af80b5708e01ac34ce76b42ab4a6eaa0288e2daf8a043 2017-11-15 01:58:58.130194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.319548 27751 container.go:409] Start housekeeping for container "/kubepods/burstable/pod4e0fac5dee63099d647b4d031a37ad7d/8da2c70a27f08a2f062af80b5708e01ac34ce76b42ab4a6eaa0288e2daf8a043"
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.324705 27751 generic.go:146] GenericPLEG: d76e26fba3bf2bfd215eb29011d55250/d5076ad0e9fb270d1b8c4ff7cdadbf32d6b1e30dc42ae24dfbc5cb01bb5aa934: non-existent -> running
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.324752 27751 generic.go:146] GenericPLEG: 4e0fac5dee63099d647b4d031a37ad7d/8da2c70a27f08a2f062af80b5708e01ac34ce76b42ab4a6eaa0288e2daf8a043: non-existent -> running
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.324763 27751 generic.go:146] GenericPLEG: 4e0fac5dee63099d647b4d031a37ad7d/19b06ca14052c17f92bf03d5316fd7642fe4fee6bf193755575355be3e87a13d: non-existent -> running
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.324774 27751 generic.go:146] GenericPLEG: bc22704d9f4dc5d62a8217cfd5c14373/413566b22305750f9a9aa46fbe256c11e75293e80f6e0d4afb7ec9e6afcdee05: non-existent -> running
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.324785 27751 generic.go:146] GenericPLEG: f49ee4da5c66af63a0b4bcea4f69baf9/272af7e3b2b1a203250d154349fdf77f296d7b7f65ce2c77b6b3a94e53dba356: non-existent -> running
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.327367 27751 kuberuntime_manager.go:833] getSandboxIDByPodUID got sandbox IDs ["d5076ad0e9fb270d1b8c4ff7cdadbf32d6b1e30dc42ae24dfbc5cb01bb5aa934"] for pod "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)"
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.328777 27751 generic.go:345] PLEG: Write status for etcd-af867b/kube-system: &container.PodStatus{ID:"d76e26fba3bf2bfd215eb29011d55250", Name:"etcd-af867b", Namespace:"kube-system", IP:"", ContainerStatuses:[]*container.ContainerStatus{}, SandboxStatuses:[]*runtime.PodSandboxStatus{(*runtime.PodSandboxStatus)(0xc4210b2730)}} (err: <nil>)
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.328871 27751 kubelet.go:1871] SyncLoop (PLEG): "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)", event: &pleg.PodLifecycleEvent{ID:"d76e26fba3bf2bfd215eb29011d55250", Type:"ContainerStarted", Data:"d5076ad0e9fb270d1b8c4ff7cdadbf32d6b1e30dc42ae24dfbc5cb01bb5aa934"}
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.329495 27751 kuberuntime_manager.go:833] getSandboxIDByPodUID got sandbox IDs ["19b06ca14052c17f92bf03d5316fd7642fe4fee6bf193755575355be3e87a13d"] for pod "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)"
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.333611 27751 generic.go:345] PLEG: Write status for kube-apiserver-af867b/kube-system: &container.PodStatus{ID:"4e0fac5dee63099d647b4d031a37ad7d", Name:"kube-apiserver-af867b", Namespace:"kube-system", IP:"", ContainerStatuses:[]*container.ContainerStatus{(*container.ContainerStatus)(0xc4200f8b60)}, SandboxStatuses:[]*runtime.PodSandboxStatus{(*runtime.PodSandboxStatus)(0xc421158550)}} (err: <nil>)
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.333697 27751 kubelet.go:1871] SyncLoop (PLEG): "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)", event: &pleg.PodLifecycleEvent{ID:"4e0fac5dee63099d647b4d031a37ad7d", Type:"ContainerStarted", Data:"8da2c70a27f08a2f062af80b5708e01ac34ce76b42ab4a6eaa0288e2daf8a043"}
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.333747 27751 kubelet.go:1871] SyncLoop (PLEG): "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)", event: &pleg.PodLifecycleEvent{ID:"4e0fac5dee63099d647b4d031a37ad7d", Type:"ContainerStarted", Data:"19b06ca14052c17f92bf03d5316fd7642fe4fee6bf193755575355be3e87a13d"}
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.333782 27751 kubelet_pods.go:1284] Generating status for "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)"
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.333847 27751 kubelet_node_status.go:280] Setting node annotation to enable volume controller attach/detach
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.335324 27751 kuberuntime_manager.go:833] getSandboxIDByPodUID got sandbox IDs ["fe18d337ce42bd3a4d2aacb1349be028274681d8073fbe5d6377946e815fa810"] for pod "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)"
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.336722 27751 kubelet_node_status.go:443] Recording NodeHasSufficientDisk event message for node af867b
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.336743 27751 kubelet_node_status.go:443] Recording NodeHasSufficientMemory event message for node af867b
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.336753 27751 kubelet_node_status.go:443] Recording NodeHasNoDiskPressure event message for node af867b
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.336943 27751 kubelet.go:1610] Creating a mirror pod for static pod "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)"
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.336971 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientMemory' Node af867b status is now: NodeHasSufficientMemory
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.336988 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientDisk' Node af867b status is now: NodeHasSufficientDisk
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.337003 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasNoDiskPressure' Node af867b status is now: NodeHasNoDiskPressure
Nov 15 01:58:58 af867b kubelet[27751]: W1115 01:58:58.339143 27751 status_manager.go:431] Failed to get status for pod "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)": Get https://10.241.226.117:6443/api/v1/namespaces/kube-system/pods/kube-apiserver-af867b: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:58 af867b kubelet[27751]: E1115 01:58:58.339240 27751 kubelet.go:1612] Failed creating a mirror pod for "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)": Post https://10.241.226.117:6443/api/v1/namespaces/kube-system/pods: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.339278 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)"
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.341569 27751 generic.go:345] PLEG: Write status for kube-scheduler-af867b/kube-system: &container.PodStatus{ID:"bc22704d9f4dc5d62a8217cfd5c14373", Name:"kube-scheduler-af867b", Namespace:"kube-system", IP:"", ContainerStatuses:[]*container.ContainerStatus{(*container.ContainerStatus)(0xc42066e700)}, SandboxStatuses:[]*runtime.PodSandboxStatus{(*runtime.PodSandboxStatus)(0xc421159d60)}} (err: <nil>)
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.341651 27751 kubelet.go:1871] SyncLoop (PLEG): "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)", event: &pleg.PodLifecycleEvent{ID:"bc22704d9f4dc5d62a8217cfd5c14373", Type:"ContainerStarted", Data:"413566b22305750f9a9aa46fbe256c11e75293e80f6e0d4afb7ec9e6afcdee05"}
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.341685 27751 kubelet_pods.go:1284] Generating status for "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)"
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.341767 27751 kubelet_node_status.go:280] Setting node annotation to enable volume controller attach/detach
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.343429 27751 kuberuntime_manager.go:833] getSandboxIDByPodUID got sandbox IDs ["439f518d46ec0d596518b7c17c503c52a85dd4da21d8a56ec174d5b6b3b98e03"] for pod "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)"
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.344546 27751 kubelet_node_status.go:443] Recording NodeHasSufficientDisk event message for node af867b
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.344571 27751 kubelet_node_status.go:443] Recording NodeHasSufficientMemory event message for node af867b
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.344580 27751 kubelet_node_status.go:443] Recording NodeHasNoDiskPressure event message for node af867b
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.344724 27751 kubelet.go:1610] Creating a mirror pod for static pod "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)"
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.344892 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientDisk' Node af867b status is now: NodeHasSufficientDisk
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.344912 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientMemory' Node af867b status is now: NodeHasSufficientMemory
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.344926 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasNoDiskPressure' Node af867b status is now: NodeHasNoDiskPressure
Nov 15 01:58:58 af867b kubelet[27751]: W1115 01:58:58.346555 27751 status_manager.go:431] Failed to get status for pod "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)": Get https://10.241.226.117:6443/api/v1/namespaces/kube-system/pods/kube-scheduler-af867b: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:58 af867b kubelet[27751]: E1115 01:58:58.346617 27751 kubelet.go:1612] Failed creating a mirror pod for "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)": Post https://10.241.226.117:6443/api/v1/namespaces/kube-system/pods: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.346660 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)"
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.348757 27751 generic.go:345] PLEG: Write status for kube-controller-manager-af867b/kube-system: &container.PodStatus{ID:"f49ee4da5c66af63a0b4bcea4f69baf9", Name:"kube-controller-manager-af867b", Namespace:"kube-system", IP:"", ContainerStatuses:[]*container.ContainerStatus{(*container.ContainerStatus)(0xc4200f9180)}, SandboxStatuses:[]*runtime.PodSandboxStatus{(*runtime.PodSandboxStatus)(0xc4211e5090)}} (err: <nil>)
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.348821 27751 kubelet.go:1871] SyncLoop (PLEG): "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)", event: &pleg.PodLifecycleEvent{ID:"f49ee4da5c66af63a0b4bcea4f69baf9", Type:"ContainerStarted", Data:"272af7e3b2b1a203250d154349fdf77f296d7b7f65ce2c77b6b3a94e53dba356"}
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.348853 27751 kubelet_pods.go:1284] Generating status for "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)"
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.348901 27751 kubelet_node_status.go:280] Setting node annotation to enable volume controller attach/detach
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.350882 27751 kubelet_node_status.go:443] Recording NodeHasSufficientDisk event message for node af867b
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.350900 27751 kubelet_node_status.go:443] Recording NodeHasSufficientMemory event message for node af867b
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.350909 27751 kubelet_node_status.go:443] Recording NodeHasNoDiskPressure event message for node af867b
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.351040 27751 kubelet.go:1610] Creating a mirror pod for static pod "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)"
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.351222 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientDisk' Node af867b status is now: NodeHasSufficientDisk
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.351242 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientMemory' Node af867b status is now: NodeHasSufficientMemory
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.351256 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasNoDiskPressure' Node af867b status is now: NodeHasNoDiskPressure
Nov 15 01:58:58 af867b kubelet[27751]: E1115 01:58:58.351481 27751 kubelet.go:1612] Failed creating a mirror pod for "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)": Post https://10.241.226.117:6443/api/v1/namespaces/kube-system/pods: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.351516 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)"
Nov 15 01:58:58 af867b kubelet[27751]: W1115 01:58:58.351643 27751 status_manager.go:431] Failed to get status for pod "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)": Get https://10.241.226.117:6443/api/v1/namespaces/kube-system/pods/kube-controller-manager-af867b: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.384333 27751 reflector.go:240] Listing and watching *v1.Node from k8s.io/kubernetes/pkg/kubelet/kubelet.go:422
Nov 15 01:58:58 af867b kubelet[27751]: E1115 01:58:58.385242 27751 reflector.go:205] k8s.io/kubernetes/pkg/kubelet/kubelet.go:422: Failed to list *v1.Node: Get https://10.241.226.117:6443/api/v1/nodes?fieldSelector=metadata.name%3Daf867b&resourceVersion=0: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.385571 27751 reflector.go:240] Listing and watching *v1.Pod from k8s.io/kubernetes/pkg/kubelet/config/apiserver.go:47
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.386551 27751 reflector.go:240] Listing and watching *v1.Service from k8s.io/kubernetes/pkg/kubelet/kubelet.go:413
Nov 15 01:58:58 af867b kubelet[27751]: E1115 01:58:58.411993 27751 reflector.go:205] k8s.io/kubernetes/pkg/kubelet/config/apiserver.go:47: Failed to list *v1.Pod: Get https://10.241.226.117:6443/api/v1/pods?fieldSelector=spec.nodeName%3Daf867b&resourceVersion=0: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.477762 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"etcd-af867b", UID:"d76e26fba3bf2bfd215eb29011d55250", APIVersion:"v1", ResourceVersion:"", FieldPath:"spec.containers{etcd}"}): type: 'Normal' reason: 'Created' Created container
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.524042 27751 kubelet.go:1911] SyncLoop (housekeeping, skipped): sources aren't ready yet.
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.611626 27751 request.go:462] Throttling request took 224.966026ms, request: GET:https://10.241.226.117:6443/api/v1/services?resourceVersion=0
Nov 15 01:58:58 af867b kubelet[27751]: E1115 01:58:58.612582 27751 reflector.go:205] k8s.io/kubernetes/pkg/kubelet/kubelet.go:413: Failed to list *v1.Service: Get https://10.241.226.117:6443/api/v1/services?resourceVersion=0: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.639700 27751 volume_manager.go:366] All volumes are attached and mounted for pod "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)"
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.639987 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:false CreateSandbox:false SandboxID:19b06ca14052c17f92bf03d5316fd7642fe4fee6bf193755575355be3e87a13d Attempt:0 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)"
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.649341 27751 volume_manager.go:366] All volumes are attached and mounted for pod "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)"
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.649486 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:false CreateSandbox:false SandboxID:fe18d337ce42bd3a4d2aacb1349be028274681d8073fbe5d6377946e815fa810 Attempt:0 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)"
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.651790 27751 volume_manager.go:366] All volumes are attached and mounted for pod "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)"
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.651889 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:false CreateSandbox:false SandboxID:439f518d46ec0d596518b7c17c503c52a85dd4da21d8a56ec174d5b6b3b98e03 Attempt:0 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)"
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.661796 27751 factory.go:112] Using factory "docker" for container "/kubepods/besteffort/podd76e26fba3bf2bfd215eb29011d55250/ab147bb1b65b2001333417bb7654896e6aadb25ce71a8c48c94ae802a2e0197f"
Nov 15 01:58:58 af867b kubelet[27751]: W1115 01:58:58.663948 27751 docker_container.go:202] Deleted previously existing symlink file: "/var/log/pods/d76e26fba3bf2bfd215eb29011d55250/etcd_0.log"
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.665156 27751 manager.go:932] Added container: "/kubepods/besteffort/podd76e26fba3bf2bfd215eb29011d55250/ab147bb1b65b2001333417bb7654896e6aadb25ce71a8c48c94ae802a2e0197f" (aliases: [k8s_etcd_etcd-af867b_kube-system_d76e26fba3bf2bfd215eb29011d55250_0 ab147bb1b65b2001333417bb7654896e6aadb25ce71a8c48c94ae802a2e0197f], namespace: "docker")
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.665292 27751 handler.go:325] Added event &{/kubepods/besteffort/podd76e26fba3bf2bfd215eb29011d55250/ab147bb1b65b2001333417bb7654896e6aadb25ce71a8c48c94ae802a2e0197f 2017-11-15 01:58:58.571194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.665334 27751 container.go:409] Start housekeeping for container "/kubepods/besteffort/podd76e26fba3bf2bfd215eb29011d55250/ab147bb1b65b2001333417bb7654896e6aadb25ce71a8c48c94ae802a2e0197f"
Nov 15 01:58:58 af867b kubelet[27751]: I1115 01:58:58.670879 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"etcd-af867b", UID:"d76e26fba3bf2bfd215eb29011d55250", APIVersion:"v1", ResourceVersion:"", FieldPath:"spec.containers{etcd}"}): type: 'Normal' reason: 'Started' Started container
Nov 15 01:58:59 af867b kubelet[27751]: E1115 01:58:59.197618 27751 event.go:209] Unable to write event: 'Post https://10.241.226.117:6443/api/v1/namespaces/default/events: dial tcp 10.241.226.117:6443: getsockopt: connection refused' (may retry after sleeping)
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.352350 27751 generic.go:146] GenericPLEG: d76e26fba3bf2bfd215eb29011d55250/ab147bb1b65b2001333417bb7654896e6aadb25ce71a8c48c94ae802a2e0197f: non-existent -> running
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.353508 27751 kuberuntime_manager.go:833] getSandboxIDByPodUID got sandbox IDs ["d5076ad0e9fb270d1b8c4ff7cdadbf32d6b1e30dc42ae24dfbc5cb01bb5aa934"] for pod "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)"
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.357306 27751 generic.go:345] PLEG: Write status for etcd-af867b/kube-system: &container.PodStatus{ID:"d76e26fba3bf2bfd215eb29011d55250", Name:"etcd-af867b", Namespace:"kube-system", IP:"", ContainerStatuses:[]*container.ContainerStatus{(*container.ContainerStatus)(0xc42066f880)}, SandboxStatuses:[]*runtime.PodSandboxStatus{(*runtime.PodSandboxStatus)(0xc4212b1b30)}} (err: <nil>)
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.357373 27751 kubelet_pods.go:1284] Generating status for "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)"
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.357455 27751 kubelet_node_status.go:280] Setting node annotation to enable volume controller attach/detach
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.358393 27751 kubelet.go:1871] SyncLoop (PLEG): "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)", event: &pleg.PodLifecycleEvent{ID:"d76e26fba3bf2bfd215eb29011d55250", Type:"ContainerStarted", Data:"ab147bb1b65b2001333417bb7654896e6aadb25ce71a8c48c94ae802a2e0197f"}
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.358448 27751 kubelet_pods.go:1284] Generating status for "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)"
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.358494 27751 kubelet_node_status.go:280] Setting node annotation to enable volume controller attach/detach
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.358516 27751 kubelet_pods.go:1284] Generating status for "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)"
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.358571 27751 kubelet_node_status.go:280] Setting node annotation to enable volume controller attach/detach
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.358593 27751 kubelet_pods.go:1284] Generating status for "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)"
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.358631 27751 kubelet_node_status.go:280] Setting node annotation to enable volume controller attach/detach
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.363984 27751 kubelet_node_status.go:443] Recording NodeHasSufficientDisk event message for node af867b
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.364011 27751 kubelet_node_status.go:443] Recording NodeHasSufficientMemory event message for node af867b
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.364025 27751 kubelet_node_status.go:443] Recording NodeHasNoDiskPressure event message for node af867b
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.364213 27751 kubelet.go:1610] Creating a mirror pod for static pod "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)"
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.364498 27751 kubelet_node_status.go:443] Recording NodeHasSufficientDisk event message for node af867b
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.364513 27751 kubelet_node_status.go:443] Recording NodeHasSufficientMemory event message for node af867b
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.364522 27751 kubelet_node_status.go:443] Recording NodeHasNoDiskPressure event message for node af867b
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.364619 27751 status_manager.go:325] Ignoring same status for pod "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)", status: {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:58 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP:10.196.65.210 StartTime:2017-11-15 01:58:55 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:kube-apiserver State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 01:58:58 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:gcr.io/google_containers/kube-apiserver-amd64:v1.8.3 ImageID:docker-pullable://gcr.io/google_containers/kube-apiserver-amd64@sha256:872e3d4286a8ef4338df59945cb0d64c2622268ceb3e8a2ce7b52243279b02d0 ContainerID:docker://8da2c70a27f08a2f062af80b5708e01ac34ce76b42ab4a6eaa0288e2daf8a043}] QOSClass:Burstable}
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.364818 27751 kubelet.go:1610] Creating a mirror pod for static pod "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)"
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.365012 27751 kubelet_node_status.go:443] Recording NodeHasSufficientDisk event message for node af867b
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.365026 27751 kubelet_node_status.go:443] Recording NodeHasSufficientMemory event message for node af867b
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.365037 27751 kubelet_node_status.go:443] Recording NodeHasNoDiskPressure event message for node af867b
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.365111 27751 status_manager.go:325] Ignoring same status for pod "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)", status: {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:58 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP:10.196.65.210 StartTime:2017-11-15 01:58:55 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:kube-scheduler State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 01:58:57 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:gcr.io/google_containers/kube-scheduler-amd64:v1.8.3 ImageID:docker-pullable://gcr.io/google_containers/kube-scheduler-amd64@sha256:c47b2438bbab28d58e8cbf64b37b7f66d26b000f5c3a31626ee829a4be8fb91e ContainerID:docker://413566b22305750f9a9aa46fbe256c11e75293e80f6e0d4afb7ec9e6afcdee05}] QOSClass:Burstable}
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.365254 27751 kubelet.go:1610] Creating a mirror pod for static pod "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)"
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.365441 27751 kubelet_node_status.go:443] Recording NodeHasSufficientDisk event message for node af867b
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.365466 27751 kubelet_node_status.go:443] Recording NodeHasSufficientMemory event message for node af867b
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.365476 27751 kubelet_node_status.go:443] Recording NodeHasNoDiskPressure event message for node af867b
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.365551 27751 status_manager.go:325] Ignoring same status for pod "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)", status: {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:58 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP:10.196.65.210 StartTime:2017-11-15 01:58:55 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:kube-controller-manager State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 01:58:57 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:gcr.io/google_containers/kube-controller-manager-amd64:v1.8.3 ImageID:docker-pullable://gcr.io/google_containers/kube-controller-manager-amd64@sha256:b6b633e3e107761d38fceb200f01bf552c51f65e3524b0aafc1a7710afff07be ContainerID:docker://272af7e3b2b1a203250d154349fdf77f296d7b7f65ce2c77b6b3a94e53dba356}] QOSClass:Burstable}
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.365672 27751 kubelet.go:1610] Creating a mirror pod for static pod "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)"
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.365771 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasNoDiskPressure' Node af867b status is now: NodeHasNoDiskPressure
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.365792 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientDisk' Node af867b status is now: NodeHasSufficientDisk
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.365805 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientMemory' Node af867b status is now: NodeHasSufficientMemory
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.365819 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasNoDiskPressure' Node af867b status is now: NodeHasNoDiskPressure
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.365867 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientDisk' Node af867b status is now: NodeHasSufficientDisk
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.365882 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientMemory' Node af867b status is now: NodeHasSufficientMemory
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.365895 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasNoDiskPressure' Node af867b status is now: NodeHasNoDiskPressure
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.365906 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientDisk' Node af867b status is now: NodeHasSufficientDisk
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.365917 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientMemory' Node af867b status is now: NodeHasSufficientMemory
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.365929 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasNoDiskPressure' Node af867b status is now: NodeHasNoDiskPressure
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.365940 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientDisk' Node af867b status is now: NodeHasSufficientDisk
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.365952 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientMemory' Node af867b status is now: NodeHasSufficientMemory
Nov 15 01:58:59 af867b kubelet[27751]: E1115 01:58:59.366022 27751 kubelet.go:1612] Failed creating a mirror pod for "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)": Post https://10.241.226.117:6443/api/v1/namespaces/kube-system/pods: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.366062 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)"
Nov 15 01:58:59 af867b kubelet[27751]: E1115 01:58:59.366146 27751 kubelet.go:1612] Failed creating a mirror pod for "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)": Post https://10.241.226.117:6443/api/v1/namespaces/kube-system/pods: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.366175 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)"
Nov 15 01:58:59 af867b kubelet[27751]: E1115 01:58:59.366239 27751 kubelet.go:1612] Failed creating a mirror pod for "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)": Post https://10.241.226.117:6443/api/v1/namespaces/kube-system/pods: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.366265 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)"
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.389829 27751 reflector.go:240] Listing and watching *v1.Node from k8s.io/kubernetes/pkg/kubelet/kubelet.go:422
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.412890 27751 reflector.go:240] Listing and watching *v1.Pod from k8s.io/kubernetes/pkg/kubelet/config/apiserver.go:47
Nov 15 01:58:59 af867b kubelet[27751]: E1115 01:58:59.414085 27751 kubelet.go:1612] Failed creating a mirror pod for "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)": Post https://10.241.226.117:6443/api/v1/namespaces/kube-system/pods: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.414170 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)"
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.611825 27751 request.go:462] Throttling request took 245.912534ms, request: GET:https://10.241.226.117:6443/api/v1/namespaces/kube-system/pods/etcd-af867b
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.612846 27751 reflector.go:240] Listing and watching *v1.Service from k8s.io/kubernetes/pkg/kubelet/kubelet.go:413
Nov 15 01:58:59 af867b kubelet[27751]: W1115 01:58:59.613045 27751 status_manager.go:431] Failed to get status for pod "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)": Get https://10.241.226.117:6443/api/v1/namespaces/kube-system/pods/etcd-af867b: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.666597 27751 volume_manager.go:366] All volumes are attached and mounted for pod "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)"
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.666813 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:false CreateSandbox:false SandboxID:fe18d337ce42bd3a4d2aacb1349be028274681d8073fbe5d6377946e815fa810 Attempt:0 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)"
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.666977 27751 volume_manager.go:366] All volumes are attached and mounted for pod "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)"
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.667062 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:false CreateSandbox:false SandboxID:19b06ca14052c17f92bf03d5316fd7642fe4fee6bf193755575355be3e87a13d Attempt:0 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)"
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.667151 27751 volume_manager.go:366] All volumes are attached and mounted for pod "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)"
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.667215 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:false CreateSandbox:false SandboxID:d5076ad0e9fb270d1b8c4ff7cdadbf32d6b1e30dc42ae24dfbc5cb01bb5aa934 Attempt:0 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)"
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.714387 27751 volume_manager.go:366] All volumes are attached and mounted for pod "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)"
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.714563 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:false CreateSandbox:false SandboxID:439f518d46ec0d596518b7c17c503c52a85dd4da21d8a56ec174d5b6b3b98e03 Attempt:0 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)"
Nov 15 01:58:59 af867b kubelet[27751]: I1115 01:58:59.811633 27751 request.go:462] Throttling request took 421.650806ms, request: GET:https://10.241.226.117:6443/api/v1/nodes?fieldSelector=metadata.name%3Daf867b&resourceVersion=0
Nov 15 01:58:59 af867b kubelet[27751]: E1115 01:58:59.813031 27751 reflector.go:205] k8s.io/kubernetes/pkg/kubelet/kubelet.go:422: Failed to list *v1.Node: Get https://10.241.226.117:6443/api/v1/nodes?fieldSelector=metadata.name%3Daf867b&resourceVersion=0: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:59:00 af867b kubelet[27751]: I1115 01:59:00.011855 27751 request.go:462] Throttling request took 598.805681ms, request: GET:https://10.241.226.117:6443/api/v1/pods?fieldSelector=spec.nodeName%3Daf867b&resourceVersion=0
Nov 15 01:59:00 af867b kubelet[27751]: E1115 01:59:00.018890 27751 reflector.go:205] k8s.io/kubernetes/pkg/kubelet/config/apiserver.go:47: Failed to list *v1.Pod: Get https://10.241.226.117:6443/api/v1/pods?fieldSelector=spec.nodeName%3Daf867b&resourceVersion=0: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:59:00 af867b kubelet[27751]: I1115 01:59:00.211647 27751 request.go:462] Throttling request took 598.663516ms, request: GET:https://10.241.226.117:6443/api/v1/services?resourceVersion=0
Nov 15 01:59:00 af867b kubelet[27751]: E1115 01:59:00.212865 27751 reflector.go:205] k8s.io/kubernetes/pkg/kubelet/kubelet.go:413: Failed to list *v1.Service: Get https://10.241.226.117:6443/api/v1/services?resourceVersion=0: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:59:00 af867b kubelet[27751]: I1115 01:59:00.360737 27751 kubelet_pods.go:1284] Generating status for "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)"
Nov 15 01:59:00 af867b kubelet[27751]: I1115 01:59:00.360816 27751 kubelet_node_status.go:280] Setting node annotation to enable volume controller attach/detach
Nov 15 01:59:00 af867b kubelet[27751]: I1115 01:59:00.363201 27751 kubelet_node_status.go:443] Recording NodeHasSufficientDisk event message for node af867b
Nov 15 01:59:00 af867b kubelet[27751]: I1115 01:59:00.363229 27751 kubelet_node_status.go:443] Recording NodeHasSufficientMemory event message for node af867b
Nov 15 01:59:00 af867b kubelet[27751]: I1115 01:59:00.363241 27751 kubelet_node_status.go:443] Recording NodeHasNoDiskPressure event message for node af867b
Nov 15 01:59:00 af867b kubelet[27751]: I1115 01:59:00.363363 27751 status_manager.go:325] Ignoring same status for pod "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)", status: {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:59 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP:10.196.65.210 StartTime:2017-11-15 01:58:55 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:etcd State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 01:58:58 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:gcr.io/google_containers/etcd-amd64:3.0.17 ImageID:docker-pullable://gcr.io/google_containers/etcd-amd64@sha256:d83d3545e06fb035db8512e33bd44afb55dea007a3abd7b17742d3ac6d235940 ContainerID:docker://ab147bb1b65b2001333417bb7654896e6aadb25ce71a8c48c94ae802a2e0197f}] QOSClass:BestEffort}
Nov 15 01:59:00 af867b kubelet[27751]: I1115 01:59:00.363522 27751 kubelet.go:1610] Creating a mirror pod for static pod "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)"
Nov 15 01:59:00 af867b kubelet[27751]: I1115 01:59:00.363648 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasNoDiskPressure' Node af867b status is now: NodeHasNoDiskPressure
Nov 15 01:59:00 af867b kubelet[27751]: I1115 01:59:00.363668 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientDisk' Node af867b status is now: NodeHasSufficientDisk
Nov 15 01:59:00 af867b kubelet[27751]: I1115 01:59:00.363680 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientMemory' Node af867b status is now: NodeHasSufficientMemory
Nov 15 01:59:00 af867b kubelet[27751]: I1115 01:59:00.415814 27751 request.go:462] Throttling request took 52.162252ms, request: POST:https://10.241.226.117:6443/api/v1/namespaces/kube-system/pods
Nov 15 01:59:00 af867b kubelet[27751]: E1115 01:59:00.416836 27751 kubelet.go:1612] Failed creating a mirror pod for "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)": Post https://10.241.226.117:6443/api/v1/namespaces/kube-system/pods: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:59:00 af867b kubelet[27751]: I1115 01:59:00.416888 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)"
Nov 15 01:59:00 af867b kubelet[27751]: I1115 01:59:00.523701 27751 kubelet.go:1911] SyncLoop (housekeeping, skipped): sources aren't ready yet.
Nov 15 01:59:00 af867b kubelet[27751]: I1115 01:59:00.717163 27751 volume_manager.go:366] All volumes are attached and mounted for pod "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)"
Nov 15 01:59:00 af867b kubelet[27751]: I1115 01:59:00.717299 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:false CreateSandbox:false SandboxID:d5076ad0e9fb270d1b8c4ff7cdadbf32d6b1e30dc42ae24dfbc5cb01bb5aa934 Attempt:0 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)"
Nov 15 01:59:00 af867b kubelet[27751]: I1115 01:59:00.796935 27751 eviction_manager.go:221] eviction manager: synchronize housekeeping
Nov 15 01:59:00 af867b kubelet[27751]: E1115 01:59:00.797006 27751 eviction_manager.go:238] eviction manager: unexpected err: failed to get node info: node 'af867b' not found
Nov 15 01:59:00 af867b kubelet[27751]: W1115 01:59:00.802145 27751 cni.go:196] Unable to update cni config: No networks found in /etc/cni/net.d
Nov 15 01:59:00 af867b kubelet[27751]: I1115 01:59:00.802294 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 01:59:00 af867b kubelet[27751]: E1115 01:59:00.802323 27751 kubelet.go:2095] Container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 01:59:00 af867b kubelet[27751]: I1115 01:59:00.813766 27751 reflector.go:240] Listing and watching *v1.Node from k8s.io/kubernetes/pkg/kubelet/kubelet.go:422
Nov 15 01:59:00 af867b kubelet[27751]: E1115 01:59:00.814565 27751 reflector.go:205] k8s.io/kubernetes/pkg/kubelet/kubelet.go:422: Failed to list *v1.Node: Get https://10.241.226.117:6443/api/v1/nodes?fieldSelector=metadata.name%3Daf867b&resourceVersion=0: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:59:01 af867b kubelet[27751]: I1115 01:59:01.019131 27751 reflector.go:240] Listing and watching *v1.Pod from k8s.io/kubernetes/pkg/kubelet/config/apiserver.go:47
Nov 15 01:59:01 af867b kubelet[27751]: E1115 01:59:01.020532 27751 reflector.go:205] k8s.io/kubernetes/pkg/kubelet/config/apiserver.go:47: Failed to list *v1.Pod: Get https://10.241.226.117:6443/api/v1/pods?fieldSelector=spec.nodeName%3Daf867b&resourceVersion=0: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:59:01 af867b kubelet[27751]: I1115 01:59:01.213083 27751 reflector.go:240] Listing and watching *v1.Service from k8s.io/kubernetes/pkg/kubelet/kubelet.go:413
Nov 15 01:59:01 af867b kubelet[27751]: E1115 01:59:01.214349 27751 reflector.go:205] k8s.io/kubernetes/pkg/kubelet/kubelet.go:413: Failed to list *v1.Service: Get https://10.241.226.117:6443/api/v1/services?resourceVersion=0: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:59:01 af867b kubelet[27751]: I1115 01:59:01.814779 27751 reflector.go:240] Listing and watching *v1.Node from k8s.io/kubernetes/pkg/kubelet/kubelet.go:422
Nov 15 01:59:01 af867b kubelet[27751]: E1115 01:59:01.816108 27751 reflector.go:205] k8s.io/kubernetes/pkg/kubelet/kubelet.go:422: Failed to list *v1.Node: Get https://10.241.226.117:6443/api/v1/nodes?fieldSelector=metadata.name%3Daf867b&resourceVersion=0: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:59:02 af867b kubelet[27751]: I1115 01:59:02.020726 27751 reflector.go:240] Listing and watching *v1.Pod from k8s.io/kubernetes/pkg/kubelet/config/apiserver.go:47
Nov 15 01:59:02 af867b kubelet[27751]: E1115 01:59:02.022074 27751 reflector.go:205] k8s.io/kubernetes/pkg/kubelet/config/apiserver.go:47: Failed to list *v1.Pod: Get https://10.241.226.117:6443/api/v1/pods?fieldSelector=spec.nodeName%3Daf867b&resourceVersion=0: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:59:02 af867b kubelet[27751]: I1115 01:59:02.214812 27751 reflector.go:240] Listing and watching *v1.Service from k8s.io/kubernetes/pkg/kubelet/kubelet.go:413
Nov 15 01:59:02 af867b kubelet[27751]: E1115 01:59:02.215854 27751 reflector.go:205] k8s.io/kubernetes/pkg/kubelet/kubelet.go:413: Failed to list *v1.Service: Get https://10.241.226.117:6443/api/v1/services?resourceVersion=0: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:59:02 af867b kubelet[27751]: I1115 01:59:02.527843 27751 kubelet.go:1911] SyncLoop (housekeeping, skipped): sources aren't ready yet.
Nov 15 01:59:02 af867b kubelet[27751]: I1115 01:59:02.816796 27751 reflector.go:240] Listing and watching *v1.Node from k8s.io/kubernetes/pkg/kubelet/kubelet.go:422
Nov 15 01:59:02 af867b kubelet[27751]: E1115 01:59:02.817699 27751 reflector.go:205] k8s.io/kubernetes/pkg/kubelet/kubelet.go:422: Failed to list *v1.Node: Get https://10.241.226.117:6443/api/v1/nodes?fieldSelector=metadata.name%3Daf867b&resourceVersion=0: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:59:03 af867b kubelet[27751]: I1115 01:59:03.022798 27751 reflector.go:240] Listing and watching *v1.Pod from k8s.io/kubernetes/pkg/kubelet/config/apiserver.go:47
Nov 15 01:59:03 af867b kubelet[27751]: E1115 01:59:03.024213 27751 reflector.go:205] k8s.io/kubernetes/pkg/kubelet/config/apiserver.go:47: Failed to list *v1.Pod: Get https://10.241.226.117:6443/api/v1/pods?fieldSelector=spec.nodeName%3Daf867b&resourceVersion=0: dial tcp 10.241.226.117:6443: getsockopt: connection refused
Nov 15 01:59:03 af867b kubelet[27751]: I1115 01:59:03.216799 27751 reflector.go:240] Listing and watching *v1.Service from k8s.io/kubernetes/pkg/kubelet/kubelet.go:413
Nov 15 01:59:03 af867b kubelet[27751]: I1115 01:59:03.264780 27751 kubelet_node_status.go:280] Setting node annotation to enable volume controller attach/detach
Nov 15 01:59:03 af867b kubelet[27751]: I1115 01:59:03.268197 27751 kubelet_node_status.go:443] Recording NodeHasSufficientDisk event message for node af867b
Nov 15 01:59:03 af867b kubelet[27751]: I1115 01:59:03.268231 27751 kubelet_node_status.go:443] Recording NodeHasSufficientMemory event message for node af867b
Nov 15 01:59:03 af867b kubelet[27751]: I1115 01:59:03.268242 27751 kubelet_node_status.go:443] Recording NodeHasNoDiskPressure event message for node af867b
Nov 15 01:59:03 af867b kubelet[27751]: I1115 01:59:03.268258 27751 kubelet_node_status.go:83] Attempting to register node af867b
Nov 15 01:59:03 af867b kubelet[27751]: I1115 01:59:03.268794 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientMemory' Node af867b status is now: NodeHasSufficientMemory
Nov 15 01:59:03 af867b kubelet[27751]: I1115 01:59:03.268827 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasSufficientDisk' Node af867b status is now: NodeHasSufficientDisk
Nov 15 01:59:03 af867b kubelet[27751]: I1115 01:59:03.268842 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeHasNoDiskPressure' Node af867b status is now: NodeHasNoDiskPressure
Nov 15 01:59:03 af867b kubelet[27751]: I1115 01:59:03.817934 27751 reflector.go:240] Listing and watching *v1.Node from k8s.io/kubernetes/pkg/kubelet/kubelet.go:422
Nov 15 01:59:04 af867b kubelet[27751]: I1115 01:59:04.024437 27751 reflector.go:240] Listing and watching *v1.Pod from k8s.io/kubernetes/pkg/kubelet/config/apiserver.go:47
Nov 15 01:59:04 af867b kubelet[27751]: I1115 01:59:04.027388 27751 config.go:282] Setting pods for source api
Nov 15 01:59:04 af867b kubelet[27751]: I1115 01:59:04.027423 27751 kubelet.go:1837] SyncLoop (ADD, "api"): ""
Nov 15 01:59:04 af867b kubelet[27751]: I1115 01:59:04.523445 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 01:59:04 af867b kubelet[27751]: I1115 01:59:04.531383 27751 kubelet_pods.go:1704] Orphaned pod "42253414d7c5f285b756a2243a4df250" found, removing pod cgroups
Nov 15 01:59:04 af867b kubelet[27751]: I1115 01:59:04.531405 27751 kubelet_pods.go:1704] Orphaned pod "9d6dd5e700f66143c0b1a919b27a8a33" found, removing pod cgroups
Nov 15 01:59:04 af867b kubelet[27751]: I1115 01:59:04.531417 27751 kubelet_pods.go:1704] Orphaned pod "b69bc062-c962-11e7-83ed-c6b053eac242" found, removing pod cgroups
Nov 15 01:59:04 af867b kubelet[27751]: I1115 01:59:04.536066 27751 manager.go:989] Destroyed container: "/kubepods/burstable/pod42253414d7c5f285b756a2243a4df250" (aliases: [], namespace: "")
Nov 15 01:59:04 af867b kubelet[27751]: I1115 01:59:04.536094 27751 handler.go:325] Added event &{/kubepods/burstable/pod42253414d7c5f285b756a2243a4df250 2017-11-15 01:59:04.536085777 +0000 UTC containerDeletion {<nil>}}
Nov 15 01:59:04 af867b kubelet[27751]: I1115 01:59:04.536123 27751 manager.go:989] Destroyed container: "/kubepods/burstable/pod9d6dd5e700f66143c0b1a919b27a8a33" (aliases: [], namespace: "")
Nov 15 01:59:04 af867b kubelet[27751]: I1115 01:59:04.536131 27751 handler.go:325] Added event &{/kubepods/burstable/pod9d6dd5e700f66143c0b1a919b27a8a33 2017-11-15 01:59:04.536128762 +0000 UTC containerDeletion {<nil>}}
Nov 15 01:59:04 af867b kubelet[27751]: I1115 01:59:04.536142 27751 manager.go:989] Destroyed container: "/kubepods/besteffort/podb69bc062-c962-11e7-83ed-c6b053eac242" (aliases: [], namespace: "")
Nov 15 01:59:04 af867b kubelet[27751]: I1115 01:59:04.536149 27751 handler.go:325] Added event &{/kubepods/besteffort/podb69bc062-c962-11e7-83ed-c6b053eac242 2017-11-15 01:59:04.536147125 +0000 UTC containerDeletion {<nil>}}
Nov 15 01:59:05 af867b kubelet[27751]: W1115 01:59:05.803524 27751 cni.go:196] Unable to update cni config: No networks found in /etc/cni/net.d
Nov 15 01:59:05 af867b kubelet[27751]: I1115 01:59:05.803667 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 01:59:05 af867b kubelet[27751]: E1115 01:59:05.803690 27751 kubelet.go:2095] Container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 01:59:06 af867b kubelet[27751]: I1115 01:59:06.523502 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 01:59:07 af867b kubelet[27751]: I1115 01:59:07.284683 27751 kubelet_node_status.go:86] Successfully registered node af867b
Nov 15 01:59:07 af867b kubelet[27751]: E1115 01:59:07.287582 27751 kubelet_node_status.go:390] Error updating node status, will retry: error getting node "af867b": nodes "af867b" not found
Nov 15 01:59:08 af867b kubelet[27751]: I1115 01:59:08.523470 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 01:59:09 af867b kubelet[27751]: I1115 01:59:09.439469 27751 request.go:462] Throttling request took 187.939948ms, request: PATCH:https://10.241.226.117:6443/api/v1/namespaces/default/events/af867b.14f71fc4e507c284
Nov 15 01:59:09 af867b kubelet[27751]: I1115 01:59:09.639478 27751 request.go:462] Throttling request took 194.092755ms, request: PATCH:https://10.241.226.117:6443/api/v1/namespaces/default/events/af867b.14f71fc4e5080948
Nov 15 01:59:09 af867b kubelet[27751]: I1115 01:59:09.839485 27751 request.go:462] Throttling request took 194.999002ms, request: PATCH:https://10.241.226.117:6443/api/v1/namespaces/default/events/af867b.14f71fc4e5082bf8
Nov 15 01:59:10 af867b kubelet[27751]: I1115 01:59:10.039470 27751 request.go:462] Throttling request took 195.012906ms, request: PATCH:https://10.241.226.117:6443/api/v1/namespaces/default/events/af867b.14f71fc4e507c284
Nov 15 01:59:10 af867b kubelet[27751]: I1115 01:59:10.239456 27751 request.go:462] Throttling request took 194.13137ms, request: PATCH:https://10.241.226.117:6443/api/v1/namespaces/default/events/af867b.14f71fc4e5080948
Nov 15 01:59:10 af867b kubelet[27751]: I1115 01:59:10.439517 27751 request.go:462] Throttling request took 194.257764ms, request: PATCH:https://10.241.226.117:6443/api/v1/namespaces/default/events/af867b.14f71fc4e5082bf8
Nov 15 01:59:10 af867b kubelet[27751]: I1115 01:59:10.523555 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 01:59:10 af867b kubelet[27751]: I1115 01:59:10.639485 27751 request.go:462] Throttling request took 191.386186ms, request: PATCH:https://10.241.226.117:6443/api/v1/namespaces/default/events/af867b.14f71fc4e507c284
Nov 15 01:59:10 af867b kubelet[27751]: I1115 01:59:10.797215 27751 eviction_manager.go:221] eviction manager: synchronize housekeeping
Nov 15 01:59:10 af867b kubelet[27751]: W1115 01:59:10.816282 27751 cni.go:196] Unable to update cni config: No networks found in /etc/cni/net.d
Nov 15 01:59:10 af867b kubelet[27751]: I1115 01:59:10.816546 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 01:59:10 af867b kubelet[27751]: E1115 01:59:10.816587 27751 kubelet.go:2095] Container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 01:59:10 af867b kubelet[27751]: I1115 01:59:10.839395 27751 request.go:462] Throttling request took 195.08274ms, request: PATCH:https://10.241.226.117:6443/api/v1/namespaces/default/events/af867b.14f71fc4e5080948
Nov 15 01:59:10 af867b kubelet[27751]: I1115 01:59:10.844679 27751 helpers.go:871] eviction manager: observations: signal=imagefs.available, available: 41099776Ki, capacity: 45Gi, time: 2017-11-15 01:59:08.748924209 +0000 UTC
Nov 15 01:59:10 af867b kubelet[27751]: I1115 01:59:10.844745 27751 helpers.go:873] eviction manager: observations: signal=allocatableMemory.available, available: 7274672Ki, capacity: 7393360Ki
Nov 15 01:59:10 af867b kubelet[27751]: I1115 01:59:10.844754 27751 helpers.go:873] eviction manager: observations: signal=allocatableNodeFs.available, available: 9624040228, capacity: 10198Mi
Nov 15 01:59:10 af867b kubelet[27751]: I1115 01:59:10.844762 27751 helpers.go:871] eviction manager: observations: signal=memory.available, available: 6666096Ki, capacity: 7393360Ki, time: 2017-11-15 01:59:08.748924209 +0000 UTC
Nov 15 01:59:10 af867b kubelet[27751]: I1115 01:59:10.844773 27751 helpers.go:871] eviction manager: observations: signal=nodefs.available, available: 7419708Ki, capacity: 10198Mi, time: 2017-11-15 01:59:08.748924209 +0000 UTC
Nov 15 01:59:10 af867b kubelet[27751]: I1115 01:59:10.844783 27751 helpers.go:871] eviction manager: observations: signal=nodefs.inodesFree, available: 10384879, capacity: 10208Ki, time: 2017-11-15 01:59:08.748924209 +0000 UTC
Nov 15 01:59:10 af867b kubelet[27751]: I1115 01:59:10.844816 27751 eviction_manager.go:325] eviction manager: no resources are starved
Nov 15 01:59:11 af867b kubelet[27751]: I1115 01:59:11.039476 27751 request.go:462] Throttling request took 188.874013ms, request: PATCH:https://10.241.226.117:6443/api/v1/namespaces/default/events/af867b.14f71fc4e5082bf8
Nov 15 01:59:11 af867b kubelet[27751]: I1115 01:59:11.239544 27751 request.go:462] Throttling request took 192.363192ms, request: PATCH:https://10.241.226.117:6443/api/v1/namespaces/default/events/af867b.14f71fc4e5082bf8
Nov 15 01:59:11 af867b kubelet[27751]: I1115 01:59:11.439482 27751 request.go:462] Throttling request took 190.719733ms, request: PATCH:https://10.241.226.117:6443/api/v1/namespaces/default/events/af867b.14f71fc4e507c284
Nov 15 01:59:11 af867b kubelet[27751]: I1115 01:59:11.639505 27751 request.go:462] Throttling request took 195.587175ms, request: PATCH:https://10.241.226.117:6443/api/v1/namespaces/default/events/af867b.14f71fc4e5080948
Nov 15 01:59:11 af867b kubelet[27751]: I1115 01:59:11.839487 27751 request.go:462] Throttling request took 193.862011ms, request: PATCH:https://10.241.226.117:6443/api/v1/namespaces/default/events/af867b.14f71fc4e507c284
Nov 15 01:59:12 af867b kubelet[27751]: I1115 01:59:12.039538 27751 request.go:462] Throttling request took 193.968694ms, request: PATCH:https://10.241.226.117:6443/api/v1/namespaces/default/events/af867b.14f71fc4e5080948
Nov 15 01:59:12 af867b kubelet[27751]: I1115 01:59:12.239498 27751 request.go:462] Throttling request took 190.282991ms, request: POST:https://10.241.226.117:6443/api/v1/namespaces/kube-system/events
Nov 15 01:59:12 af867b kubelet[27751]: I1115 01:59:12.439482 27751 request.go:462] Throttling request took 196.077185ms, request: POST:https://10.241.226.117:6443/api/v1/namespaces/kube-system/events
Nov 15 01:59:12 af867b kubelet[27751]: I1115 01:59:12.523872 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 01:59:12 af867b kubelet[27751]: I1115 01:59:12.639550 27751 request.go:462] Throttling request took 195.008881ms, request: POST:https://10.241.226.117:6443/api/v1/namespaces/kube-system/events
Nov 15 01:59:12 af867b kubelet[27751]: I1115 01:59:12.839499 27751 request.go:462] Throttling request took 193.958228ms, request: POST:https://10.241.226.117:6443/api/v1/namespaces/kube-system/events
Nov 15 01:59:13 af867b kubelet[27751]: I1115 01:59:13.039565 27751 request.go:462] Throttling request took 196.053303ms, request: POST:https://10.241.226.117:6443/api/v1/namespaces/kube-system/events
Nov 15 01:59:13 af867b kubelet[27751]: I1115 01:59:13.239565 27751 request.go:462] Throttling request took 194.031519ms, request: POST:https://10.241.226.117:6443/api/v1/namespaces/kube-system/events
Nov 15 01:59:13 af867b kubelet[27751]: I1115 01:59:13.439491 27751 request.go:462] Throttling request took 195.025944ms, request: POST:https://10.241.226.117:6443/api/v1/namespaces/kube-system/events
Nov 15 01:59:13 af867b kubelet[27751]: I1115 01:59:13.617341 27751 prober.go:160] HTTP-Probe Host: https://127.0.0.1, Port: 6443, Path: /healthz
Nov 15 01:59:13 af867b kubelet[27751]: I1115 01:59:13.617375 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 01:59:13 af867b kubelet[27751]: I1115 01:59:13.630228 27751 http.go:96] Probe succeeded for https://127.0.0.1:6443/healthz, Response: {200 OK 200 HTTP/2.0 2 0 map[Content-Type:[text/plain; charset=utf-8] Content-Length:[2] Date:[Wed, 15 Nov 2017 01:59:13 GMT]] 0xc42186b360 2 [] false false map[] 0xc420a31300 0xc42107d760}
Nov 15 01:59:13 af867b kubelet[27751]: I1115 01:59:13.630288 27751 prober.go:113] Liveness probe for "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d):kube-apiserver" succeeded
Nov 15 01:59:13 af867b kubelet[27751]: I1115 01:59:13.639469 27751 request.go:462] Throttling request took 195.415143ms, request: POST:https://10.241.226.117:6443/api/v1/namespaces/kube-system/events
Nov 15 01:59:13 af867b kubelet[27751]: I1115 01:59:13.839469 27751 request.go:462] Throttling request took 196.54814ms, request: POST:https://10.241.226.117:6443/api/v1/namespaces/kube-system/events
Nov 15 01:59:14 af867b kubelet[27751]: I1115 01:59:14.039464 27751 request.go:462] Throttling request took 197.024593ms, request: POST:https://10.241.226.117:6443/api/v1/namespaces/kube-system/events
Nov 15 01:59:14 af867b kubelet[27751]: I1115 01:59:14.239463 27751 request.go:462] Throttling request took 195.99837ms, request: POST:https://10.241.226.117:6443/api/v1/namespaces/kube-system/events
Nov 15 01:59:14 af867b kubelet[27751]: I1115 01:59:14.439455 27751 request.go:462] Throttling request took 194.821127ms, request: POST:https://10.241.226.117:6443/api/v1/namespaces/kube-system/events
Nov 15 01:59:14 af867b kubelet[27751]: I1115 01:59:14.524403 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 01:59:14 af867b kubelet[27751]: I1115 01:59:14.642803 27751 request.go:462] Throttling request took 198.765503ms, request: POST:https://10.241.226.117:6443/api/v1/namespaces/kube-system/events
Nov 15 01:59:14 af867b kubelet[27751]: I1115 01:59:14.839498 27751 request.go:462] Throttling request took 192.497221ms, request: POST:https://10.241.226.117:6443/api/v1/namespaces/kube-system/events
Nov 15 01:59:15 af867b kubelet[27751]: I1115 01:59:15.039479 27751 request.go:462] Throttling request took 196.12028ms, request: POST:https://10.241.226.117:6443/api/v1/namespaces/kube-system/events
Nov 15 01:59:15 af867b kubelet[27751]: I1115 01:59:15.240168 27751 request.go:462] Throttling request took 197.134005ms, request: POST:https://10.241.226.117:6443/api/v1/namespaces/kube-system/events
Nov 15 01:59:15 af867b kubelet[27751]: I1115 01:59:15.439452 27751 request.go:462] Throttling request took 193.069837ms, request: POST:https://10.241.226.117:6443/api/v1/namespaces/kube-system/events
Nov 15 01:59:15 af867b kubelet[27751]: I1115 01:59:15.640872 27751 request.go:462] Throttling request took 198.081111ms, request: POST:https://10.241.226.117:6443/api/v1/namespaces/kube-system/events
Nov 15 01:59:15 af867b kubelet[27751]: W1115 01:59:15.817942 27751 cni.go:196] Unable to update cni config: No networks found in /etc/cni/net.d
Nov 15 01:59:15 af867b kubelet[27751]: I1115 01:59:15.818694 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 01:59:15 af867b kubelet[27751]: E1115 01:59:15.818759 27751 kubelet.go:2095] Container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 01:59:15 af867b kubelet[27751]: I1115 01:59:15.839455 27751 request.go:462] Throttling request took 185.97892ms, request: POST:https://10.241.226.117:6443/api/v1/namespaces/kube-system/events
Nov 15 01:59:16 af867b kubelet[27751]: I1115 01:59:16.039492 27751 request.go:462] Throttling request took 195.094844ms, request: POST:https://10.241.226.117:6443/api/v1/namespaces/kube-system/events
Nov 15 01:59:16 af867b kubelet[27751]: I1115 01:59:16.239457 27751 request.go:462] Throttling request took 194.804621ms, request: POST:https://10.241.226.117:6443/api/v1/namespaces/kube-system/events
Nov 15 01:59:16 af867b kubelet[27751]: I1115 01:59:16.439400 27751 request.go:462] Throttling request took 195.383661ms, request: POST:https://10.241.226.117:6443/api/v1/namespaces/kube-system/events
Nov 15 01:59:16 af867b kubelet[27751]: I1115 01:59:16.523461 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 01:59:17 af867b kubelet[27751]: I1115 01:59:17.506213 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10252, Path: /healthz
Nov 15 01:59:17 af867b kubelet[27751]: I1115 01:59:17.506264 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 01:59:17 af867b kubelet[27751]: I1115 01:59:17.510625 27751 http.go:96] Probe succeeded for http://127.0.0.1:10252/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 01:59:17 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc420cbbe80 2 [] true false map[] 0xc420afca00 <nil>}
Nov 15 01:59:17 af867b kubelet[27751]: I1115 01:59:17.510703 27751 prober.go:113] Liveness probe for "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9):kube-controller-manager" succeeded
Nov 15 01:59:17 af867b kubelet[27751]: I1115 01:59:17.681687 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10251, Path: /healthz
Nov 15 01:59:17 af867b kubelet[27751]: I1115 01:59:17.681733 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 01:59:17 af867b kubelet[27751]: I1115 01:59:17.683669 27751 http.go:96] Probe succeeded for http://127.0.0.1:10251/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 01:59:17 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc420aa7940 2 [] true false map[] 0xc420d4a700 <nil>}
Nov 15 01:59:17 af867b kubelet[27751]: I1115 01:59:17.683713 27751 prober.go:113] Liveness probe for "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373):kube-scheduler" succeeded
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.525427 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.591992 27751 config.go:282] Setting pods for source api
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.592201 27751 config.go:404] Receiving a new pod "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.592264 27751 kubelet.go:1837] SyncLoop (ADD, "api"): "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.592389 27751 kubelet_pods.go:1284] Generating status for "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.592666 27751 qos_container_manager_linux.go:320] [ContainerManager]: Updated QoS cgroup configuration
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.600132 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.600352 27751 factory.go:116] Factory "docker" was unable to handle container "/kubepods/besteffort/pod9729c03a-c9a8-11e7-89f4-c6b053eac242"
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.600379 27751 factory.go:105] Error trying to work out if we can handle /kubepods/besteffort/pod9729c03a-c9a8-11e7-89f4-c6b053eac242: /kubepods/besteffort/pod9729c03a-c9a8-11e7-89f4-c6b053eac242 not handled by systemd handler
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.600423 27751 factory.go:116] Factory "systemd" was unable to handle container "/kubepods/besteffort/pod9729c03a-c9a8-11e7-89f4-c6b053eac242"
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.600461 27751 factory.go:112] Using factory "raw" for container "/kubepods/besteffort/pod9729c03a-c9a8-11e7-89f4-c6b053eac242"
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.600685 27751 manager.go:932] Added container: "/kubepods/besteffort/pod9729c03a-c9a8-11e7-89f4-c6b053eac242" (aliases: [], namespace: "")
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.600819 27751 handler.go:325] Added event &{/kubepods/besteffort/pod9729c03a-c9a8-11e7-89f4-c6b053eac242 2017-11-15 01:59:18.597194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.600866 27751 container.go:409] Start housekeeping for container "/kubepods/besteffort/pod9729c03a-c9a8-11e7-89f4-c6b053eac242"
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.614966 27751 status_manager.go:451] Status for pod "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)" updated successfully: (1, {Phase:Pending Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:59:18 +0000 UTC Reason: Message:} {Type:Ready Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:59:18 +0000 UTC Reason:ContainersNotReady Message:containers with unready status: [kube-proxy]}] Message: Reason: HostIP:10.196.65.210 PodIP:10.196.65.210 StartTime:2017-11-15 01:59:18 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:kube-proxy State:{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,} Running:nil Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:false RestartCount:0 Image:gcr.io/google_containers/kube-proxy-amd64:v1.8.3 ImageID: ContainerID:}] QOSClass:BestEffort})
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.615109 27751 config.go:282] Setting pods for source api
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.615370 27751 kubelet.go:1850] SyncLoop (RECONCILE, "api"): "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.776553 27751 reconciler.go:212] operationExecutor.VerifyControllerAttachedVolume started for volume "kube-proxy" (UniqueName: "kubernetes.io/configmap/9729c03a-c9a8-11e7-89f4-c6b053eac242-kube-proxy") pod "kube-proxy-nnsjf" (UID: "9729c03a-c9a8-11e7-89f4-c6b053eac242")
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.776640 27751 reconciler.go:212] operationExecutor.VerifyControllerAttachedVolume started for volume "xtables-lock" (UniqueName: "kubernetes.io/host-path/9729c03a-c9a8-11e7-89f4-c6b053eac242-xtables-lock") pod "kube-proxy-nnsjf" (UID: "9729c03a-c9a8-11e7-89f4-c6b053eac242")
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.776780 27751 reconciler.go:212] operationExecutor.VerifyControllerAttachedVolume started for volume "kube-proxy-token-gqhfs" (UniqueName: "kubernetes.io/secret/9729c03a-c9a8-11e7-89f4-c6b053eac242-kube-proxy-token-gqhfs") pod "kube-proxy-nnsjf" (UID: "9729c03a-c9a8-11e7-89f4-c6b053eac242")
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.877079 27751 reconciler.go:257] operationExecutor.MountVolume started for volume "kube-proxy" (UniqueName: "kubernetes.io/configmap/9729c03a-c9a8-11e7-89f4-c6b053eac242-kube-proxy") pod "kube-proxy-nnsjf" (UID: "9729c03a-c9a8-11e7-89f4-c6b053eac242")
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.877156 27751 reconciler.go:257] operationExecutor.MountVolume started for volume "xtables-lock" (UniqueName: "kubernetes.io/host-path/9729c03a-c9a8-11e7-89f4-c6b053eac242-xtables-lock") pod "kube-proxy-nnsjf" (UID: "9729c03a-c9a8-11e7-89f4-c6b053eac242")
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.877214 27751 reconciler.go:257] operationExecutor.MountVolume started for volume "kube-proxy-token-gqhfs" (UniqueName: "kubernetes.io/secret/9729c03a-c9a8-11e7-89f4-c6b053eac242-kube-proxy-token-gqhfs") pod "kube-proxy-nnsjf" (UID: "9729c03a-c9a8-11e7-89f4-c6b053eac242")
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.877242 27751 configmap.go:187] Setting up volume kube-proxy for pod 9729c03a-c9a8-11e7-89f4-c6b053eac242 at /var/lib/kubelet/pods/9729c03a-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~configmap/kube-proxy
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.877266 27751 secret.go:186] Setting up volume kube-proxy-token-gqhfs for pod 9729c03a-c9a8-11e7-89f4-c6b053eac242 at /var/lib/kubelet/pods/9729c03a-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/kube-proxy-token-gqhfs
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.877362 27751 operation_generator.go:484] MountVolume.SetUp succeeded for volume "xtables-lock" (UniqueName: "kubernetes.io/host-path/9729c03a-c9a8-11e7-89f4-c6b053eac242-xtables-lock") pod "kube-proxy-nnsjf" (UID: "9729c03a-c9a8-11e7-89f4-c6b053eac242")
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.877773 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-proxy-nnsjf", UID:"9729c03a-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"287", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "xtables-lock"
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.879861 27751 empty_dir.go:264] pod 9729c03a-c9a8-11e7-89f4-c6b053eac242: mounting tmpfs for volume wrapped_kube-proxy-token-gqhfs
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.879884 27751 mount_linux.go:135] Mounting cmd (systemd-run) with arguments ([--description=Kubernetes transient mount for /var/lib/kubelet/pods/9729c03a-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/kube-proxy-token-gqhfs --scope -- mount -t tmpfs tmpfs /var/lib/kubelet/pods/9729c03a-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/kube-proxy-token-gqhfs])
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.895791 27751 configmap.go:218] Received configMap kube-system/kube-proxy containing (1) pieces of data, 407 total bytes
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.895871 27751 atomic_writer.go:145] pod kube-system/kube-proxy-nnsjf volume kube-proxy: write required for target directory /var/lib/kubelet/pods/9729c03a-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~configmap/kube-proxy
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.895977 27751 atomic_writer.go:160] pod kube-system/kube-proxy-nnsjf volume kube-proxy: performed write of new data to ts data directory: /var/lib/kubelet/pods/9729c03a-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~configmap/kube-proxy/..119811_15_11_01_59_18.128125473
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.896083 27751 operation_generator.go:484] MountVolume.SetUp succeeded for volume "kube-proxy" (UniqueName: "kubernetes.io/configmap/9729c03a-c9a8-11e7-89f4-c6b053eac242-kube-proxy") pod "kube-proxy-nnsjf" (UID: "9729c03a-c9a8-11e7-89f4-c6b053eac242")
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.896121 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-proxy-nnsjf", UID:"9729c03a-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"287", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "kube-proxy"
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.900747 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/run-28837.scope"
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.900775 27751 factory.go:105] Error trying to work out if we can handle /system.slice/run-28837.scope: /system.slice/run-28837.scope not handled by systemd handler
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.900782 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/run-28837.scope"
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.900790 27751 factory.go:112] Using factory "raw" for container "/system.slice/run-28837.scope"
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.900939 27751 manager.go:932] Added container: "/system.slice/run-28837.scope" (aliases: [], namespace: "")
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.901034 27751 handler.go:325] Added event &{/system.slice/run-28837.scope 2017-11-15 01:59:18.887194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.901061 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-29c1400e918b28b9982788b4c8b07c46b1767efcd8b96000e44bc78da639994b.mount"
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.901072 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-29c1400e918b28b9982788b4c8b07c46b1767efcd8b96000e44bc78da639994b.mount", but ignoring.
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.901082 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-29c1400e918b28b9982788b4c8b07c46b1767efcd8b96000e44bc78da639994b.mount"
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.901090 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-9d97c593c8a0b1ef2442048619ba4466fad52c408e686c9285a48450f3a6669f.mount"
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.901097 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-9d97c593c8a0b1ef2442048619ba4466fad52c408e686c9285a48450f3a6669f.mount", but ignoring.
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.901107 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-9d97c593c8a0b1ef2442048619ba4466fad52c408e686c9285a48450f3a6669f.mount"
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.901115 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-faffc5ce0cf0da39c68377dff0a611eb436ee74dfce11692aecd1e723311ef80.mount"
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.901122 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-faffc5ce0cf0da39c68377dff0a611eb436ee74dfce11692aecd1e723311ef80.mount", but ignoring.
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.901130 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-faffc5ce0cf0da39c68377dff0a611eb436ee74dfce11692aecd1e723311ef80.mount"
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.901138 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-fe18d337ce42bd3a4d2aacb1349be028274681d8073fbe5d6377946e815fa810-shm.mount"
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.901147 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-fe18d337ce42bd3a4d2aacb1349be028274681d8073fbe5d6377946e815fa810-shm.mount", but ignoring.
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.901155 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-containers-fe18d337ce42bd3a4d2aacb1349be028274681d8073fbe5d6377946e815fa810-shm.mount"
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.901194 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-82408ec054eef4c0bce3519663bc7e5cbaff7a4fea1d8b7490fa3cc5a74bac60.mount"
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.901202 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-82408ec054eef4c0bce3519663bc7e5cbaff7a4fea1d8b7490fa3cc5a74bac60.mount", but ignoring.
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.901211 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-82408ec054eef4c0bce3519663bc7e5cbaff7a4fea1d8b7490fa3cc5a74bac60.mount"
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.901219 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-19b06ca14052c17f92bf03d5316fd7642fe4fee6bf193755575355be3e87a13d-shm.mount"
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.901227 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-19b06ca14052c17f92bf03d5316fd7642fe4fee6bf193755575355be3e87a13d-shm.mount", but ignoring.
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.901236 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-containers-19b06ca14052c17f92bf03d5316fd7642fe4fee6bf193755575355be3e87a13d-shm.mount"
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.901244 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-d5076ad0e9fb270d1b8c4ff7cdadbf32d6b1e30dc42ae24dfbc5cb01bb5aa934-shm.mount"
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.901251 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-d5076ad0e9fb270d1b8c4ff7cdadbf32d6b1e30dc42ae24dfbc5cb01bb5aa934-shm.mount", but ignoring.
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.901259 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-containers-d5076ad0e9fb270d1b8c4ff7cdadbf32d6b1e30dc42ae24dfbc5cb01bb5aa934-shm.mount"
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.901267 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-cd27b64716d8032aa3ea595ec121a222dc749f2f903149a780eb711106fb0c53.mount"
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.901285 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-cd27b64716d8032aa3ea595ec121a222dc749f2f903149a780eb711106fb0c53.mount", but ignoring.
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.901293 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-cd27b64716d8032aa3ea595ec121a222dc749f2f903149a780eb711106fb0c53.mount"
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.901301 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-8c430fd9b74591760bf92ca717db0989293743473d027f75b1dfded4d0661504.mount"
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.901308 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-8c430fd9b74591760bf92ca717db0989293743473d027f75b1dfded4d0661504.mount", but ignoring.
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.901317 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-8c430fd9b74591760bf92ca717db0989293743473d027f75b1dfded4d0661504.mount"
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.901325 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-439f518d46ec0d596518b7c17c503c52a85dd4da21d8a56ec174d5b6b3b98e03-shm.mount"
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.901333 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-439f518d46ec0d596518b7c17c503c52a85dd4da21d8a56ec174d5b6b3b98e03-shm.mount", but ignoring.
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.901354 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-containers-439f518d46ec0d596518b7c17c503c52a85dd4da21d8a56ec174d5b6b3b98e03-shm.mount"
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.901361 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-23d8892c53722cfb4d3aa6c5b126ad2674a7b8041b4873d30b98faefad6f5b44.mount"
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.901369 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-23d8892c53722cfb4d3aa6c5b126ad2674a7b8041b4873d30b98faefad6f5b44.mount", but ignoring.
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.901378 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-23d8892c53722cfb4d3aa6c5b126ad2674a7b8041b4873d30b98faefad6f5b44.mount"
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.901388 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-6ddc96a63fb65ccd1743152170b39274e0eb3e4ccc4948bf2c0f5a6130ee9ce7.mount"
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.901394 27751 container.go:409] Start housekeeping for container "/system.slice/run-28837.scope"
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.901396 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-6ddc96a63fb65ccd1743152170b39274e0eb3e4ccc4948bf2c0f5a6130ee9ce7.mount", but ignoring.
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.901970 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-6ddc96a63fb65ccd1743152170b39274e0eb3e4ccc4948bf2c0f5a6130ee9ce7.mount"
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.911799 27751 manager.go:989] Destroyed container: "/system.slice/run-28837.scope" (aliases: [], namespace: "")
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.911816 27751 handler.go:325] Added event &{/system.slice/run-28837.scope 2017-11-15 01:59:18.911812017 +0000 UTC containerDeletion {<nil>}}
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.913500 27751 secret.go:217] Received secret kube-system/kube-proxy-token-gqhfs containing (3) pieces of data, 1904 total bytes
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.913564 27751 atomic_writer.go:145] pod kube-system/kube-proxy-nnsjf volume kube-proxy-token-gqhfs: write required for target directory /var/lib/kubelet/pods/9729c03a-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/kube-proxy-token-gqhfs
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.913642 27751 atomic_writer.go:160] pod kube-system/kube-proxy-nnsjf volume kube-proxy-token-gqhfs: performed write of new data to ts data directory: /var/lib/kubelet/pods/9729c03a-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/kube-proxy-token-gqhfs/..119811_15_11_01_59_18.005299980
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.913748 27751 operation_generator.go:484] MountVolume.SetUp succeeded for volume "kube-proxy-token-gqhfs" (UniqueName: "kubernetes.io/secret/9729c03a-c9a8-11e7-89f4-c6b053eac242-kube-proxy-token-gqhfs") pod "kube-proxy-nnsjf" (UID: "9729c03a-c9a8-11e7-89f4-c6b053eac242")
Nov 15 01:59:18 af867b kubelet[27751]: I1115 01:59:18.913780 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-proxy-nnsjf", UID:"9729c03a-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"287", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "kube-proxy-token-gqhfs"
Nov 15 01:59:19 af867b kubelet[27751]: I1115 01:59:19.200464 27751 volume_manager.go:366] All volumes are attached and mounted for pod "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 01:59:19 af867b kubelet[27751]: I1115 01:59:19.200584 27751 kuberuntime_manager.go:370] No sandbox for pod "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)" can be found. Need to start a new one
Nov 15 01:59:19 af867b kubelet[27751]: I1115 01:59:19.200613 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:true CreateSandbox:true SandboxID: Attempt:0 NextInitContainerToStart:nil ContainersToStart:[0] ContainersToKill:map[]} for pod "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 01:59:19 af867b kubelet[27751]: I1115 01:59:19.200801 27751 kuberuntime_manager.go:565] SyncPod received new pod "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)", will create a sandbox for it
Nov 15 01:59:19 af867b kubelet[27751]: I1115 01:59:19.200832 27751 kuberuntime_manager.go:574] Stopping PodSandbox for "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)", will start new one
Nov 15 01:59:19 af867b kubelet[27751]: I1115 01:59:19.200878 27751 kuberuntime_manager.go:626] Creating sandbox for pod "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 01:59:19 af867b kubelet[27751]: I1115 01:59:19.205243 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/besteffort/pod9729c03a-c9a8-11e7-89f4-c6b053eac242"
Nov 15 01:59:19 af867b kubelet[27751]: I1115 01:59:19.205269 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/besteffort/pod9729c03a-c9a8-11e7-89f4-c6b053eac242"
Nov 15 01:59:19 af867b kubelet[27751]: I1115 01:59:19.646297 27751 factory.go:112] Using factory "docker" for container "/kubepods/besteffort/pod9729c03a-c9a8-11e7-89f4-c6b053eac242/91510886f3beb621e5d04309d502352ff78392e46405f21633802da4f7047069"
Nov 15 01:59:19 af867b kubelet[27751]: I1115 01:59:19.647984 27751 docker_sandbox.go:691] Will attempt to re-write config file /var/lib/docker/containers/91510886f3beb621e5d04309d502352ff78392e46405f21633802da4f7047069/resolv.conf with:
Nov 15 01:59:19 af867b kubelet[27751]: [nameserver 10.196.65.209 search opcwlaas.oraclecloud.internal. opcwlaas.oraclecloud.internal.]
Nov 15 01:59:19 af867b kubelet[27751]: I1115 01:59:19.648846 27751 manager.go:932] Added container: "/kubepods/besteffort/pod9729c03a-c9a8-11e7-89f4-c6b053eac242/91510886f3beb621e5d04309d502352ff78392e46405f21633802da4f7047069" (aliases: [k8s_POD_kube-proxy-nnsjf_kube-system_9729c03a-c9a8-11e7-89f4-c6b053eac242_0 91510886f3beb621e5d04309d502352ff78392e46405f21633802da4f7047069], namespace: "docker")
Nov 15 01:59:19 af867b kubelet[27751]: I1115 01:59:19.648974 27751 handler.go:325] Added event &{/kubepods/besteffort/pod9729c03a-c9a8-11e7-89f4-c6b053eac242/91510886f3beb621e5d04309d502352ff78392e46405f21633802da4f7047069 2017-11-15 01:59:19.585194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:59:19 af867b kubelet[27751]: I1115 01:59:19.649014 27751 container.go:409] Start housekeeping for container "/kubepods/besteffort/pod9729c03a-c9a8-11e7-89f4-c6b053eac242/91510886f3beb621e5d04309d502352ff78392e46405f21633802da4f7047069"
Nov 15 01:59:19 af867b kubelet[27751]: I1115 01:59:19.651424 27751 kuberuntime_manager.go:640] Created PodSandbox "91510886f3beb621e5d04309d502352ff78392e46405f21633802da4f7047069" for pod "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 01:59:19 af867b kubelet[27751]: I1115 01:59:19.655159 27751 kuberuntime_manager.go:705] Creating container &Container{Name:kube-proxy,Image:gcr.io/google_containers/kube-proxy-amd64:v1.8.3,Command:[/usr/local/bin/kube-proxy --kubeconfig=/var/lib/kube-proxy/kubeconfig.conf],Args:[],WorkingDir:,Ports:[],Env:[],Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[{kube-proxy false /var/lib/kube-proxy <nil>} {xtables-lock false /run/xtables.lock <nil>} {kube-proxy-token-gqhfs true /var/run/secrets/kubernetes.io/serviceaccount <nil>}],LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[],TerminationMessagePolicy:File,} in pod kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)
Nov 15 01:59:19 af867b kubelet[27751]: I1115 01:59:19.657471 27751 kuberuntime_container.go:100] Generating ref for container kube-proxy: &v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-proxy-nnsjf", UID:"9729c03a-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"287", FieldPath:"spec.containers{kube-proxy}"}
Nov 15 01:59:19 af867b kubelet[27751]: I1115 01:59:19.657517 27751 container_manager_linux.go:634] Calling devicePluginHandler AllocateDevices
Nov 15 01:59:19 af867b kubelet[27751]: I1115 01:59:19.657557 27751 kubelet_pods.go:123] container: kube-system/kube-proxy-nnsjf/kube-proxy podIP: "10.196.65.210" creating hosts mount: true
Nov 15 01:59:19 af867b kubelet[27751]: I1115 01:59:19.659933 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-proxy-nnsjf", UID:"9729c03a-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"287", FieldPath:"spec.containers{kube-proxy}"}): type: 'Normal' reason: 'Pulled' Container image "gcr.io/google_containers/kube-proxy-amd64:v1.8.3" already present on machine
Nov 15 01:59:19 af867b kubelet[27751]: I1115 01:59:19.662148 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/besteffort/pod9729c03a-c9a8-11e7-89f4-c6b053eac242"
Nov 15 01:59:19 af867b kubelet[27751]: I1115 01:59:19.983686 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-proxy-nnsjf", UID:"9729c03a-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"287", FieldPath:"spec.containers{kube-proxy}"}): type: 'Normal' reason: 'Created' Created container
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.158025 27751 factory.go:112] Using factory "docker" for container "/kubepods/besteffort/pod9729c03a-c9a8-11e7-89f4-c6b053eac242/7ea397ec6048e25ce044b9edba43fe0ef2ed54803c9f3516c7b6e780d2c60a34"
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.159890 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-proxy-nnsjf", UID:"9729c03a-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"287", FieldPath:"spec.containers{kube-proxy}"}): type: 'Normal' reason: 'Started' Started container
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.164369 27751 manager.go:932] Added container: "/kubepods/besteffort/pod9729c03a-c9a8-11e7-89f4-c6b053eac242/7ea397ec6048e25ce044b9edba43fe0ef2ed54803c9f3516c7b6e780d2c60a34" (aliases: [k8s_kube-proxy_kube-proxy-nnsjf_kube-system_9729c03a-c9a8-11e7-89f4-c6b053eac242_0 7ea397ec6048e25ce044b9edba43fe0ef2ed54803c9f3516c7b6e780d2c60a34], namespace: "docker")
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.164555 27751 handler.go:325] Added event &{/kubepods/besteffort/pod9729c03a-c9a8-11e7-89f4-c6b053eac242/7ea397ec6048e25ce044b9edba43fe0ef2ed54803c9f3516c7b6e780d2c60a34 2017-11-15 01:59:20.076194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.164597 27751 container.go:409] Start housekeeping for container "/kubepods/besteffort/pod9729c03a-c9a8-11e7-89f4-c6b053eac242/7ea397ec6048e25ce044b9edba43fe0ef2ed54803c9f3516c7b6e780d2c60a34"
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.458359 27751 generic.go:146] GenericPLEG: 9729c03a-c9a8-11e7-89f4-c6b053eac242/7ea397ec6048e25ce044b9edba43fe0ef2ed54803c9f3516c7b6e780d2c60a34: non-existent -> running
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.458385 27751 generic.go:146] GenericPLEG: 9729c03a-c9a8-11e7-89f4-c6b053eac242/91510886f3beb621e5d04309d502352ff78392e46405f21633802da4f7047069: non-existent -> running
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.460320 27751 kuberuntime_manager.go:833] getSandboxIDByPodUID got sandbox IDs ["91510886f3beb621e5d04309d502352ff78392e46405f21633802da4f7047069"] for pod "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.464727 27751 generic.go:345] PLEG: Write status for kube-proxy-nnsjf/kube-system: &container.PodStatus{ID:"9729c03a-c9a8-11e7-89f4-c6b053eac242", Name:"kube-proxy-nnsjf", Namespace:"kube-system", IP:"", ContainerStatuses:[]*container.ContainerStatus{(*container.ContainerStatus)(0xc4214881c0)}, SandboxStatuses:[]*runtime.PodSandboxStatus{(*runtime.PodSandboxStatus)(0xc4212b15e0)}} (err: <nil>)
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.464798 27751 kubelet.go:1871] SyncLoop (PLEG): "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)", event: &pleg.PodLifecycleEvent{ID:"9729c03a-c9a8-11e7-89f4-c6b053eac242", Type:"ContainerStarted", Data:"7ea397ec6048e25ce044b9edba43fe0ef2ed54803c9f3516c7b6e780d2c60a34"}
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.464838 27751 kubelet.go:1871] SyncLoop (PLEG): "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)", event: &pleg.PodLifecycleEvent{ID:"9729c03a-c9a8-11e7-89f4-c6b053eac242", Type:"ContainerStarted", Data:"91510886f3beb621e5d04309d502352ff78392e46405f21633802da4f7047069"}
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.464883 27751 kubelet_pods.go:1284] Generating status for "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.465135 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.473355 27751 status_manager.go:451] Status for pod "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)" updated successfully: (2, {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:59:18 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:59:20 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP:10.196.65.210 StartTime:2017-11-15 01:59:18 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:kube-proxy State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 01:59:20 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:gcr.io/google_containers/kube-proxy-amd64:v1.8.3 ImageID:docker-pullable://gcr.io/google_containers/kube-proxy-amd64@sha256:63210bc9690144d41126a646caf03a3d76ddc6d06b8bad119d468193c3e90c24 ContainerID:docker://7ea397ec6048e25ce044b9edba43fe0ef2ed54803c9f3516c7b6e780d2c60a34}] QOSClass:BestEffort})
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.473573 27751 config.go:282] Setting pods for source api
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.473907 27751 kubelet.go:1850] SyncLoop (RECONCILE, "api"): "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.484552 27751 secret.go:186] Setting up volume kube-proxy-token-gqhfs for pod 9729c03a-c9a8-11e7-89f4-c6b053eac242 at /var/lib/kubelet/pods/9729c03a-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/kube-proxy-token-gqhfs
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.485242 27751 configmap.go:187] Setting up volume kube-proxy for pod 9729c03a-c9a8-11e7-89f4-c6b053eac242 at /var/lib/kubelet/pods/9729c03a-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~configmap/kube-proxy
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.488387 27751 configmap.go:218] Received configMap kube-system/kube-proxy containing (1) pieces of data, 407 total bytes
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.488492 27751 atomic_writer.go:142] pod kube-system/kube-proxy-nnsjf volume kube-proxy: no update required for target directory /var/lib/kubelet/pods/9729c03a-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~configmap/kube-proxy
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.488584 27751 secret.go:217] Received secret kube-system/kube-proxy-token-gqhfs containing (3) pieces of data, 1904 total bytes
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.488699 27751 atomic_writer.go:142] pod kube-system/kube-proxy-nnsjf volume kube-proxy-token-gqhfs: no update required for target directory /var/lib/kubelet/pods/9729c03a-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/kube-proxy-token-gqhfs
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.523514 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.549604 27751 factory.go:116] Factory "docker" was unable to handle container "/kubepods/besteffort/pod9729c03a-c9a8-11e7-89f4-c6b053eac242/7ea397ec6048e25ce044b9edba43fe0ef2ed54803c9f3516c7b6e780d2c60a34/kube-proxy"
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.549627 27751 factory.go:105] Error trying to work out if we can handle /kubepods/besteffort/pod9729c03a-c9a8-11e7-89f4-c6b053eac242/7ea397ec6048e25ce044b9edba43fe0ef2ed54803c9f3516c7b6e780d2c60a34/kube-proxy: /kubepods/besteffort/pod9729c03a-c9a8-11e7-89f4-c6b053eac242/7ea397ec6048e25ce044b9edba43fe0ef2ed54803c9f3516c7b6e780d2c60a34/kube-proxy not handled by systemd handler
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.549634 27751 factory.go:116] Factory "systemd" was unable to handle container "/kubepods/besteffort/pod9729c03a-c9a8-11e7-89f4-c6b053eac242/7ea397ec6048e25ce044b9edba43fe0ef2ed54803c9f3516c7b6e780d2c60a34/kube-proxy"
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.549645 27751 factory.go:112] Using factory "raw" for container "/kubepods/besteffort/pod9729c03a-c9a8-11e7-89f4-c6b053eac242/7ea397ec6048e25ce044b9edba43fe0ef2ed54803c9f3516c7b6e780d2c60a34/kube-proxy"
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.549844 27751 manager.go:932] Added container: "/kubepods/besteffort/pod9729c03a-c9a8-11e7-89f4-c6b053eac242/7ea397ec6048e25ce044b9edba43fe0ef2ed54803c9f3516c7b6e780d2c60a34/kube-proxy" (aliases: [], namespace: "")
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.549964 27751 handler.go:325] Added event &{/kubepods/besteffort/pod9729c03a-c9a8-11e7-89f4-c6b053eac242/7ea397ec6048e25ce044b9edba43fe0ef2ed54803c9f3516c7b6e780d2c60a34/kube-proxy 2017-11-15 01:59:20.548194711 +0000 UTC containerCreation {<nil>}}
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.549997 27751 container.go:409] Start housekeeping for container "/kubepods/besteffort/pod9729c03a-c9a8-11e7-89f4-c6b053eac242/7ea397ec6048e25ce044b9edba43fe0ef2ed54803c9f3516c7b6e780d2c60a34/kube-proxy"
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.765395 27751 volume_manager.go:366] All volumes are attached and mounted for pod "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.765536 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:false CreateSandbox:false SandboxID:91510886f3beb621e5d04309d502352ff78392e46405f21633802da4f7047069 Attempt:0 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 01:59:20 af867b kubelet[27751]: W1115 01:59:20.819991 27751 cni.go:196] Unable to update cni config: No networks found in /etc/cni/net.d
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.820550 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 01:59:20 af867b kubelet[27751]: E1115 01:59:20.820579 27751 kubelet.go:2095] Container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.844977 27751 eviction_manager.go:221] eviction manager: synchronize housekeeping
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.898283 27751 helpers.go:871] eviction manager: observations: signal=imagefs.available, available: 41099776Ki, capacity: 45Gi, time: 2017-11-15 01:59:08.748924209 +0000 UTC
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.898344 27751 helpers.go:873] eviction manager: observations: signal=allocatableMemory.available, available: 7095676Ki, capacity: 7393360Ki
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.898357 27751 helpers.go:873] eviction manager: observations: signal=allocatableNodeFs.available, available: 9624040228, capacity: 10198Mi
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.898377 27751 helpers.go:871] eviction manager: observations: signal=memory.available, available: 6666096Ki, capacity: 7393360Ki, time: 2017-11-15 01:59:08.748924209 +0000 UTC
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.898393 27751 helpers.go:871] eviction manager: observations: signal=nodefs.available, available: 7419708Ki, capacity: 10198Mi, time: 2017-11-15 01:59:08.748924209 +0000 UTC
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.898406 27751 helpers.go:871] eviction manager: observations: signal=nodefs.inodesFree, available: 10384879, capacity: 10208Ki, time: 2017-11-15 01:59:08.748924209 +0000 UTC
Nov 15 01:59:20 af867b kubelet[27751]: I1115 01:59:20.898438 27751 eviction_manager.go:325] eviction manager: no resources are starved
Nov 15 01:59:21 af867b kubelet[27751]: I1115 01:59:21.471105 27751 kubelet_pods.go:1284] Generating status for "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 01:59:21 af867b kubelet[27751]: I1115 01:59:21.471583 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 01:59:21 af867b kubelet[27751]: I1115 01:59:21.484123 27751 status_manager.go:451] Status for pod "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)" updated successfully: (3, {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:59:18 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:59:20 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:59:20 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP:10.196.65.210 StartTime:2017-11-15 01:59:18 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:kube-proxy State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 01:59:20 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:gcr.io/google_containers/kube-proxy-amd64:v1.8.3 ImageID:docker-pullable://gcr.io/google_containers/kube-proxy-amd64@sha256:63210bc9690144d41126a646caf03a3d76ddc6d06b8bad119d468193c3e90c24 ContainerID:docker://7ea397ec6048e25ce044b9edba43fe0ef2ed54803c9f3516c7b6e780d2c60a34}] QOSClass:BestEffort})
Nov 15 01:59:21 af867b kubelet[27751]: I1115 01:59:21.484525 27751 config.go:282] Setting pods for source api
Nov 15 01:59:21 af867b kubelet[27751]: I1115 01:59:21.487641 27751 kubelet.go:1850] SyncLoop (RECONCILE, "api"): "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 01:59:21 af867b kubelet[27751]: I1115 01:59:21.489760 27751 configmap.go:187] Setting up volume kube-proxy for pod 9729c03a-c9a8-11e7-89f4-c6b053eac242 at /var/lib/kubelet/pods/9729c03a-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~configmap/kube-proxy
Nov 15 01:59:21 af867b kubelet[27751]: I1115 01:59:21.489962 27751 secret.go:186] Setting up volume kube-proxy-token-gqhfs for pod 9729c03a-c9a8-11e7-89f4-c6b053eac242 at /var/lib/kubelet/pods/9729c03a-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/kube-proxy-token-gqhfs
Nov 15 01:59:21 af867b kubelet[27751]: I1115 01:59:21.492438 27751 configmap.go:218] Received configMap kube-system/kube-proxy containing (1) pieces of data, 407 total bytes
Nov 15 01:59:21 af867b kubelet[27751]: I1115 01:59:21.492539 27751 secret.go:217] Received secret kube-system/kube-proxy-token-gqhfs containing (3) pieces of data, 1904 total bytes
Nov 15 01:59:21 af867b kubelet[27751]: I1115 01:59:21.492737 27751 atomic_writer.go:142] pod kube-system/kube-proxy-nnsjf volume kube-proxy-token-gqhfs: no update required for target directory /var/lib/kubelet/pods/9729c03a-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/kube-proxy-token-gqhfs
Nov 15 01:59:21 af867b kubelet[27751]: I1115 01:59:21.492548 27751 atomic_writer.go:142] pod kube-system/kube-proxy-nnsjf volume kube-proxy: no update required for target directory /var/lib/kubelet/pods/9729c03a-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~configmap/kube-proxy
Nov 15 01:59:21 af867b kubelet[27751]: I1115 01:59:21.772037 27751 volume_manager.go:366] All volumes are attached and mounted for pod "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 01:59:21 af867b kubelet[27751]: I1115 01:59:21.772223 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:false CreateSandbox:false SandboxID:91510886f3beb621e5d04309d502352ff78392e46405f21633802da4f7047069 Attempt:0 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 01:59:22 af867b kubelet[27751]: I1115 01:59:22.523519 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 01:59:22 af867b kubelet[27751]: I1115 01:59:22.960004 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 2379, Path: /health
Nov 15 01:59:22 af867b kubelet[27751]: I1115 01:59:22.960066 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 01:59:23 af867b kubelet[27751]: I1115 01:59:23.461824 27751 http.go:96] Probe succeeded for http://127.0.0.1:2379/health, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 01:59:23 GMT] Content-Length:[18] Content-Type:[text/plain; charset=utf-8]] 0xc421528c60 18 [] true false map[] 0xc421179500 <nil>}
Nov 15 01:59:23 af867b kubelet[27751]: I1115 01:59:23.461900 27751 prober.go:113] Liveness probe for "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250):etcd" succeeded
Nov 15 01:59:23 af867b kubelet[27751]: I1115 01:59:23.617422 27751 prober.go:160] HTTP-Probe Host: https://127.0.0.1, Port: 6443, Path: /healthz
Nov 15 01:59:23 af867b kubelet[27751]: I1115 01:59:23.617473 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 01:59:23 af867b kubelet[27751]: I1115 01:59:23.625211 27751 http.go:96] Probe succeeded for https://127.0.0.1:6443/healthz, Response: {200 OK 200 HTTP/2.0 2 0 map[Content-Type:[text/plain; charset=utf-8] Content-Length:[2] Date:[Wed, 15 Nov 2017 01:59:23 GMT]] 0xc420f3e640 2 [] false false map[] 0xc421179700 0xc42179cb00}
Nov 15 01:59:23 af867b kubelet[27751]: I1115 01:59:23.625301 27751 prober.go:113] Liveness probe for "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d):kube-apiserver" succeeded
Nov 15 01:59:24 af867b kubelet[27751]: I1115 01:59:24.523463 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 01:59:25 af867b kubelet[27751]: W1115 01:59:25.821983 27751 cni.go:196] Unable to update cni config: No networks found in /etc/cni/net.d
Nov 15 01:59:25 af867b kubelet[27751]: I1115 01:59:25.823149 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 01:59:25 af867b kubelet[27751]: E1115 01:59:25.823186 27751 kubelet.go:2095] Container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 01:59:26 af867b kubelet[27751]: I1115 01:59:26.523545 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 01:59:27 af867b kubelet[27751]: I1115 01:59:27.506003 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10252, Path: /healthz
Nov 15 01:59:27 af867b kubelet[27751]: I1115 01:59:27.506027 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 01:59:27 af867b kubelet[27751]: I1115 01:59:27.508388 27751 http.go:96] Probe succeeded for http://127.0.0.1:10252/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 01:59:27 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc4201a0500 2 [] true false map[] 0xc420d4a300 <nil>}
Nov 15 01:59:27 af867b kubelet[27751]: I1115 01:59:27.508425 27751 prober.go:113] Liveness probe for "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9):kube-controller-manager" succeeded
Nov 15 01:59:27 af867b kubelet[27751]: I1115 01:59:27.681677 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10251, Path: /healthz
Nov 15 01:59:27 af867b kubelet[27751]: I1115 01:59:27.681746 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 01:59:27 af867b kubelet[27751]: I1115 01:59:27.683459 27751 http.go:96] Probe succeeded for http://127.0.0.1:10251/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 01:59:27 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc4201a0a20 2 [] true false map[] 0xc420d4a400 <nil>}
Nov 15 01:59:27 af867b kubelet[27751]: I1115 01:59:27.683501 27751 prober.go:113] Liveness probe for "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373):kube-scheduler" succeeded
Nov 15 01:59:28 af867b kubelet[27751]: I1115 01:59:28.523440 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 01:59:30 af867b kubelet[27751]: I1115 01:59:30.523527 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 01:59:30 af867b kubelet[27751]: W1115 01:59:30.824684 27751 cni.go:196] Unable to update cni config: No networks found in /etc/cni/net.d
Nov 15 01:59:30 af867b kubelet[27751]: I1115 01:59:30.825042 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 01:59:30 af867b kubelet[27751]: E1115 01:59:30.825065 27751 kubelet.go:2095] Container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 01:59:30 af867b kubelet[27751]: I1115 01:59:30.898653 27751 eviction_manager.go:221] eviction manager: synchronize housekeeping
Nov 15 01:59:30 af867b kubelet[27751]: I1115 01:59:30.967184 27751 helpers.go:871] eviction manager: observations: signal=nodefs.available, available: 7419108Ki, capacity: 10198Mi, time: 2017-11-15 01:59:24.804361047 +0000 UTC
Nov 15 01:59:30 af867b kubelet[27751]: I1115 01:59:30.967236 27751 helpers.go:871] eviction manager: observations: signal=nodefs.inodesFree, available: 10384825, capacity: 10208Ki, time: 2017-11-15 01:59:24.804361047 +0000 UTC
Nov 15 01:59:30 af867b kubelet[27751]: I1115 01:59:30.967248 27751 helpers.go:871] eviction manager: observations: signal=imagefs.available, available: 41082368Ki, capacity: 45Gi, time: 2017-11-15 01:59:24.804361047 +0000 UTC
Nov 15 01:59:30 af867b kubelet[27751]: I1115 01:59:30.967257 27751 helpers.go:873] eviction manager: observations: signal=allocatableMemory.available, available: 7047332Ki, capacity: 7393360Ki
Nov 15 01:59:30 af867b kubelet[27751]: I1115 01:59:30.967265 27751 helpers.go:873] eviction manager: observations: signal=allocatableNodeFs.available, available: 9624040228, capacity: 10198Mi
Nov 15 01:59:30 af867b kubelet[27751]: I1115 01:59:30.967273 27751 helpers.go:871] eviction manager: observations: signal=memory.available, available: 6625680Ki, capacity: 7393360Ki, time: 2017-11-15 01:59:24.804361047 +0000 UTC
Nov 15 01:59:30 af867b kubelet[27751]: I1115 01:59:30.967316 27751 eviction_manager.go:325] eviction manager: no resources are starved
Nov 15 01:59:32 af867b kubelet[27751]: I1115 01:59:32.523973 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 01:59:32 af867b kubelet[27751]: I1115 01:59:32.959962 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 2379, Path: /health
Nov 15 01:59:32 af867b kubelet[27751]: I1115 01:59:32.960024 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 01:59:33 af867b kubelet[27751]: I1115 01:59:33.462606 27751 http.go:96] Probe succeeded for http://127.0.0.1:2379/health, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 01:59:33 GMT] Content-Length:[18] Content-Type:[text/plain; charset=utf-8]] 0xc420b4d540 18 [] true false map[] 0xc420d4a100 <nil>}
Nov 15 01:59:33 af867b kubelet[27751]: I1115 01:59:33.462671 27751 prober.go:113] Liveness probe for "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250):etcd" succeeded
Nov 15 01:59:33 af867b kubelet[27751]: I1115 01:59:33.617335 27751 prober.go:160] HTTP-Probe Host: https://127.0.0.1, Port: 6443, Path: /healthz
Nov 15 01:59:33 af867b kubelet[27751]: I1115 01:59:33.617370 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 01:59:33 af867b kubelet[27751]: I1115 01:59:33.623899 27751 http.go:96] Probe succeeded for https://127.0.0.1:6443/healthz, Response: {200 OK 200 HTTP/2.0 2 0 map[Content-Type:[text/plain; charset=utf-8] Content-Length:[2] Date:[Wed, 15 Nov 2017 01:59:33 GMT]] 0xc42017f960 2 [] false false map[] 0xc4200dcd00 0xc4212f4370}
Nov 15 01:59:33 af867b kubelet[27751]: I1115 01:59:33.623958 27751 prober.go:113] Liveness probe for "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d):kube-apiserver" succeeded
Nov 15 01:59:34 af867b kubelet[27751]: I1115 01:59:34.523457 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 01:59:35 af867b kubelet[27751]: W1115 01:59:35.826324 27751 cni.go:196] Unable to update cni config: No networks found in /etc/cni/net.d
Nov 15 01:59:35 af867b kubelet[27751]: I1115 01:59:35.826482 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 01:59:35 af867b kubelet[27751]: E1115 01:59:35.826509 27751 kubelet.go:2095] Container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 01:59:36 af867b kubelet[27751]: I1115 01:59:36.523471 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 01:59:37 af867b kubelet[27751]: I1115 01:59:37.506112 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10252, Path: /healthz
Nov 15 01:59:37 af867b kubelet[27751]: I1115 01:59:37.506150 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 01:59:37 af867b kubelet[27751]: I1115 01:59:37.507596 27751 http.go:96] Probe succeeded for http://127.0.0.1:10252/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 01:59:37 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc4214c6420 2 [] true false map[] 0xc420431b00 <nil>}
Nov 15 01:59:37 af867b kubelet[27751]: I1115 01:59:37.507642 27751 prober.go:113] Liveness probe for "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9):kube-controller-manager" succeeded
Nov 15 01:59:37 af867b kubelet[27751]: I1115 01:59:37.681696 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10251, Path: /healthz
Nov 15 01:59:37 af867b kubelet[27751]: I1115 01:59:37.681770 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 01:59:37 af867b kubelet[27751]: I1115 01:59:37.683183 27751 http.go:96] Probe succeeded for http://127.0.0.1:10251/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 01:59:37 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc4215283e0 2 [] true false map[] 0xc420ee2100 <nil>}
Nov 15 01:59:37 af867b kubelet[27751]: I1115 01:59:37.683236 27751 prober.go:113] Liveness probe for "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373):kube-scheduler" succeeded
Nov 15 01:59:38 af867b kubelet[27751]: I1115 01:59:38.523457 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 01:59:40 af867b kubelet[27751]: I1115 01:59:40.523523 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 01:59:40 af867b kubelet[27751]: W1115 01:59:40.827678 27751 cni.go:196] Unable to update cni config: No networks found in /etc/cni/net.d
Nov 15 01:59:40 af867b kubelet[27751]: I1115 01:59:40.827859 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 01:59:40 af867b kubelet[27751]: E1115 01:59:40.827884 27751 kubelet.go:2095] Container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 01:59:40 af867b kubelet[27751]: I1115 01:59:40.967493 27751 eviction_manager.go:221] eviction manager: synchronize housekeeping
Nov 15 01:59:41 af867b kubelet[27751]: I1115 01:59:41.011291 27751 helpers.go:871] eviction manager: observations: signal=memory.available, available: 6621280Ki, capacity: 7393360Ki, time: 2017-11-15 01:59:36.700965635 +0000 UTC
Nov 15 01:59:41 af867b kubelet[27751]: I1115 01:59:41.011334 27751 helpers.go:871] eviction manager: observations: signal=nodefs.available, available: 7419912Ki, capacity: 10198Mi, time: 2017-11-15 01:59:36.700965635 +0000 UTC
Nov 15 01:59:41 af867b kubelet[27751]: I1115 01:59:41.011345 27751 helpers.go:871] eviction manager: observations: signal=nodefs.inodesFree, available: 10384825, capacity: 10208Ki, time: 2017-11-15 01:59:36.700965635 +0000 UTC
Nov 15 01:59:41 af867b kubelet[27751]: I1115 01:59:41.011354 27751 helpers.go:871] eviction manager: observations: signal=imagefs.available, available: 41082368Ki, capacity: 45Gi, time: 2017-11-15 01:59:36.700965635 +0000 UTC
Nov 15 01:59:41 af867b kubelet[27751]: I1115 01:59:41.011363 27751 helpers.go:873] eviction manager: observations: signal=allocatableMemory.available, available: 7044832Ki, capacity: 7393360Ki
Nov 15 01:59:41 af867b kubelet[27751]: I1115 01:59:41.011369 27751 helpers.go:873] eviction manager: observations: signal=allocatableNodeFs.available, available: 9624040228, capacity: 10198Mi
Nov 15 01:59:41 af867b kubelet[27751]: I1115 01:59:41.011389 27751 eviction_manager.go:325] eviction manager: no resources are starved
Nov 15 01:59:42 af867b kubelet[27751]: I1115 01:59:42.523518 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 01:59:42 af867b kubelet[27751]: I1115 01:59:42.959945 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 2379, Path: /health
Nov 15 01:59:42 af867b kubelet[27751]: I1115 01:59:42.960001 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 01:59:43 af867b kubelet[27751]: I1115 01:59:43.461406 27751 http.go:96] Probe succeeded for http://127.0.0.1:2379/health, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 01:59:43 GMT] Content-Length:[18] Content-Type:[text/plain; charset=utf-8]] 0xc421438600 18 [] true false map[] 0xc420afd600 <nil>}
Nov 15 01:59:43 af867b kubelet[27751]: I1115 01:59:43.461479 27751 prober.go:113] Liveness probe for "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250):etcd" succeeded
Nov 15 01:59:43 af867b kubelet[27751]: I1115 01:59:43.617343 27751 prober.go:160] HTTP-Probe Host: https://127.0.0.1, Port: 6443, Path: /healthz
Nov 15 01:59:43 af867b kubelet[27751]: I1115 01:59:43.617382 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 01:59:43 af867b kubelet[27751]: I1115 01:59:43.624651 27751 http.go:96] Probe succeeded for https://127.0.0.1:6443/healthz, Response: {200 OK 200 HTTP/2.0 2 0 map[Content-Type:[text/plain; charset=utf-8] Content-Length:[2] Date:[Wed, 15 Nov 2017 01:59:43 GMT]] 0xc4212e1420 2 [] false false map[] 0xc420a30700 0xc421002840}
Nov 15 01:59:43 af867b kubelet[27751]: I1115 01:59:43.624709 27751 prober.go:113] Liveness probe for "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d):kube-apiserver" succeeded
Nov 15 01:59:44 af867b kubelet[27751]: I1115 01:59:44.523440 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 01:59:45 af867b kubelet[27751]: W1115 01:59:45.829608 27751 cni.go:196] Unable to update cni config: No networks found in /etc/cni/net.d
Nov 15 01:59:45 af867b kubelet[27751]: I1115 01:59:45.829816 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 01:59:45 af867b kubelet[27751]: E1115 01:59:45.829840 27751 kubelet.go:2095] Container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 01:59:46 af867b kubelet[27751]: I1115 01:59:46.523452 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 01:59:47 af867b kubelet[27751]: I1115 01:59:47.506092 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10252, Path: /healthz
Nov 15 01:59:47 af867b kubelet[27751]: I1115 01:59:47.506129 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 01:59:47 af867b kubelet[27751]: I1115 01:59:47.507009 27751 http.go:96] Probe succeeded for http://127.0.0.1:10252/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Length:[2] Content-Type:[text/plain; charset=utf-8] Date:[Wed, 15 Nov 2017 01:59:47 GMT]] 0xc420b4df60 2 [] true false map[] 0xc42110a600 <nil>}
Nov 15 01:59:47 af867b kubelet[27751]: I1115 01:59:47.507043 27751 prober.go:113] Liveness probe for "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9):kube-controller-manager" succeeded
Nov 15 01:59:47 af867b kubelet[27751]: I1115 01:59:47.681637 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10251, Path: /healthz
Nov 15 01:59:47 af867b kubelet[27751]: I1115 01:59:47.681683 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 01:59:47 af867b kubelet[27751]: I1115 01:59:47.682860 27751 http.go:96] Probe succeeded for http://127.0.0.1:10251/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 01:59:47 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc421211aa0 2 [] true false map[] 0xc421178c00 <nil>}
Nov 15 01:59:47 af867b kubelet[27751]: I1115 01:59:47.682904 27751 prober.go:113] Liveness probe for "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373):kube-scheduler" succeeded
Nov 15 01:59:48 af867b kubelet[27751]: I1115 01:59:48.523444 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.523525 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.525879 27751 qos_container_manager_linux.go:320] [ContainerManager]: Updated QoS cgroup configuration
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.542147 27751 kubelet.go:1222] Container garbage collection succeeded
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823071 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/run-user-1000.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823117 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/run-user-1000.mount", but ignoring.
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823127 27751 manager.go:901] ignoring container "/system.slice/run-user-1000.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823134 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/sys-kernel-debug.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823140 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/sys-kernel-debug.mount", but ignoring.
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823147 27751 manager.go:901] ignoring container "/system.slice/sys-kernel-debug.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823153 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-19b06ca14052c17f92bf03d5316fd7642fe4fee6bf193755575355be3e87a13d-shm.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823161 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-19b06ca14052c17f92bf03d5316fd7642fe4fee6bf193755575355be3e87a13d-shm.mount", but ignoring.
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823171 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-containers-19b06ca14052c17f92bf03d5316fd7642fe4fee6bf193755575355be3e87a13d-shm.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823179 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/dev-hugepages.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823185 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/dev-hugepages.mount", but ignoring.
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823192 27751 manager.go:901] ignoring container "/system.slice/dev-hugepages.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823197 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-23d8892c53722cfb4d3aa6c5b126ad2674a7b8041b4873d30b98faefad6f5b44.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823205 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-23d8892c53722cfb4d3aa6c5b126ad2674a7b8041b4873d30b98faefad6f5b44.mount", but ignoring.
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823213 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-23d8892c53722cfb4d3aa6c5b126ad2674a7b8041b4873d30b98faefad6f5b44.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823220 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/-.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823226 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/-.mount", but ignoring.
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823233 27751 manager.go:901] ignoring container "/system.slice/-.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823237 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/dev-mqueue.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823243 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/dev-mqueue.mount", but ignoring.
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823250 27751 manager.go:901] ignoring container "/system.slice/dev-mqueue.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823255 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-faffc5ce0cf0da39c68377dff0a611eb436ee74dfce11692aecd1e723311ef80.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823263 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-faffc5ce0cf0da39c68377dff0a611eb436ee74dfce11692aecd1e723311ef80.mount", but ignoring.
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823272 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-faffc5ce0cf0da39c68377dff0a611eb436ee74dfce11692aecd1e723311ef80.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823279 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-d5076ad0e9fb270d1b8c4ff7cdadbf32d6b1e30dc42ae24dfbc5cb01bb5aa934-shm.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823286 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-d5076ad0e9fb270d1b8c4ff7cdadbf32d6b1e30dc42ae24dfbc5cb01bb5aa934-shm.mount", but ignoring.
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823294 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-containers-d5076ad0e9fb270d1b8c4ff7cdadbf32d6b1e30dc42ae24dfbc5cb01bb5aa934-shm.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823301 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-6ddc96a63fb65ccd1743152170b39274e0eb3e4ccc4948bf2c0f5a6130ee9ce7.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823308 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-6ddc96a63fb65ccd1743152170b39274e0eb3e4ccc4948bf2c0f5a6130ee9ce7.mount", but ignoring.
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823316 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-6ddc96a63fb65ccd1743152170b39274e0eb3e4ccc4948bf2c0f5a6130ee9ce7.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823323 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/proc-xen.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823329 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/proc-xen.mount", but ignoring.
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823335 27751 manager.go:901] ignoring container "/system.slice/proc-xen.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823340 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/run-docker-netns-default.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823346 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/run-docker-netns-default.mount", but ignoring.
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823353 27751 manager.go:901] ignoring container "/system.slice/run-docker-netns-default.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823358 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-9d97c593c8a0b1ef2442048619ba4466fad52c408e686c9285a48450f3a6669f.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823365 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-9d97c593c8a0b1ef2442048619ba4466fad52c408e686c9285a48450f3a6669f.mount", but ignoring.
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823374 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-9d97c593c8a0b1ef2442048619ba4466fad52c408e686c9285a48450f3a6669f.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823381 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/boot.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823386 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/boot.mount", but ignoring.
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823393 27751 manager.go:901] ignoring container "/system.slice/boot.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823397 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/sys-kernel-config.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823404 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/sys-kernel-config.mount", but ignoring.
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823411 27751 manager.go:901] ignoring container "/system.slice/sys-kernel-config.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823416 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-fe18d337ce42bd3a4d2aacb1349be028274681d8073fbe5d6377946e815fa810-shm.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823423 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-fe18d337ce42bd3a4d2aacb1349be028274681d8073fbe5d6377946e815fa810-shm.mount", but ignoring.
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823431 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-containers-fe18d337ce42bd3a4d2aacb1349be028274681d8073fbe5d6377946e815fa810-shm.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823438 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-8c430fd9b74591760bf92ca717db0989293743473d027f75b1dfded4d0661504.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823445 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-8c430fd9b74591760bf92ca717db0989293743473d027f75b1dfded4d0661504.mount", but ignoring.
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823454 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-8c430fd9b74591760bf92ca717db0989293743473d027f75b1dfded4d0661504.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823461 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823468 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper.mount", but ignoring.
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823475 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823480 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-439f518d46ec0d596518b7c17c503c52a85dd4da21d8a56ec174d5b6b3b98e03-shm.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823487 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-439f518d46ec0d596518b7c17c503c52a85dd4da21d8a56ec174d5b6b3b98e03-shm.mount", but ignoring.
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823496 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-containers-439f518d46ec0d596518b7c17c503c52a85dd4da21d8a56ec174d5b6b3b98e03-shm.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823503 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/u01-applicationSpace.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823509 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/u01-applicationSpace.mount", but ignoring.
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823516 27751 manager.go:901] ignoring container "/system.slice/u01-applicationSpace.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823521 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-29c1400e918b28b9982788b4c8b07c46b1767efcd8b96000e44bc78da639994b.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823528 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-29c1400e918b28b9982788b4c8b07c46b1767efcd8b96000e44bc78da639994b.mount", but ignoring.
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823538 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-29c1400e918b28b9982788b4c8b07c46b1767efcd8b96000e44bc78da639994b.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823545 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-82408ec054eef4c0bce3519663bc7e5cbaff7a4fea1d8b7490fa3cc5a74bac60.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823552 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-82408ec054eef4c0bce3519663bc7e5cbaff7a4fea1d8b7490fa3cc5a74bac60.mount", but ignoring.
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823561 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-82408ec054eef4c0bce3519663bc7e5cbaff7a4fea1d8b7490fa3cc5a74bac60.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823569 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-cd27b64716d8032aa3ea595ec121a222dc749f2f903149a780eb711106fb0c53.mount"
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823576 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-cd27b64716d8032aa3ea595ec121a222dc749f2f903149a780eb711106fb0c53.mount", but ignoring.
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.823584 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-cd27b64716d8032aa3ea595ec121a222dc749f2f903149a780eb711106fb0c53.mount"
Nov 15 01:59:50 af867b kubelet[27751]: W1115 01:59:50.831733 27751 cni.go:196] Unable to update cni config: No networks found in /etc/cni/net.d
Nov 15 01:59:50 af867b kubelet[27751]: I1115 01:59:50.831896 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 01:59:50 af867b kubelet[27751]: E1115 01:59:50.831922 27751 kubelet.go:2095] Container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 01:59:51 af867b kubelet[27751]: I1115 01:59:51.011664 27751 eviction_manager.go:221] eviction manager: synchronize housekeeping
Nov 15 01:59:51 af867b kubelet[27751]: I1115 01:59:51.074753 27751 helpers.go:871] eviction manager: observations: signal=imagefs.available, available: 41082368Ki, capacity: 45Gi, time: 2017-11-15 01:59:36.700965635 +0000 UTC
Nov 15 01:59:51 af867b kubelet[27751]: I1115 01:59:51.074808 27751 helpers.go:873] eviction manager: observations: signal=allocatableMemory.available, available: 7042836Ki, capacity: 7393360Ki
Nov 15 01:59:51 af867b kubelet[27751]: I1115 01:59:51.074817 27751 helpers.go:873] eviction manager: observations: signal=allocatableNodeFs.available, available: 9624040228, capacity: 10198Mi
Nov 15 01:59:51 af867b kubelet[27751]: I1115 01:59:51.074826 27751 helpers.go:871] eviction manager: observations: signal=memory.available, available: 6621280Ki, capacity: 7393360Ki, time: 2017-11-15 01:59:36.700965635 +0000 UTC
Nov 15 01:59:51 af867b kubelet[27751]: I1115 01:59:51.074836 27751 helpers.go:871] eviction manager: observations: signal=nodefs.available, available: 7419912Ki, capacity: 10198Mi, time: 2017-11-15 01:59:36.700965635 +0000 UTC
Nov 15 01:59:51 af867b kubelet[27751]: I1115 01:59:51.074848 27751 helpers.go:871] eviction manager: observations: signal=nodefs.inodesFree, available: 10384825, capacity: 10208Ki, time: 2017-11-15 01:59:36.700965635 +0000 UTC
Nov 15 01:59:51 af867b kubelet[27751]: I1115 01:59:51.074875 27751 eviction_manager.go:325] eviction manager: no resources are starved
Nov 15 01:59:52 af867b kubelet[27751]: I1115 01:59:52.523554 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 01:59:52 af867b kubelet[27751]: I1115 01:59:52.960029 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 2379, Path: /health
Nov 15 01:59:52 af867b kubelet[27751]: I1115 01:59:52.960085 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 01:59:53 af867b kubelet[27751]: I1115 01:59:53.461958 27751 http.go:96] Probe succeeded for http://127.0.0.1:2379/health, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 01:59:53 GMT] Content-Length:[18] Content-Type:[text/plain; charset=utf-8]] 0xc420db2200 18 [] true false map[] 0xc420c90100 <nil>}
Nov 15 01:59:53 af867b kubelet[27751]: I1115 01:59:53.462087 27751 prober.go:113] Liveness probe for "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250):etcd" succeeded
Nov 15 01:59:53 af867b kubelet[27751]: I1115 01:59:53.617367 27751 prober.go:160] HTTP-Probe Host: https://127.0.0.1, Port: 6443, Path: /healthz
Nov 15 01:59:53 af867b kubelet[27751]: I1115 01:59:53.617423 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 01:59:53 af867b kubelet[27751]: I1115 01:59:53.626431 27751 http.go:96] Probe succeeded for https://127.0.0.1:6443/healthz, Response: {200 OK 200 HTTP/2.0 2 0 map[Content-Type:[text/plain; charset=utf-8] Content-Length:[2] Date:[Wed, 15 Nov 2017 01:59:53 GMT]] 0xc42103a6e0 2 [] false false map[] 0xc420430a00 0xc420e78c60}
Nov 15 01:59:53 af867b kubelet[27751]: I1115 01:59:53.626526 27751 prober.go:113] Liveness probe for "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d):kube-apiserver" succeeded
Nov 15 01:59:54 af867b kubelet[27751]: I1115 01:59:54.523466 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 01:59:55 af867b kubelet[27751]: W1115 01:59:55.833311 27751 cni.go:196] Unable to update cni config: No networks found in /etc/cni/net.d
Nov 15 01:59:55 af867b kubelet[27751]: I1115 01:59:55.833480 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 01:59:55 af867b kubelet[27751]: E1115 01:59:55.833512 27751 kubelet.go:2095] Container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 01:59:56 af867b kubelet[27751]: I1115 01:59:56.523454 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 01:59:57 af867b kubelet[27751]: I1115 01:59:57.506158 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10252, Path: /healthz
Nov 15 01:59:57 af867b kubelet[27751]: I1115 01:59:57.506206 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 01:59:57 af867b kubelet[27751]: I1115 01:59:57.507135 27751 http.go:96] Probe succeeded for http://127.0.0.1:10252/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 01:59:57 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc4206ab160 2 [] true false map[] 0xc420a30300 <nil>}
Nov 15 01:59:57 af867b kubelet[27751]: I1115 01:59:57.507172 27751 prober.go:113] Liveness probe for "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9):kube-controller-manager" succeeded
Nov 15 01:59:57 af867b kubelet[27751]: I1115 01:59:57.681741 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10251, Path: /healthz
Nov 15 01:59:57 af867b kubelet[27751]: I1115 01:59:57.681787 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 01:59:57 af867b kubelet[27751]: I1115 01:59:57.683527 27751 http.go:96] Probe succeeded for http://127.0.0.1:10251/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 01:59:57 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc420c98b80 2 [] true false map[] 0xc420a30700 <nil>}
Nov 15 01:59:57 af867b kubelet[27751]: I1115 01:59:57.683586 27751 prober.go:113] Liveness probe for "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373):kube-scheduler" succeeded
Nov 15 01:59:58 af867b kubelet[27751]: I1115 01:59:58.523483 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:00:00 af867b kubelet[27751]: I1115 02:00:00.523681 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:00:00 af867b kubelet[27751]: W1115 02:00:00.834797 27751 cni.go:196] Unable to update cni config: No networks found in /etc/cni/net.d
Nov 15 02:00:00 af867b kubelet[27751]: I1115 02:00:00.835770 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 02:00:00 af867b kubelet[27751]: E1115 02:00:00.835806 27751 kubelet.go:2095] Container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 02:00:01 af867b kubelet[27751]: I1115 02:00:01.075087 27751 eviction_manager.go:221] eviction manager: synchronize housekeeping
Nov 15 02:00:01 af867b kubelet[27751]: I1115 02:00:01.120313 27751 helpers.go:871] eviction manager: observations: signal=memory.available, available: 6617608Ki, capacity: 7393360Ki, time: 2017-11-15 01:59:52.517760792 +0000 UTC
Nov 15 02:00:01 af867b kubelet[27751]: I1115 02:00:01.120356 27751 helpers.go:871] eviction manager: observations: signal=nodefs.available, available: 7419916Ki, capacity: 10198Mi, time: 2017-11-15 01:59:52.517760792 +0000 UTC
Nov 15 02:00:01 af867b kubelet[27751]: I1115 02:00:01.120367 27751 helpers.go:871] eviction manager: observations: signal=nodefs.inodesFree, available: 10384831, capacity: 10208Ki, time: 2017-11-15 01:59:52.517760792 +0000 UTC
Nov 15 02:00:01 af867b kubelet[27751]: I1115 02:00:01.120375 27751 helpers.go:871] eviction manager: observations: signal=imagefs.available, available: 41082368Ki, capacity: 45Gi, time: 2017-11-15 01:59:52.517760792 +0000 UTC
Nov 15 02:00:01 af867b kubelet[27751]: I1115 02:00:01.120385 27751 helpers.go:873] eviction manager: observations: signal=allocatableMemory.available, available: 7041148Ki, capacity: 7393360Ki
Nov 15 02:00:01 af867b kubelet[27751]: I1115 02:00:01.120392 27751 helpers.go:873] eviction manager: observations: signal=allocatableNodeFs.available, available: 9624040228, capacity: 10198Mi
Nov 15 02:00:01 af867b kubelet[27751]: I1115 02:00:01.120414 27751 eviction_manager.go:325] eviction manager: no resources are starved
Nov 15 02:00:02 af867b kubelet[27751]: I1115 02:00:02.523566 27751 kubelet.go:1890] SyncLoop (SYNC): 1 pods; kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)
Nov 15 02:00:02 af867b kubelet[27751]: I1115 02:00:02.523659 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:00:02 af867b kubelet[27751]: I1115 02:00:02.526759 27751 kubelet_pods.go:1284] Generating status for "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)"
Nov 15 02:00:02 af867b kubelet[27751]: I1115 02:00:02.529722 27751 status_manager.go:325] Ignoring same status for pod "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)", status: {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:58 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP:10.196.65.210 StartTime:2017-11-15 01:58:55 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:kube-controller-manager State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 01:58:57 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:gcr.io/google_containers/kube-controller-manager-amd64:v1.8.3 ImageID:docker-pullable://gcr.io/google_containers/kube-controller-manager-amd64@sha256:b6b633e3e107761d38fceb200f01bf552c51f65e3524b0aafc1a7710afff07be ContainerID:docker://272af7e3b2b1a203250d154349fdf77f296d7b7f65ce2c77b6b3a94e53dba356}] QOSClass:Burstable}
Nov 15 02:00:02 af867b kubelet[27751]: I1115 02:00:02.530006 27751 kubelet.go:1610] Creating a mirror pod for static pod "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)"
Nov 15 02:00:02 af867b kubelet[27751]: I1115 02:00:02.546808 27751 config.go:282] Setting pods for source api
Nov 15 02:00:02 af867b kubelet[27751]: I1115 02:00:02.547226 27751 config.go:404] Receiving a new pod "kube-controller-manager-af867b_kube-system(b15cbd7d-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:00:02 af867b kubelet[27751]: I1115 02:00:02.547285 27751 kubelet.go:1837] SyncLoop (ADD, "api"): "kube-controller-manager-af867b_kube-system(b15cbd7d-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:00:02 af867b kubelet[27751]: I1115 02:00:02.546811 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)"
Nov 15 02:00:02 af867b kubelet[27751]: I1115 02:00:02.847616 27751 volume_manager.go:366] All volumes are attached and mounted for pod "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)"
Nov 15 02:00:02 af867b kubelet[27751]: I1115 02:00:02.847846 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:false CreateSandbox:false SandboxID:439f518d46ec0d596518b7c17c503c52a85dd4da21d8a56ec174d5b6b3b98e03 Attempt:0 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)"
Nov 15 02:00:02 af867b kubelet[27751]: I1115 02:00:02.960006 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 2379, Path: /health
Nov 15 02:00:02 af867b kubelet[27751]: I1115 02:00:02.960060 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:00:03 af867b kubelet[27751]: I1115 02:00:03.461643 27751 http.go:96] Probe succeeded for http://127.0.0.1:2379/health, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:00:03 GMT] Content-Length:[18] Content-Type:[text/plain; charset=utf-8]] 0xc421786220 18 [] true false map[] 0xc420a30900 <nil>}
Nov 15 02:00:03 af867b kubelet[27751]: I1115 02:00:03.461717 27751 prober.go:113] Liveness probe for "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250):etcd" succeeded
Nov 15 02:00:03 af867b kubelet[27751]: I1115 02:00:03.523459 27751 kubelet.go:1890] SyncLoop (SYNC): 1 pods; kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)
Nov 15 02:00:03 af867b kubelet[27751]: I1115 02:00:03.523582 27751 kubelet_pods.go:1284] Generating status for "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)"
Nov 15 02:00:03 af867b kubelet[27751]: I1115 02:00:03.523810 27751 status_manager.go:325] Ignoring same status for pod "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)", status: {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:58 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP:10.196.65.210 StartTime:2017-11-15 01:58:55 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:kube-scheduler State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 01:58:57 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:gcr.io/google_containers/kube-scheduler-amd64:v1.8.3 ImageID:docker-pullable://gcr.io/google_containers/kube-scheduler-amd64@sha256:c47b2438bbab28d58e8cbf64b37b7f66d26b000f5c3a31626ee829a4be8fb91e ContainerID:docker://413566b22305750f9a9aa46fbe256c11e75293e80f6e0d4afb7ec9e6afcdee05}] QOSClass:Burstable}
Nov 15 02:00:03 af867b kubelet[27751]: I1115 02:00:03.524003 27751 kubelet.go:1610] Creating a mirror pod for static pod "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)"
Nov 15 02:00:03 af867b kubelet[27751]: I1115 02:00:03.530005 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)"
Nov 15 02:00:03 af867b kubelet[27751]: I1115 02:00:03.530173 27751 config.go:282] Setting pods for source api
Nov 15 02:00:03 af867b kubelet[27751]: I1115 02:00:03.530617 27751 config.go:404] Receiving a new pod "kube-scheduler-af867b_kube-system(b1f2f82f-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:00:03 af867b kubelet[27751]: I1115 02:00:03.530672 27751 kubelet.go:1837] SyncLoop (ADD, "api"): "kube-scheduler-af867b_kube-system(b1f2f82f-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:00:03 af867b kubelet[27751]: I1115 02:00:03.617380 27751 prober.go:160] HTTP-Probe Host: https://127.0.0.1, Port: 6443, Path: /healthz
Nov 15 02:00:03 af867b kubelet[27751]: I1115 02:00:03.617432 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:00:03 af867b kubelet[27751]: I1115 02:00:03.624624 27751 http.go:96] Probe succeeded for https://127.0.0.1:6443/healthz, Response: {200 OK 200 HTTP/2.0 2 0 map[Content-Type:[text/plain; charset=utf-8] Content-Length:[2] Date:[Wed, 15 Nov 2017 02:00:03 GMT]] 0xc4216f0fc0 2 [] false false map[] 0xc420d4a600 0xc42127ca50}
Nov 15 02:00:03 af867b kubelet[27751]: I1115 02:00:03.624663 27751 prober.go:113] Liveness probe for "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d):kube-apiserver" succeeded
Nov 15 02:00:03 af867b kubelet[27751]: I1115 02:00:03.658331 27751 kubelet_pods.go:1284] Generating status for "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)"
Nov 15 02:00:03 af867b kubelet[27751]: I1115 02:00:03.658511 27751 status_manager.go:325] Ignoring same status for pod "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)", status: {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:58 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP:10.196.65.210 StartTime:2017-11-15 01:58:55 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:kube-controller-manager State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 01:58:57 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:gcr.io/google_containers/kube-controller-manager-amd64:v1.8.3 ImageID:docker-pullable://gcr.io/google_containers/kube-controller-manager-amd64@sha256:b6b633e3e107761d38fceb200f01bf552c51f65e3524b0aafc1a7710afff07be ContainerID:docker://272af7e3b2b1a203250d154349fdf77f296d7b7f65ce2c77b6b3a94e53dba356}] QOSClass:Burstable}
Nov 15 02:00:03 af867b kubelet[27751]: I1115 02:00:03.658695 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)"
Nov 15 02:00:03 af867b kubelet[27751]: I1115 02:00:03.830256 27751 volume_manager.go:366] All volumes are attached and mounted for pod "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)"
Nov 15 02:00:03 af867b kubelet[27751]: I1115 02:00:03.830448 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:false CreateSandbox:false SandboxID:fe18d337ce42bd3a4d2aacb1349be028274681d8073fbe5d6377946e815fa810 Attempt:0 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)"
Nov 15 02:00:03 af867b kubelet[27751]: I1115 02:00:03.958994 27751 volume_manager.go:366] All volumes are attached and mounted for pod "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)"
Nov 15 02:00:03 af867b kubelet[27751]: I1115 02:00:03.959207 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:false CreateSandbox:false SandboxID:439f518d46ec0d596518b7c17c503c52a85dd4da21d8a56ec174d5b6b3b98e03 Attempt:0 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)"
Nov 15 02:00:04 af867b kubelet[27751]: I1115 02:00:04.523465 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:00:04 af867b kubelet[27751]: I1115 02:00:04.661690 27751 kubelet_pods.go:1284] Generating status for "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)"
Nov 15 02:00:04 af867b kubelet[27751]: I1115 02:00:04.661883 27751 status_manager.go:325] Ignoring same status for pod "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)", status: {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:58 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP:10.196.65.210 StartTime:2017-11-15 01:58:55 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:kube-scheduler State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 01:58:57 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:gcr.io/google_containers/kube-scheduler-amd64:v1.8.3 ImageID:docker-pullable://gcr.io/google_containers/kube-scheduler-amd64@sha256:c47b2438bbab28d58e8cbf64b37b7f66d26b000f5c3a31626ee829a4be8fb91e ContainerID:docker://413566b22305750f9a9aa46fbe256c11e75293e80f6e0d4afb7ec9e6afcdee05}] QOSClass:Burstable}
Nov 15 02:00:04 af867b kubelet[27751]: I1115 02:00:04.662048 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)"
Nov 15 02:00:04 af867b kubelet[27751]: I1115 02:00:04.962285 27751 volume_manager.go:366] All volumes are attached and mounted for pod "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)"
Nov 15 02:00:04 af867b kubelet[27751]: I1115 02:00:04.962460 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:false CreateSandbox:false SandboxID:fe18d337ce42bd3a4d2aacb1349be028274681d8073fbe5d6377946e815fa810 Attempt:0 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)"
Nov 15 02:00:05 af867b kubelet[27751]: W1115 02:00:05.836982 27751 cni.go:196] Unable to update cni config: No networks found in /etc/cni/net.d
Nov 15 02:00:05 af867b kubelet[27751]: I1115 02:00:05.837604 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 02:00:05 af867b kubelet[27751]: E1115 02:00:05.837631 27751 kubelet.go:2095] Container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 02:00:06 af867b kubelet[27751]: I1115 02:00:06.523449 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:00:07 af867b kubelet[27751]: I1115 02:00:07.506120 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10252, Path: /healthz
Nov 15 02:00:07 af867b kubelet[27751]: I1115 02:00:07.506162 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:00:07 af867b kubelet[27751]: I1115 02:00:07.507285 27751 http.go:96] Probe succeeded for http://127.0.0.1:10252/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Type:[text/plain; charset=utf-8] Date:[Wed, 15 Nov 2017 02:00:07 GMT] Content-Length:[2]] 0xc421171d20 2 [] true false map[] 0xc420431400 <nil>}
Nov 15 02:00:07 af867b kubelet[27751]: I1115 02:00:07.507329 27751 prober.go:113] Liveness probe for "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9):kube-controller-manager" succeeded
Nov 15 02:00:07 af867b kubelet[27751]: I1115 02:00:07.681634 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10251, Path: /healthz
Nov 15 02:00:07 af867b kubelet[27751]: I1115 02:00:07.681656 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:00:07 af867b kubelet[27751]: I1115 02:00:07.682836 27751 http.go:96] Probe succeeded for http://127.0.0.1:10251/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:00:07 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc4212e0e60 2 [] true false map[] 0xc420431800 <nil>}
Nov 15 02:00:07 af867b kubelet[27751]: I1115 02:00:07.682884 27751 prober.go:113] Liveness probe for "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373):kube-scheduler" succeeded
Nov 15 02:00:08 af867b kubelet[27751]: I1115 02:00:08.523445 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:00:10 af867b kubelet[27751]: I1115 02:00:10.523462 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:00:10 af867b kubelet[27751]: I1115 02:00:10.530782 27751 config.go:282] Setting pods for source api
Nov 15 02:00:10 af867b kubelet[27751]: I1115 02:00:10.531554 27751 status_manager.go:451] Status for pod "kube-controller-manager-af867b_kube-system(b15cbd7d-c9a8-11e7-89f4-c6b053eac242)" updated successfully: (2, {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:58 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP:10.196.65.210 StartTime:2017-11-15 01:58:55 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:kube-controller-manager State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 01:58:57 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:gcr.io/google_containers/kube-controller-manager-amd64:v1.8.3 ImageID:docker-pullable://gcr.io/google_containers/kube-controller-manager-amd64@sha256:b6b633e3e107761d38fceb200f01bf552c51f65e3524b0aafc1a7710afff07be ContainerID:docker://272af7e3b2b1a203250d154349fdf77f296d7b7f65ce2c77b6b3a94e53dba356}] QOSClass:Burstable})
Nov 15 02:00:10 af867b kubelet[27751]: I1115 02:00:10.539640 27751 kubelet.go:1850] SyncLoop (RECONCILE, "api"): "kube-controller-manager-af867b_kube-system(b15cbd7d-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:00:10 af867b kubelet[27751]: I1115 02:00:10.543515 27751 status_manager.go:451] Status for pod "kube-scheduler-af867b_kube-system(b1f2f82f-c9a8-11e7-89f4-c6b053eac242)" updated successfully: (2, {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:58 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP:10.196.65.210 StartTime:2017-11-15 01:58:55 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:kube-scheduler State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 01:58:57 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:gcr.io/google_containers/kube-scheduler-amd64:v1.8.3 ImageID:docker-pullable://gcr.io/google_containers/kube-scheduler-amd64@sha256:c47b2438bbab28d58e8cbf64b37b7f66d26b000f5c3a31626ee829a4be8fb91e ContainerID:docker://413566b22305750f9a9aa46fbe256c11e75293e80f6e0d4afb7ec9e6afcdee05}] QOSClass:Burstable})
Nov 15 02:00:10 af867b kubelet[27751]: I1115 02:00:10.543690 27751 config.go:282] Setting pods for source api
Nov 15 02:00:10 af867b kubelet[27751]: I1115 02:00:10.544287 27751 kubelet.go:1850] SyncLoop (RECONCILE, "api"): "kube-scheduler-af867b_kube-system(b1f2f82f-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:00:10 af867b kubelet[27751]: W1115 02:00:10.838909 27751 cni.go:196] Unable to update cni config: No networks found in /etc/cni/net.d
Nov 15 02:00:10 af867b kubelet[27751]: I1115 02:00:10.839087 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 02:00:10 af867b kubelet[27751]: E1115 02:00:10.839109 27751 kubelet.go:2095] Container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 02:00:11 af867b kubelet[27751]: I1115 02:00:11.120578 27751 eviction_manager.go:221] eviction manager: synchronize housekeeping
Nov 15 02:00:11 af867b kubelet[27751]: I1115 02:00:11.163882 27751 helpers.go:871] eviction manager: observations: signal=memory.available, available: 6617280Ki, capacity: 7393360Ki, time: 2017-11-15 02:00:05.296675797 +0000 UTC
Nov 15 02:00:11 af867b kubelet[27751]: I1115 02:00:11.163921 27751 helpers.go:871] eviction manager: observations: signal=nodefs.available, available: 7420096Ki, capacity: 10198Mi, time: 2017-11-15 02:00:05.296675797 +0000 UTC
Nov 15 02:00:11 af867b kubelet[27751]: I1115 02:00:11.163931 27751 helpers.go:871] eviction manager: observations: signal=nodefs.inodesFree, available: 10384831, capacity: 10208Ki, time: 2017-11-15 02:00:05.296675797 +0000 UTC
Nov 15 02:00:11 af867b kubelet[27751]: I1115 02:00:11.163939 27751 helpers.go:871] eviction manager: observations: signal=imagefs.available, available: 40107Mi, capacity: 45Gi, time: 2017-11-15 02:00:05.296675797 +0000 UTC
Nov 15 02:00:11 af867b kubelet[27751]: I1115 02:00:11.163947 27751 helpers.go:873] eviction manager: observations: signal=allocatableMemory.available, available: 7040360Ki, capacity: 7393360Ki
Nov 15 02:00:11 af867b kubelet[27751]: I1115 02:00:11.163954 27751 helpers.go:873] eviction manager: observations: signal=allocatableNodeFs.available, available: 9624040228, capacity: 10198Mi
Nov 15 02:00:11 af867b kubelet[27751]: I1115 02:00:11.163972 27751 eviction_manager.go:325] eviction manager: no resources are starved
Nov 15 02:00:11 af867b kubelet[27751]: I1115 02:00:11.523448 27751 kubelet.go:1890] SyncLoop (SYNC): 1 pods; etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)
Nov 15 02:00:11 af867b kubelet[27751]: I1115 02:00:11.523563 27751 kubelet_pods.go:1284] Generating status for "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)"
Nov 15 02:00:11 af867b kubelet[27751]: I1115 02:00:11.524141 27751 status_manager.go:325] Ignoring same status for pod "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)", status: {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:59 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP:10.196.65.210 StartTime:2017-11-15 01:58:55 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:etcd State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 01:58:58 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:gcr.io/google_containers/etcd-amd64:3.0.17 ImageID:docker-pullable://gcr.io/google_containers/etcd-amd64@sha256:d83d3545e06fb035db8512e33bd44afb55dea007a3abd7b17742d3ac6d235940 ContainerID:docker://ab147bb1b65b2001333417bb7654896e6aadb25ce71a8c48c94ae802a2e0197f}] QOSClass:BestEffort}
Nov 15 02:00:11 af867b kubelet[27751]: I1115 02:00:11.524857 27751 kubelet.go:1610] Creating a mirror pod for static pod "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)"
Nov 15 02:00:11 af867b kubelet[27751]: I1115 02:00:11.536029 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)"
Nov 15 02:00:11 af867b kubelet[27751]: I1115 02:00:11.536189 27751 config.go:282] Setting pods for source api
Nov 15 02:00:11 af867b kubelet[27751]: I1115 02:00:11.536879 27751 config.go:404] Receiving a new pod "etcd-af867b_kube-system(b6b89970-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:00:11 af867b kubelet[27751]: I1115 02:00:11.536932 27751 kubelet.go:1837] SyncLoop (ADD, "api"): "etcd-af867b_kube-system(b6b89970-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:00:11 af867b kubelet[27751]: I1115 02:00:11.836282 27751 volume_manager.go:366] All volumes are attached and mounted for pod "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)"
Nov 15 02:00:11 af867b kubelet[27751]: I1115 02:00:11.836452 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:false CreateSandbox:false SandboxID:d5076ad0e9fb270d1b8c4ff7cdadbf32d6b1e30dc42ae24dfbc5cb01bb5aa934 Attempt:0 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)"
Nov 15 02:00:12 af867b kubelet[27751]: I1115 02:00:12.523432 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:00:12 af867b kubelet[27751]: I1115 02:00:12.693892 27751 kubelet_pods.go:1284] Generating status for "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)"
Nov 15 02:00:12 af867b kubelet[27751]: I1115 02:00:12.694076 27751 status_manager.go:325] Ignoring same status for pod "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)", status: {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:59 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP:10.196.65.210 StartTime:2017-11-15 01:58:55 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:etcd State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 01:58:58 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:gcr.io/google_containers/etcd-amd64:3.0.17 ImageID:docker-pullable://gcr.io/google_containers/etcd-amd64@sha256:d83d3545e06fb035db8512e33bd44afb55dea007a3abd7b17742d3ac6d235940 ContainerID:docker://ab147bb1b65b2001333417bb7654896e6aadb25ce71a8c48c94ae802a2e0197f}] QOSClass:BestEffort}
Nov 15 02:00:12 af867b kubelet[27751]: I1115 02:00:12.694273 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)"
Nov 15 02:00:12 af867b kubelet[27751]: I1115 02:00:12.809140 27751 config.go:282] Setting pods for source api
Nov 15 02:00:12 af867b kubelet[27751]: I1115 02:00:12.809734 27751 config.go:404] Receiving a new pod "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:00:12 af867b kubelet[27751]: I1115 02:00:12.810111 27751 kubelet.go:1837] SyncLoop (ADD, "api"): "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:00:12 af867b kubelet[27751]: I1115 02:00:12.810267 27751 kubelet_pods.go:1284] Generating status for "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:00:12 af867b kubelet[27751]: I1115 02:00:12.810576 27751 qos_container_manager_linux.go:320] [ContainerManager]: Updated QoS cgroup configuration
Nov 15 02:00:12 af867b kubelet[27751]: I1115 02:00:12.814331 27751 factory.go:116] Factory "docker" was unable to handle container "/kubepods/burstable/podb77b0858-c9a8-11e7-89f4-c6b053eac242"
Nov 15 02:00:12 af867b kubelet[27751]: I1115 02:00:12.814356 27751 factory.go:105] Error trying to work out if we can handle /kubepods/burstable/podb77b0858-c9a8-11e7-89f4-c6b053eac242: /kubepods/burstable/podb77b0858-c9a8-11e7-89f4-c6b053eac242 not handled by systemd handler
Nov 15 02:00:12 af867b kubelet[27751]: I1115 02:00:12.814364 27751 factory.go:116] Factory "systemd" was unable to handle container "/kubepods/burstable/podb77b0858-c9a8-11e7-89f4-c6b053eac242"
Nov 15 02:00:12 af867b kubelet[27751]: I1115 02:00:12.814372 27751 factory.go:112] Using factory "raw" for container "/kubepods/burstable/podb77b0858-c9a8-11e7-89f4-c6b053eac242"
Nov 15 02:00:12 af867b kubelet[27751]: I1115 02:00:12.814584 27751 manager.go:932] Added container: "/kubepods/burstable/podb77b0858-c9a8-11e7-89f4-c6b053eac242" (aliases: [], namespace: "")
Nov 15 02:00:12 af867b kubelet[27751]: I1115 02:00:12.814698 27751 handler.go:325] Added event &{/kubepods/burstable/podb77b0858-c9a8-11e7-89f4-c6b053eac242 2017-11-15 02:00:12.813194711 +0000 UTC containerCreation {<nil>}}
Nov 15 02:00:12 af867b kubelet[27751]: I1115 02:00:12.814750 27751 container.go:409] Start housekeeping for container "/kubepods/burstable/podb77b0858-c9a8-11e7-89f4-c6b053eac242"
Nov 15 02:00:12 af867b kubelet[27751]: I1115 02:00:12.817255 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:00:12 af867b kubelet[27751]: I1115 02:00:12.829638 27751 status_manager.go:451] Status for pod "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242)" updated successfully: (1, {Phase:Pending Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:00:12 +0000 UTC Reason: Message:} {Type:Ready Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:00:12 +0000 UTC Reason:ContainersNotReady Message:containers with unready status: [weave weave-npc]}] Message: Reason: HostIP:10.196.65.210 PodIP:10.196.65.210 StartTime:2017-11-15 02:00:12 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:weave State:{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,} Running:nil Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:false RestartCount:0 Image:weaveworks/weave-kube:2.0.5 ImageID: ContainerID:} {Name:weave-npc State:{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,} Running:nil Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:false RestartCount:0 Image:weaveworks/weave-npc:2.0.5 ImageID: ContainerID:}] QOSClass:Burstable})
Nov 15 02:00:12 af867b kubelet[27751]: I1115 02:00:12.830235 27751 config.go:282] Setting pods for source api
Nov 15 02:00:12 af867b kubelet[27751]: I1115 02:00:12.831180 27751 kubelet.go:1850] SyncLoop (RECONCILE, "api"): "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:00:12 af867b kubelet[27751]: I1115 02:00:12.920554 27751 reconciler.go:212] operationExecutor.VerifyControllerAttachedVolume started for volume "cni-conf" (UniqueName: "kubernetes.io/host-path/b77b0858-c9a8-11e7-89f4-c6b053eac242-cni-conf") pod "weave-net-rg7fn" (UID: "b77b0858-c9a8-11e7-89f4-c6b053eac242")
Nov 15 02:00:12 af867b kubelet[27751]: I1115 02:00:12.920610 27751 reconciler.go:212] operationExecutor.VerifyControllerAttachedVolume started for volume "cni-bin2" (UniqueName: "kubernetes.io/host-path/b77b0858-c9a8-11e7-89f4-c6b053eac242-cni-bin2") pod "weave-net-rg7fn" (UID: "b77b0858-c9a8-11e7-89f4-c6b053eac242")
Nov 15 02:00:12 af867b kubelet[27751]: I1115 02:00:12.920642 27751 reconciler.go:212] operationExecutor.VerifyControllerAttachedVolume started for volume "xtables-lock" (UniqueName: "kubernetes.io/host-path/b77b0858-c9a8-11e7-89f4-c6b053eac242-xtables-lock") pod "weave-net-rg7fn" (UID: "b77b0858-c9a8-11e7-89f4-c6b053eac242")
Nov 15 02:00:12 af867b kubelet[27751]: I1115 02:00:12.920670 27751 reconciler.go:212] operationExecutor.VerifyControllerAttachedVolume started for volume "cni-bin" (UniqueName: "kubernetes.io/host-path/b77b0858-c9a8-11e7-89f4-c6b053eac242-cni-bin") pod "weave-net-rg7fn" (UID: "b77b0858-c9a8-11e7-89f4-c6b053eac242")
Nov 15 02:00:12 af867b kubelet[27751]: I1115 02:00:12.920695 27751 reconciler.go:212] operationExecutor.VerifyControllerAttachedVolume started for volume "dbus" (UniqueName: "kubernetes.io/host-path/b77b0858-c9a8-11e7-89f4-c6b053eac242-dbus") pod "weave-net-rg7fn" (UID: "b77b0858-c9a8-11e7-89f4-c6b053eac242")
Nov 15 02:00:12 af867b kubelet[27751]: I1115 02:00:12.920741 27751 reconciler.go:212] operationExecutor.VerifyControllerAttachedVolume started for volume "weavedb" (UniqueName: "kubernetes.io/host-path/b77b0858-c9a8-11e7-89f4-c6b053eac242-weavedb") pod "weave-net-rg7fn" (UID: "b77b0858-c9a8-11e7-89f4-c6b053eac242")
Nov 15 02:00:12 af867b kubelet[27751]: I1115 02:00:12.920833 27751 reconciler.go:212] operationExecutor.VerifyControllerAttachedVolume started for volume "lib-modules" (UniqueName: "kubernetes.io/host-path/b77b0858-c9a8-11e7-89f4-c6b053eac242-lib-modules") pod "weave-net-rg7fn" (UID: "b77b0858-c9a8-11e7-89f4-c6b053eac242")
Nov 15 02:00:12 af867b kubelet[27751]: I1115 02:00:12.920900 27751 reconciler.go:212] operationExecutor.VerifyControllerAttachedVolume started for volume "weave-net-token-rn6j7" (UniqueName: "kubernetes.io/secret/b77b0858-c9a8-11e7-89f4-c6b053eac242-weave-net-token-rn6j7") pod "weave-net-rg7fn" (UID: "b77b0858-c9a8-11e7-89f4-c6b053eac242")
Nov 15 02:00:12 af867b kubelet[27751]: I1115 02:00:12.959969 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 2379, Path: /health
Nov 15 02:00:12 af867b kubelet[27751]: I1115 02:00:12.960014 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:00:12 af867b kubelet[27751]: I1115 02:00:12.994554 27751 volume_manager.go:366] All volumes are attached and mounted for pod "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)"
Nov 15 02:00:12 af867b kubelet[27751]: I1115 02:00:12.994698 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:false CreateSandbox:false SandboxID:d5076ad0e9fb270d1b8c4ff7cdadbf32d6b1e30dc42ae24dfbc5cb01bb5aa934 Attempt:0 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)"
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.021193 27751 reconciler.go:257] operationExecutor.MountVolume started for volume "dbus" (UniqueName: "kubernetes.io/host-path/b77b0858-c9a8-11e7-89f4-c6b053eac242-dbus") pod "weave-net-rg7fn" (UID: "b77b0858-c9a8-11e7-89f4-c6b053eac242")
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.021253 27751 reconciler.go:257] operationExecutor.MountVolume started for volume "cni-bin" (UniqueName: "kubernetes.io/host-path/b77b0858-c9a8-11e7-89f4-c6b053eac242-cni-bin") pod "weave-net-rg7fn" (UID: "b77b0858-c9a8-11e7-89f4-c6b053eac242")
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.021289 27751 reconciler.go:257] operationExecutor.MountVolume started for volume "weavedb" (UniqueName: "kubernetes.io/host-path/b77b0858-c9a8-11e7-89f4-c6b053eac242-weavedb") pod "weave-net-rg7fn" (UID: "b77b0858-c9a8-11e7-89f4-c6b053eac242")
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.021324 27751 reconciler.go:257] operationExecutor.MountVolume started for volume "lib-modules" (UniqueName: "kubernetes.io/host-path/b77b0858-c9a8-11e7-89f4-c6b053eac242-lib-modules") pod "weave-net-rg7fn" (UID: "b77b0858-c9a8-11e7-89f4-c6b053eac242")
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.021366 27751 reconciler.go:257] operationExecutor.MountVolume started for volume "weave-net-token-rn6j7" (UniqueName: "kubernetes.io/secret/b77b0858-c9a8-11e7-89f4-c6b053eac242-weave-net-token-rn6j7") pod "weave-net-rg7fn" (UID: "b77b0858-c9a8-11e7-89f4-c6b053eac242")
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.021393 27751 reconciler.go:257] operationExecutor.MountVolume started for volume "cni-bin2" (UniqueName: "kubernetes.io/host-path/b77b0858-c9a8-11e7-89f4-c6b053eac242-cni-bin2") pod "weave-net-rg7fn" (UID: "b77b0858-c9a8-11e7-89f4-c6b053eac242")
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.021428 27751 reconciler.go:257] operationExecutor.MountVolume started for volume "cni-conf" (UniqueName: "kubernetes.io/host-path/b77b0858-c9a8-11e7-89f4-c6b053eac242-cni-conf") pod "weave-net-rg7fn" (UID: "b77b0858-c9a8-11e7-89f4-c6b053eac242")
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.021459 27751 reconciler.go:257] operationExecutor.MountVolume started for volume "xtables-lock" (UniqueName: "kubernetes.io/host-path/b77b0858-c9a8-11e7-89f4-c6b053eac242-xtables-lock") pod "weave-net-rg7fn" (UID: "b77b0858-c9a8-11e7-89f4-c6b053eac242")
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.021546 27751 operation_generator.go:484] MountVolume.SetUp succeeded for volume "xtables-lock" (UniqueName: "kubernetes.io/host-path/b77b0858-c9a8-11e7-89f4-c6b053eac242-xtables-lock") pod "weave-net-rg7fn" (UID: "b77b0858-c9a8-11e7-89f4-c6b053eac242")
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.022012 27751 operation_generator.go:484] MountVolume.SetUp succeeded for volume "dbus" (UniqueName: "kubernetes.io/host-path/b77b0858-c9a8-11e7-89f4-c6b053eac242-dbus") pod "weave-net-rg7fn" (UID: "b77b0858-c9a8-11e7-89f4-c6b053eac242")
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.022114 27751 operation_generator.go:484] MountVolume.SetUp succeeded for volume "cni-bin" (UniqueName: "kubernetes.io/host-path/b77b0858-c9a8-11e7-89f4-c6b053eac242-cni-bin") pod "weave-net-rg7fn" (UID: "b77b0858-c9a8-11e7-89f4-c6b053eac242")
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.022158 27751 operation_generator.go:484] MountVolume.SetUp succeeded for volume "weavedb" (UniqueName: "kubernetes.io/host-path/b77b0858-c9a8-11e7-89f4-c6b053eac242-weavedb") pod "weave-net-rg7fn" (UID: "b77b0858-c9a8-11e7-89f4-c6b053eac242")
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.022197 27751 operation_generator.go:484] MountVolume.SetUp succeeded for volume "lib-modules" (UniqueName: "kubernetes.io/host-path/b77b0858-c9a8-11e7-89f4-c6b053eac242-lib-modules") pod "weave-net-rg7fn" (UID: "b77b0858-c9a8-11e7-89f4-c6b053eac242")
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.022231 27751 secret.go:186] Setting up volume weave-net-token-rn6j7 for pod b77b0858-c9a8-11e7-89f4-c6b053eac242 at /var/lib/kubelet/pods/b77b0858-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/weave-net-token-rn6j7
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.022433 27751 empty_dir.go:264] pod b77b0858-c9a8-11e7-89f4-c6b053eac242: mounting tmpfs for volume wrapped_weave-net-token-rn6j7
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.022452 27751 mount_linux.go:135] Mounting cmd (systemd-run) with arguments ([--description=Kubernetes transient mount for /var/lib/kubelet/pods/b77b0858-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/weave-net-token-rn6j7 --scope -- mount -t tmpfs tmpfs /var/lib/kubelet/pods/b77b0858-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/weave-net-token-rn6j7])
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.024265 27751 operation_generator.go:484] MountVolume.SetUp succeeded for volume "cni-conf" (UniqueName: "kubernetes.io/host-path/b77b0858-c9a8-11e7-89f4-c6b053eac242-cni-conf") pod "weave-net-rg7fn" (UID: "b77b0858-c9a8-11e7-89f4-c6b053eac242")
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.024399 27751 operation_generator.go:484] MountVolume.SetUp succeeded for volume "cni-bin2" (UniqueName: "kubernetes.io/host-path/b77b0858-c9a8-11e7-89f4-c6b053eac242-cni-bin2") pod "weave-net-rg7fn" (UID: "b77b0858-c9a8-11e7-89f4-c6b053eac242")
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.024437 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"weave-net-rg7fn", UID:"b77b0858-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"381", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "xtables-lock"
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.024467 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"weave-net-rg7fn", UID:"b77b0858-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"381", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "dbus"
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.024495 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"weave-net-rg7fn", UID:"b77b0858-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"381", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "cni-bin"
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.024505 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"weave-net-rg7fn", UID:"b77b0858-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"381", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "weavedb"
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.024523 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"weave-net-rg7fn", UID:"b77b0858-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"381", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "lib-modules"
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.024532 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"weave-net-rg7fn", UID:"b77b0858-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"381", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "cni-conf"
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.024556 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"weave-net-rg7fn", UID:"b77b0858-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"381", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "cni-bin2"
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.033146 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/run-29268.scope"
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.033169 27751 factory.go:105] Error trying to work out if we can handle /system.slice/run-29268.scope: /system.slice/run-29268.scope not handled by systemd handler
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.033175 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/run-29268.scope"
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.033182 27751 factory.go:112] Using factory "raw" for container "/system.slice/run-29268.scope"
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.033358 27751 manager.go:932] Added container: "/system.slice/run-29268.scope" (aliases: [], namespace: "")
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.033476 27751 handler.go:325] Added event &{/system.slice/run-29268.scope 2017-11-15 02:00:13.031194711 +0000 UTC containerCreation {<nil>}}
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.033506 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-984a35c7cf6d04c0af5fcd260872a7b6556aa9cd8b38bc1a1cf635972c1d6511.mount"
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.033518 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-984a35c7cf6d04c0af5fcd260872a7b6556aa9cd8b38bc1a1cf635972c1d6511.mount", but ignoring.
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.033528 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-984a35c7cf6d04c0af5fcd260872a7b6556aa9cd8b38bc1a1cf635972c1d6511.mount"
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.033537 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-91510886f3beb621e5d04309d502352ff78392e46405f21633802da4f7047069-shm.mount"
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.033545 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-91510886f3beb621e5d04309d502352ff78392e46405f21633802da4f7047069-shm.mount", but ignoring.
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.033554 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-containers-91510886f3beb621e5d04309d502352ff78392e46405f21633802da4f7047069-shm.mount"
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.033577 27751 container.go:409] Start housekeeping for container "/system.slice/run-29268.scope"
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.039932 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-kubelet-pods-9729c03a\\x2dc9a8\\x2d11e7\\x2d89f4\\x2dc6b053eac242-volumes-kubernetes.io\\x7esecret-kube\\x2dproxy\\x2dtoken\\x2dgqhfs.mount"
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.039963 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-kubelet-pods-9729c03a\\x2dc9a8\\x2d11e7\\x2d89f4\\x2dc6b053eac242-volumes-kubernetes.io\\x7esecret-kube\\x2dproxy\\x2dtoken\\x2dgqhfs.mount", but ignoring.
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.039981 27751 manager.go:901] ignoring container "/system.slice/var-lib-kubelet-pods-9729c03a\\x2dc9a8\\x2d11e7\\x2d89f4\\x2dc6b053eac242-volumes-kubernetes.io\\x7esecret-kube\\x2dproxy\\x2dtoken\\x2dgqhfs.mount"
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.039995 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-7d2ebdadb8c541a4839058dd46b491ef9e14209821b47e7d86b539de402fbef5.mount"
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.040005 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-7d2ebdadb8c541a4839058dd46b491ef9e14209821b47e7d86b539de402fbef5.mount", but ignoring.
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.040015 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-7d2ebdadb8c541a4839058dd46b491ef9e14209821b47e7d86b539de402fbef5.mount"
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.042819 27751 secret.go:217] Received secret kube-system/weave-net-token-rn6j7 containing (3) pieces of data, 1900 total bytes
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.042893 27751 atomic_writer.go:145] pod kube-system/weave-net-rg7fn volume weave-net-token-rn6j7: write required for target directory /var/lib/kubelet/pods/b77b0858-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/weave-net-token-rn6j7
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.042968 27751 atomic_writer.go:160] pod kube-system/weave-net-rg7fn volume weave-net-token-rn6j7: performed write of new data to ts data directory: /var/lib/kubelet/pods/b77b0858-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/weave-net-token-rn6j7/..119811_15_11_02_00_13.100287739
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.043046 27751 operation_generator.go:484] MountVolume.SetUp succeeded for volume "weave-net-token-rn6j7" (UniqueName: "kubernetes.io/secret/b77b0858-c9a8-11e7-89f4-c6b053eac242-weave-net-token-rn6j7") pod "weave-net-rg7fn" (UID: "b77b0858-c9a8-11e7-89f4-c6b053eac242")
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.043075 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"weave-net-rg7fn", UID:"b77b0858-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"381", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "weave-net-token-rn6j7"
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.044941 27751 manager.go:989] Destroyed container: "/system.slice/run-29268.scope" (aliases: [], namespace: "")
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.044961 27751 handler.go:325] Added event &{/system.slice/run-29268.scope 2017-11-15 02:00:13.044956891 +0000 UTC containerDeletion {<nil>}}
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.117571 27751 volume_manager.go:366] All volumes are attached and mounted for pod "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.117606 27751 kuberuntime_manager.go:370] No sandbox for pod "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242)" can be found. Need to start a new one
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.117619 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:true CreateSandbox:true SandboxID: Attempt:0 NextInitContainerToStart:nil ContainersToStart:[0 1] ContainersToKill:map[]} for pod "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.117688 27751 kuberuntime_manager.go:565] SyncPod received new pod "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242)", will create a sandbox for it
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.117699 27751 kuberuntime_manager.go:574] Stopping PodSandbox for "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242)", will start new one
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.117720 27751 kuberuntime_manager.go:626] Creating sandbox for pod "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.119592 27751 expiration_cache.go:98] Entry version: {key:version obj:0xc4206acc80} has expired
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.120191 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/burstable/podb77b0858-c9a8-11e7-89f4-c6b053eac242"
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.120217 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/burstable/podb77b0858-c9a8-11e7-89f4-c6b053eac242"
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.211220 27751 http.go:96] Probe succeeded for http://127.0.0.1:2379/health, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:00:13 GMT] Content-Length:[18] Content-Type:[text/plain; charset=utf-8]] 0xc420f0dbc0 18 [] true false map[] 0xc420afd200 <nil>}
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.211285 27751 prober.go:113] Liveness probe for "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250):etcd" succeeded
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.425860 27751 factory.go:112] Using factory "docker" for container "/kubepods/burstable/podb77b0858-c9a8-11e7-89f4-c6b053eac242/be14b0552100f929052c8e94e1b7cf587c4218e321d811314e58f2cf312f1c0f"
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.427245 27751 docker_sandbox.go:691] Will attempt to re-write config file /var/lib/docker/containers/be14b0552100f929052c8e94e1b7cf587c4218e321d811314e58f2cf312f1c0f/resolv.conf with:
Nov 15 02:00:13 af867b kubelet[27751]: [nameserver 10.196.65.209 search opcwlaas.oraclecloud.internal. opcwlaas.oraclecloud.internal.]
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.427250 27751 manager.go:932] Added container: "/kubepods/burstable/podb77b0858-c9a8-11e7-89f4-c6b053eac242/be14b0552100f929052c8e94e1b7cf587c4218e321d811314e58f2cf312f1c0f" (aliases: [k8s_POD_weave-net-rg7fn_kube-system_b77b0858-c9a8-11e7-89f4-c6b053eac242_0 be14b0552100f929052c8e94e1b7cf587c4218e321d811314e58f2cf312f1c0f], namespace: "docker")
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.427395 27751 handler.go:325] Added event &{/kubepods/burstable/podb77b0858-c9a8-11e7-89f4-c6b053eac242/be14b0552100f929052c8e94e1b7cf587c4218e321d811314e58f2cf312f1c0f 2017-11-15 02:00:13.367194711 +0000 UTC containerCreation {<nil>}}
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.427435 27751 container.go:409] Start housekeeping for container "/kubepods/burstable/podb77b0858-c9a8-11e7-89f4-c6b053eac242/be14b0552100f929052c8e94e1b7cf587c4218e321d811314e58f2cf312f1c0f"
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.427597 27751 kuberuntime_manager.go:640] Created PodSandbox "be14b0552100f929052c8e94e1b7cf587c4218e321d811314e58f2cf312f1c0f" for pod "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.429920 27751 kuberuntime_manager.go:705] Creating container &Container{Name:weave,Image:weaveworks/weave-kube:2.0.5,Command:[/home/weave/launch.sh],Args:[],WorkingDir:,Ports:[],Env:[{HOSTNAME EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,}}],Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {<nil>} 10m DecimalSI},},},VolumeMounts:[{weavedb false /weavedb <nil>} {cni-bin false /host/opt <nil>} {cni-bin2 false /host/home <nil>} {cni-conf false /host/etc <nil>} {dbus false /host/var/lib/dbus <nil>} {lib-modules false /lib/modules <nil>} {xtables-lock false /run/xtables.lock <nil>} {weave-net-token-rn6j7 true /var/run/secrets/kubernetes.io/serviceaccount <nil>}],LivenessProbe:&Probe{Handler:Handler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/status,Port:6784,Host:127.0.0.1,Scheme:HTTP,HTTPHeaders:[],},TCPSocket:nil,},InitialDelaySeconds:30,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[],TerminationMessagePolicy:File,} in pod weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242)
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.431028 27751 provider.go:119] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.431060 27751 config.go:131] looking for config.json at /var/lib/kubelet/config.json
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.431096 27751 config.go:131] looking for config.json at /config.json
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.431108 27751 config.go:131] looking for config.json at /.docker/config.json
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.431117 27751 config.go:131] looking for config.json at /.docker/config.json
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.431129 27751 config.go:101] looking for .dockercfg at /var/lib/kubelet/.dockercfg
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.431140 27751 config.go:101] looking for .dockercfg at /.dockercfg
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.431150 27751 config.go:101] looking for .dockercfg at /.dockercfg
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.431160 27751 config.go:101] looking for .dockercfg at /.dockercfg
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.431170 27751 provider.go:89] Unable to parse Docker config file: couldn't find valid .dockercfg after checking in [/var/lib/kubelet /]
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.431183 27751 kuberuntime_image.go:46] Pulling image "weaveworks/weave-kube:2.0.5" without credentials
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.431256 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"weave-net-rg7fn", UID:"b77b0858-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"381", FieldPath:"spec.containers{weave}"}): type: 'Normal' reason: 'Pulling' pulling image "weaveworks/weave-kube:2.0.5"
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.617344 27751 prober.go:160] HTTP-Probe Host: https://127.0.0.1, Port: 6443, Path: /healthz
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.617380 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.623731 27751 http.go:96] Probe succeeded for https://127.0.0.1:6443/healthz, Response: {200 OK 200 HTTP/2.0 2 0 map[Content-Type:[text/plain; charset=utf-8] Content-Length:[2] Date:[Wed, 15 Nov 2017 02:00:13 GMT]] 0xc42138a9a0 2 [] false false map[] 0xc420c91300 0xc420b44d10}
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.623775 27751 prober.go:113] Liveness probe for "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d):kube-apiserver" succeeded
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.698029 27751 generic.go:146] GenericPLEG: b77b0858-c9a8-11e7-89f4-c6b053eac242/be14b0552100f929052c8e94e1b7cf587c4218e321d811314e58f2cf312f1c0f: non-existent -> running
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.699194 27751 kuberuntime_manager.go:833] getSandboxIDByPodUID got sandbox IDs ["be14b0552100f929052c8e94e1b7cf587c4218e321d811314e58f2cf312f1c0f"] for pod "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.701376 27751 generic.go:345] PLEG: Write status for weave-net-rg7fn/kube-system: &container.PodStatus{ID:"b77b0858-c9a8-11e7-89f4-c6b053eac242", Name:"weave-net-rg7fn", Namespace:"kube-system", IP:"", ContainerStatuses:[]*container.ContainerStatus{}, SandboxStatuses:[]*runtime.PodSandboxStatus{(*runtime.PodSandboxStatus)(0xc4211e51d0)}} (err: <nil>)
Nov 15 02:00:13 af867b kubelet[27751]: I1115 02:00:13.701447 27751 kubelet.go:1871] SyncLoop (PLEG): "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242)", event: &pleg.PodLifecycleEvent{ID:"b77b0858-c9a8-11e7-89f4-c6b053eac242", Type:"ContainerStarted", Data:"be14b0552100f929052c8e94e1b7cf587c4218e321d811314e58f2cf312f1c0f"}
Nov 15 02:00:14 af867b kubelet[27751]: I1115 02:00:14.523469 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:00:15 af867b kubelet[27751]: I1115 02:00:15.362590 27751 worker.go:164] Probe target container not found: weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242) - weave
Nov 15 02:00:15 af867b kubelet[27751]: I1115 02:00:15.523554 27751 kubelet.go:1890] SyncLoop (SYNC): 1 pods; kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)
Nov 15 02:00:15 af867b kubelet[27751]: I1115 02:00:15.523818 27751 kubelet_pods.go:1284] Generating status for "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)"
Nov 15 02:00:15 af867b kubelet[27751]: I1115 02:00:15.524269 27751 status_manager.go:325] Ignoring same status for pod "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)", status: {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:58 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP:10.196.65.210 StartTime:2017-11-15 01:58:55 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:kube-apiserver State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 01:58:58 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:gcr.io/google_containers/kube-apiserver-amd64:v1.8.3 ImageID:docker-pullable://gcr.io/google_containers/kube-apiserver-amd64@sha256:872e3d4286a8ef4338df59945cb0d64c2622268ceb3e8a2ce7b52243279b02d0 ContainerID:docker://8da2c70a27f08a2f062af80b5708e01ac34ce76b42ab4a6eaa0288e2daf8a043}] QOSClass:Burstable}
Nov 15 02:00:15 af867b kubelet[27751]: I1115 02:00:15.524607 27751 kubelet.go:1610] Creating a mirror pod for static pod "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)"
Nov 15 02:00:15 af867b kubelet[27751]: I1115 02:00:15.539785 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)"
Nov 15 02:00:15 af867b kubelet[27751]: I1115 02:00:15.539976 27751 config.go:282] Setting pods for source api
Nov 15 02:00:15 af867b kubelet[27751]: I1115 02:00:15.540524 27751 config.go:404] Receiving a new pod "kube-apiserver-af867b_kube-system(b91a02e8-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:00:15 af867b kubelet[27751]: I1115 02:00:15.541971 27751 kubelet.go:1837] SyncLoop (ADD, "api"): "kube-apiserver-af867b_kube-system(b91a02e8-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:00:15 af867b kubelet[27751]: I1115 02:00:15.840265 27751 volume_manager.go:366] All volumes are attached and mounted for pod "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)"
Nov 15 02:00:15 af867b kubelet[27751]: I1115 02:00:15.840451 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:false CreateSandbox:false SandboxID:19b06ca14052c17f92bf03d5316fd7642fe4fee6bf193755575355be3e87a13d Attempt:0 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)"
Nov 15 02:00:15 af867b kubelet[27751]: W1115 02:00:15.841565 27751 cni.go:196] Unable to update cni config: No networks found in /etc/cni/net.d
Nov 15 02:00:15 af867b kubelet[27751]: I1115 02:00:15.841822 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 02:00:15 af867b kubelet[27751]: E1115 02:00:15.841881 27751 kubelet.go:2095] Container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 02:00:16 af867b kubelet[27751]: I1115 02:00:16.523482 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:00:16 af867b kubelet[27751]: I1115 02:00:16.713391 27751 kubelet_pods.go:1284] Generating status for "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)"
Nov 15 02:00:16 af867b kubelet[27751]: I1115 02:00:16.713624 27751 status_manager.go:325] Ignoring same status for pod "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)", status: {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:58 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP:10.196.65.210 StartTime:2017-11-15 01:58:55 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:kube-apiserver State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 01:58:58 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:gcr.io/google_containers/kube-apiserver-amd64:v1.8.3 ImageID:docker-pullable://gcr.io/google_containers/kube-apiserver-amd64@sha256:872e3d4286a8ef4338df59945cb0d64c2622268ceb3e8a2ce7b52243279b02d0 ContainerID:docker://8da2c70a27f08a2f062af80b5708e01ac34ce76b42ab4a6eaa0288e2daf8a043}] QOSClass:Burstable}
Nov 15 02:00:16 af867b kubelet[27751]: I1115 02:00:16.713817 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)"
Nov 15 02:00:17 af867b kubelet[27751]: I1115 02:00:17.014096 27751 volume_manager.go:366] All volumes are attached and mounted for pod "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)"
Nov 15 02:00:17 af867b kubelet[27751]: I1115 02:00:17.014297 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:false CreateSandbox:false SandboxID:19b06ca14052c17f92bf03d5316fd7642fe4fee6bf193755575355be3e87a13d Attempt:0 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)"
Nov 15 02:00:17 af867b kubelet[27751]: I1115 02:00:17.506132 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10252, Path: /healthz
Nov 15 02:00:17 af867b kubelet[27751]: I1115 02:00:17.506179 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:00:17 af867b kubelet[27751]: I1115 02:00:17.507170 27751 http.go:96] Probe succeeded for http://127.0.0.1:10252/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:00:17 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc4216c48c0 2 [] true false map[] 0xc4200dd600 <nil>}
Nov 15 02:00:17 af867b kubelet[27751]: I1115 02:00:17.507222 27751 prober.go:113] Liveness probe for "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9):kube-controller-manager" succeeded
Nov 15 02:00:17 af867b kubelet[27751]: I1115 02:00:17.681694 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10251, Path: /healthz
Nov 15 02:00:17 af867b kubelet[27751]: I1115 02:00:17.681744 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:00:17 af867b kubelet[27751]: I1115 02:00:17.682678 27751 http.go:96] Probe succeeded for http://127.0.0.1:10251/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:00:17 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc42159a9e0 2 [] true false map[] 0xc4200dd800 <nil>}
Nov 15 02:00:17 af867b kubelet[27751]: I1115 02:00:17.682717 27751 prober.go:113] Liveness probe for "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373):kube-scheduler" succeeded
Nov 15 02:00:18 af867b kubelet[27751]: I1115 02:00:18.525670 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:00:20 af867b kubelet[27751]: I1115 02:00:20.523542 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:00:20 af867b kubelet[27751]: I1115 02:00:20.532140 27751 status_manager.go:451] Status for pod "etcd-af867b_kube-system(b6b89970-c9a8-11e7-89f4-c6b053eac242)" updated successfully: (2, {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:59 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP:10.196.65.210 StartTime:2017-11-15 01:58:55 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:etcd State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 01:58:58 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:gcr.io/google_containers/etcd-amd64:3.0.17 ImageID:docker-pullable://gcr.io/google_containers/etcd-amd64@sha256:d83d3545e06fb035db8512e33bd44afb55dea007a3abd7b17742d3ac6d235940 ContainerID:docker://ab147bb1b65b2001333417bb7654896e6aadb25ce71a8c48c94ae802a2e0197f}] QOSClass:BestEffort})
Nov 15 02:00:20 af867b kubelet[27751]: I1115 02:00:20.532466 27751 config.go:282] Setting pods for source api
Nov 15 02:00:20 af867b kubelet[27751]: I1115 02:00:20.540181 27751 config.go:282] Setting pods for source api
Nov 15 02:00:20 af867b kubelet[27751]: I1115 02:00:20.540703 27751 status_manager.go:451] Status for pod "kube-apiserver-af867b_kube-system(b91a02e8-c9a8-11e7-89f4-c6b053eac242)" updated successfully: (2, {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:58 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP:10.196.65.210 StartTime:2017-11-15 01:58:55 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:kube-apiserver State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 01:58:58 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:gcr.io/google_containers/kube-apiserver-amd64:v1.8.3 ImageID:docker-pullable://gcr.io/google_containers/kube-apiserver-amd64@sha256:872e3d4286a8ef4338df59945cb0d64c2622268ceb3e8a2ce7b52243279b02d0 ContainerID:docker://8da2c70a27f08a2f062af80b5708e01ac34ce76b42ab4a6eaa0288e2daf8a043}] QOSClass:Burstable})
Nov 15 02:00:20 af867b kubelet[27751]: I1115 02:00:20.550969 27751 kubelet.go:1850] SyncLoop (RECONCILE, "api"): "etcd-af867b_kube-system(b6b89970-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:00:20 af867b kubelet[27751]: I1115 02:00:20.551029 27751 kubelet.go:1850] SyncLoop (RECONCILE, "api"): "kube-apiserver-af867b_kube-system(b91a02e8-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:00:20 af867b kubelet[27751]: W1115 02:00:20.843145 27751 cni.go:196] Unable to update cni config: No networks found in /etc/cni/net.d
Nov 15 02:00:20 af867b kubelet[27751]: I1115 02:00:20.843297 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 02:00:20 af867b kubelet[27751]: E1115 02:00:20.843320 27751 kubelet.go:2095] Container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 02:00:21 af867b kubelet[27751]: I1115 02:00:21.164195 27751 eviction_manager.go:221] eviction manager: synchronize housekeeping
Nov 15 02:00:21 af867b kubelet[27751]: I1115 02:00:21.228893 27751 helpers.go:873] eviction manager: observations: signal=allocatableMemory.available, available: 7040084Ki, capacity: 7393360Ki
Nov 15 02:00:21 af867b kubelet[27751]: I1115 02:00:21.228930 27751 helpers.go:873] eviction manager: observations: signal=allocatableNodeFs.available, available: 9624040228, capacity: 10198Mi
Nov 15 02:00:21 af867b kubelet[27751]: I1115 02:00:21.228939 27751 helpers.go:871] eviction manager: observations: signal=memory.available, available: 6617280Ki, capacity: 7393360Ki, time: 2017-11-15 02:00:05.296675797 +0000 UTC
Nov 15 02:00:21 af867b kubelet[27751]: I1115 02:00:21.228966 27751 helpers.go:871] eviction manager: observations: signal=nodefs.available, available: 7420096Ki, capacity: 10198Mi, time: 2017-11-15 02:00:05.296675797 +0000 UTC
Nov 15 02:00:21 af867b kubelet[27751]: I1115 02:00:21.228976 27751 helpers.go:871] eviction manager: observations: signal=nodefs.inodesFree, available: 10384831, capacity: 10208Ki, time: 2017-11-15 02:00:05.296675797 +0000 UTC
Nov 15 02:00:21 af867b kubelet[27751]: I1115 02:00:21.228986 27751 helpers.go:871] eviction manager: observations: signal=imagefs.available, available: 40107Mi, capacity: 45Gi, time: 2017-11-15 02:00:05.296675797 +0000 UTC
Nov 15 02:00:21 af867b kubelet[27751]: I1115 02:00:21.229009 27751 eviction_manager.go:325] eviction manager: no resources are starved
Nov 15 02:00:22 af867b kubelet[27751]: I1115 02:00:22.523505 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:00:22 af867b kubelet[27751]: I1115 02:00:22.961802 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 2379, Path: /health
Nov 15 02:00:22 af867b kubelet[27751]: I1115 02:00:22.961848 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:00:23 af867b kubelet[27751]: I1115 02:00:23.470109 27751 http.go:96] Probe succeeded for http://127.0.0.1:2379/health, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:00:23 GMT] Content-Length:[18] Content-Type:[text/plain; charset=utf-8]] 0xc421438720 18 [] true false map[] 0xc420ee3100 <nil>}
Nov 15 02:00:23 af867b kubelet[27751]: I1115 02:00:23.470171 27751 prober.go:113] Liveness probe for "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250):etcd" succeeded
Nov 15 02:00:23 af867b kubelet[27751]: I1115 02:00:23.617358 27751 prober.go:160] HTTP-Probe Host: https://127.0.0.1, Port: 6443, Path: /healthz
Nov 15 02:00:23 af867b kubelet[27751]: I1115 02:00:23.617401 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:00:23 af867b kubelet[27751]: I1115 02:00:23.624664 27751 http.go:96] Probe succeeded for https://127.0.0.1:6443/healthz, Response: {200 OK 200 HTTP/2.0 2 0 map[Content-Length:[2] Date:[Wed, 15 Nov 2017 02:00:23 GMT] Content-Type:[text/plain; charset=utf-8]] 0xc420f64580 2 [] false false map[] 0xc420d4be00 0xc42131e790}
Nov 15 02:00:23 af867b kubelet[27751]: I1115 02:00:23.624732 27751 prober.go:113] Liveness probe for "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d):kube-apiserver" succeeded
Nov 15 02:00:24 af867b kubelet[27751]: I1115 02:00:24.523808 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:00:24 af867b kubelet[27751]: I1115 02:00:24.790344 27751 kube_docker_client.go:330] Pulling image "weaveworks/weave-kube:2.0.5": "27c6c140ede3: Extracting [====================================> ] 5.407MB/7.44MB"
Nov 15 02:00:25 af867b kubelet[27751]: I1115 02:00:25.362712 27751 worker.go:164] Probe target container not found: weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242) - weave
Nov 15 02:00:25 af867b kubelet[27751]: W1115 02:00:25.844428 27751 cni.go:196] Unable to update cni config: No networks found in /etc/cni/net.d
Nov 15 02:00:25 af867b kubelet[27751]: I1115 02:00:25.844571 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 02:00:25 af867b kubelet[27751]: E1115 02:00:25.844594 27751 kubelet.go:2095] Container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 02:00:26 af867b kubelet[27751]: I1115 02:00:26.523483 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:00:27 af867b kubelet[27751]: I1115 02:00:27.506176 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10252, Path: /healthz
Nov 15 02:00:27 af867b kubelet[27751]: I1115 02:00:27.506227 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:00:27 af867b kubelet[27751]: I1115 02:00:27.510746 27751 http.go:96] Probe succeeded for http://127.0.0.1:10252/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:00:27 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc421bf5060 2 [] true false map[] 0xc421179100 <nil>}
Nov 15 02:00:27 af867b kubelet[27751]: I1115 02:00:27.510818 27751 prober.go:113] Liveness probe for "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9):kube-controller-manager" succeeded
Nov 15 02:00:27 af867b kubelet[27751]: I1115 02:00:27.681816 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10251, Path: /healthz
Nov 15 02:00:27 af867b kubelet[27751]: I1115 02:00:27.681858 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:00:27 af867b kubelet[27751]: I1115 02:00:27.682690 27751 http.go:96] Probe succeeded for http://127.0.0.1:10251/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Length:[2] Content-Type:[text/plain; charset=utf-8] Date:[Wed, 15 Nov 2017 02:00:27 GMT]] 0xc421bf52c0 2 [] true false map[] 0xc420a31f00 <nil>}
Nov 15 02:00:27 af867b kubelet[27751]: I1115 02:00:27.682761 27751 prober.go:113] Liveness probe for "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373):kube-scheduler" succeeded
Nov 15 02:00:28 af867b kubelet[27751]: I1115 02:00:28.523555 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:00:30 af867b kubelet[27751]: I1115 02:00:30.523545 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:00:30 af867b kubelet[27751]: W1115 02:00:30.845392 27751 cni.go:196] Unable to update cni config: No networks found in /etc/cni/net.d
Nov 15 02:00:30 af867b kubelet[27751]: I1115 02:00:30.845496 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 02:00:30 af867b kubelet[27751]: E1115 02:00:30.845520 27751 kubelet.go:2095] Container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 02:00:31 af867b kubelet[27751]: I1115 02:00:31.229267 27751 eviction_manager.go:221] eviction manager: synchronize housekeeping
Nov 15 02:00:31 af867b kubelet[27751]: I1115 02:00:31.291198 27751 helpers.go:871] eviction manager: observations: signal=memory.available, available: 6606584Ki, capacity: 7393360Ki, time: 2017-11-15 02:00:22.740279886 +0000 UTC
Nov 15 02:00:31 af867b kubelet[27751]: I1115 02:00:31.291241 27751 helpers.go:871] eviction manager: observations: signal=nodefs.available, available: 7399360Ki, capacity: 10198Mi, time: 2017-11-15 02:00:22.740279886 +0000 UTC
Nov 15 02:00:31 af867b kubelet[27751]: I1115 02:00:31.291253 27751 helpers.go:871] eviction manager: observations: signal=nodefs.inodesFree, available: 10384791, capacity: 10208Ki, time: 2017-11-15 02:00:22.740279886 +0000 UTC
Nov 15 02:00:31 af867b kubelet[27751]: I1115 02:00:31.291263 27751 helpers.go:871] eviction manager: observations: signal=imagefs.available, available: 41037312Ki, capacity: 45Gi, time: 2017-11-15 02:00:22.740279886 +0000 UTC
Nov 15 02:00:31 af867b kubelet[27751]: I1115 02:00:31.291272 27751 helpers.go:873] eviction manager: observations: signal=allocatableMemory.available, available: 7038388Ki, capacity: 7393360Ki
Nov 15 02:00:31 af867b kubelet[27751]: I1115 02:00:31.291279 27751 helpers.go:873] eviction manager: observations: signal=allocatableNodeFs.available, available: 9624040228, capacity: 10198Mi
Nov 15 02:00:31 af867b kubelet[27751]: I1115 02:00:31.291298 27751 eviction_manager.go:325] eviction manager: no resources are starved
Nov 15 02:00:32 af867b kubelet[27751]: I1115 02:00:32.523518 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:00:32 af867b kubelet[27751]: I1115 02:00:32.959993 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 2379, Path: /health
Nov 15 02:00:32 af867b kubelet[27751]: I1115 02:00:32.960040 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:00:33 af867b kubelet[27751]: I1115 02:00:33.461409 27751 http.go:96] Probe succeeded for http://127.0.0.1:2379/health, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:00:33 GMT] Content-Length:[18] Content-Type:[text/plain; charset=utf-8]] 0xc420d82480 18 [] true false map[] 0xc4200dd700 <nil>}
Nov 15 02:00:33 af867b kubelet[27751]: I1115 02:00:33.461483 27751 prober.go:113] Liveness probe for "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250):etcd" succeeded
Nov 15 02:00:33 af867b kubelet[27751]: I1115 02:00:33.617291 27751 prober.go:160] HTTP-Probe Host: https://127.0.0.1, Port: 6443, Path: /healthz
Nov 15 02:00:33 af867b kubelet[27751]: I1115 02:00:33.617318 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:00:33 af867b kubelet[27751]: I1115 02:00:33.623565 27751 http.go:96] Probe succeeded for https://127.0.0.1:6443/healthz, Response: {200 OK 200 HTTP/2.0 2 0 map[Content-Type:[text/plain; charset=utf-8] Content-Length:[2] Date:[Wed, 15 Nov 2017 02:00:33 GMT]] 0xc4209c55c0 2 [] false false map[] 0xc420d4ac00 0xc421867c30}
Nov 15 02:00:33 af867b kubelet[27751]: I1115 02:00:33.623600 27751 prober.go:113] Liveness probe for "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d):kube-apiserver" succeeded
Nov 15 02:00:34 af867b kubelet[27751]: I1115 02:00:34.523633 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:00:34 af867b kubelet[27751]: I1115 02:00:34.790393 27751 kube_docker_client.go:330] Pulling image "weaveworks/weave-kube:2.0.5": "35577841e8d1: Downloading [===========> ] 2.391MB/10.09MB"
Nov 15 02:00:35 af867b kubelet[27751]: I1115 02:00:35.362720 27751 worker.go:164] Probe target container not found: weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242) - weave
Nov 15 02:00:35 af867b kubelet[27751]: W1115 02:00:35.846344 27751 cni.go:196] Unable to update cni config: No networks found in /etc/cni/net.d
Nov 15 02:00:35 af867b kubelet[27751]: I1115 02:00:35.846496 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 02:00:35 af867b kubelet[27751]: E1115 02:00:35.846518 27751 kubelet.go:2095] Container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 02:00:36 af867b kubelet[27751]: I1115 02:00:36.523462 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:00:37 af867b kubelet[27751]: I1115 02:00:37.506115 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10252, Path: /healthz
Nov 15 02:00:37 af867b kubelet[27751]: I1115 02:00:37.506140 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:00:37 af867b kubelet[27751]: I1115 02:00:37.506840 27751 http.go:96] Probe succeeded for http://127.0.0.1:10252/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:00:37 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc421b39c40 2 [] true false map[] 0xc420d4b100 <nil>}
Nov 15 02:00:37 af867b kubelet[27751]: I1115 02:00:37.506874 27751 prober.go:113] Liveness probe for "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9):kube-controller-manager" succeeded
Nov 15 02:00:37 af867b kubelet[27751]: I1115 02:00:37.681639 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10251, Path: /healthz
Nov 15 02:00:37 af867b kubelet[27751]: I1115 02:00:37.681667 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:00:37 af867b kubelet[27751]: I1115 02:00:37.682324 27751 http.go:96] Probe succeeded for http://127.0.0.1:10251/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Type:[text/plain; charset=utf-8] Date:[Wed, 15 Nov 2017 02:00:37 GMT] Content-Length:[2]] 0xc421c17160 2 [] true false map[] 0xc4200dd500 <nil>}
Nov 15 02:00:37 af867b kubelet[27751]: I1115 02:00:37.682360 27751 prober.go:113] Liveness probe for "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373):kube-scheduler" succeeded
Nov 15 02:00:38 af867b kubelet[27751]: I1115 02:00:38.523449 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:00:40 af867b kubelet[27751]: I1115 02:00:40.523834 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:00:40 af867b kubelet[27751]: W1115 02:00:40.847393 27751 cni.go:196] Unable to update cni config: No networks found in /etc/cni/net.d
Nov 15 02:00:40 af867b kubelet[27751]: I1115 02:00:40.847526 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 02:00:40 af867b kubelet[27751]: E1115 02:00:40.847559 27751 kubelet.go:2095] Container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 02:00:41 af867b kubelet[27751]: I1115 02:00:41.291487 27751 eviction_manager.go:221] eviction manager: synchronize housekeeping
Nov 15 02:00:41 af867b kubelet[27751]: I1115 02:00:41.336486 27751 helpers.go:871] eviction manager: observations: signal=memory.available, available: 6596220Ki, capacity: 7393360Ki, time: 2017-11-15 02:00:34.61443142 +0000 UTC
Nov 15 02:00:41 af867b kubelet[27751]: I1115 02:00:41.336524 27751 helpers.go:871] eviction manager: observations: signal=nodefs.available, available: 7415348Ki, capacity: 10198Mi, time: 2017-11-15 02:00:34.61443142 +0000 UTC
Nov 15 02:00:41 af867b kubelet[27751]: I1115 02:00:41.336535 27751 helpers.go:871] eviction manager: observations: signal=nodefs.inodesFree, available: 10384744, capacity: 10208Ki, time: 2017-11-15 02:00:34.61443142 +0000 UTC
Nov 15 02:00:41 af867b kubelet[27751]: I1115 02:00:41.336543 27751 helpers.go:871] eviction manager: observations: signal=imagefs.available, available: 39981Mi, capacity: 45Gi, time: 2017-11-15 02:00:34.61443142 +0000 UTC
Nov 15 02:00:41 af867b kubelet[27751]: I1115 02:00:41.336552 27751 helpers.go:873] eviction manager: observations: signal=allocatableMemory.available, available: 7038116Ki, capacity: 7393360Ki
Nov 15 02:00:41 af867b kubelet[27751]: I1115 02:00:41.336559 27751 helpers.go:873] eviction manager: observations: signal=allocatableNodeFs.available, available: 9624040228, capacity: 10198Mi
Nov 15 02:00:41 af867b kubelet[27751]: I1115 02:00:41.336577 27751 eviction_manager.go:325] eviction manager: no resources are starved
Nov 15 02:00:42 af867b kubelet[27751]: I1115 02:00:42.523536 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:00:42 af867b kubelet[27751]: I1115 02:00:42.959963 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 2379, Path: /health
Nov 15 02:00:42 af867b kubelet[27751]: I1115 02:00:42.959993 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:00:43 af867b kubelet[27751]: I1115 02:00:43.464519 27751 http.go:96] Probe succeeded for http://127.0.0.1:2379/health, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:00:43 GMT] Content-Length:[18] Content-Type:[text/plain; charset=utf-8]] 0xc421329ee0 18 [] true false map[] 0xc420a31900 <nil>}
Nov 15 02:00:43 af867b kubelet[27751]: I1115 02:00:43.464644 27751 prober.go:113] Liveness probe for "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250):etcd" succeeded
Nov 15 02:00:43 af867b kubelet[27751]: I1115 02:00:43.617323 27751 prober.go:160] HTTP-Probe Host: https://127.0.0.1, Port: 6443, Path: /healthz
Nov 15 02:00:43 af867b kubelet[27751]: I1115 02:00:43.617362 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:00:43 af867b kubelet[27751]: I1115 02:00:43.624374 27751 http.go:96] Probe succeeded for https://127.0.0.1:6443/healthz, Response: {200 OK 200 HTTP/2.0 2 0 map[Content-Type:[text/plain; charset=utf-8] Content-Length:[2] Date:[Wed, 15 Nov 2017 02:00:43 GMT]] 0xc420db2160 2 [] false false map[] 0xc420a31b00 0xc4212f5ce0}
Nov 15 02:00:43 af867b kubelet[27751]: I1115 02:00:43.624418 27751 prober.go:113] Liveness probe for "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d):kube-apiserver" succeeded
Nov 15 02:00:44 af867b kubelet[27751]: I1115 02:00:44.523474 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:00:44 af867b kubelet[27751]: I1115 02:00:44.790319 27751 kube_docker_client.go:330] Pulling image "weaveworks/weave-kube:2.0.5": "35577841e8d1: Downloading [===================> ] 3.997MB/10.09MB"
Nov 15 02:00:45 af867b kubelet[27751]: I1115 02:00:45.362891 27751 worker.go:164] Probe target container not found: weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242) - weave
Nov 15 02:00:45 af867b kubelet[27751]: I1115 02:00:45.527527 27751 kubelet.go:1890] SyncLoop (SYNC): 1 pods; kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)
Nov 15 02:00:45 af867b kubelet[27751]: I1115 02:00:45.527626 27751 kubelet_pods.go:1284] Generating status for "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:00:45 af867b kubelet[27751]: I1115 02:00:45.527824 27751 status_manager.go:325] Ignoring same status for pod "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)", status: {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:59:18 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:59:20 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:59:20 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP:10.196.65.210 StartTime:2017-11-15 01:59:18 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:kube-proxy State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 01:59:20 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:gcr.io/google_containers/kube-proxy-amd64:v1.8.3 ImageID:docker-pullable://gcr.io/google_containers/kube-proxy-amd64@sha256:63210bc9690144d41126a646caf03a3d76ddc6d06b8bad119d468193c3e90c24 ContainerID:docker://7ea397ec6048e25ce044b9edba43fe0ef2ed54803c9f3516c7b6e780d2c60a34}] QOSClass:BestEffort}
Nov 15 02:00:45 af867b kubelet[27751]: I1115 02:00:45.528000 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:00:45 af867b kubelet[27751]: I1115 02:00:45.540039 27751 configmap.go:187] Setting up volume kube-proxy for pod 9729c03a-c9a8-11e7-89f4-c6b053eac242 at /var/lib/kubelet/pods/9729c03a-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~configmap/kube-proxy
Nov 15 02:00:45 af867b kubelet[27751]: I1115 02:00:45.540392 27751 secret.go:186] Setting up volume kube-proxy-token-gqhfs for pod 9729c03a-c9a8-11e7-89f4-c6b053eac242 at /var/lib/kubelet/pods/9729c03a-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/kube-proxy-token-gqhfs
Nov 15 02:00:45 af867b kubelet[27751]: I1115 02:00:45.542843 27751 secret.go:217] Received secret kube-system/kube-proxy-token-gqhfs containing (3) pieces of data, 1904 total bytes
Nov 15 02:00:45 af867b kubelet[27751]: I1115 02:00:45.542983 27751 atomic_writer.go:142] pod kube-system/kube-proxy-nnsjf volume kube-proxy-token-gqhfs: no update required for target directory /var/lib/kubelet/pods/9729c03a-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/kube-proxy-token-gqhfs
Nov 15 02:00:45 af867b kubelet[27751]: I1115 02:00:45.543073 27751 configmap.go:218] Received configMap kube-system/kube-proxy containing (1) pieces of data, 407 total bytes
Nov 15 02:00:45 af867b kubelet[27751]: I1115 02:00:45.543151 27751 atomic_writer.go:142] pod kube-system/kube-proxy-nnsjf volume kube-proxy: no update required for target directory /var/lib/kubelet/pods/9729c03a-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~configmap/kube-proxy
Nov 15 02:00:45 af867b kubelet[27751]: I1115 02:00:45.828275 27751 volume_manager.go:366] All volumes are attached and mounted for pod "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:00:45 af867b kubelet[27751]: I1115 02:00:45.828410 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:false CreateSandbox:false SandboxID:91510886f3beb621e5d04309d502352ff78392e46405f21633802da4f7047069 Attempt:0 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:00:45 af867b kubelet[27751]: W1115 02:00:45.849136 27751 cni.go:196] Unable to update cni config: No networks found in /etc/cni/net.d
Nov 15 02:00:45 af867b kubelet[27751]: I1115 02:00:45.849285 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 02:00:45 af867b kubelet[27751]: E1115 02:00:45.849303 27751 kubelet.go:2095] Container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 02:00:46 af867b kubelet[27751]: I1115 02:00:46.523460 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:00:47 af867b kubelet[27751]: I1115 02:00:47.506130 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10252, Path: /healthz
Nov 15 02:00:47 af867b kubelet[27751]: I1115 02:00:47.506165 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:00:47 af867b kubelet[27751]: I1115 02:00:47.507081 27751 http.go:96] Probe succeeded for http://127.0.0.1:10252/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Length:[2] Content-Type:[text/plain; charset=utf-8] Date:[Wed, 15 Nov 2017 02:00:47 GMT]] 0xc421a53d80 2 [] true false map[] 0xc42110ae00 <nil>}
Nov 15 02:00:47 af867b kubelet[27751]: I1115 02:00:47.507119 27751 prober.go:113] Liveness probe for "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9):kube-controller-manager" succeeded
Nov 15 02:00:47 af867b kubelet[27751]: I1115 02:00:47.684180 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10251, Path: /healthz
Nov 15 02:00:47 af867b kubelet[27751]: I1115 02:00:47.684225 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:00:47 af867b kubelet[27751]: I1115 02:00:47.685836 27751 http.go:96] Probe succeeded for http://127.0.0.1:10251/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:00:47 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc4219eda60 2 [] true false map[] 0xc420afcd00 <nil>}
Nov 15 02:00:47 af867b kubelet[27751]: I1115 02:00:47.685893 27751 prober.go:113] Liveness probe for "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373):kube-scheduler" succeeded
Nov 15 02:00:48 af867b kubelet[27751]: I1115 02:00:48.523431 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.523449 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.528897 27751 qos_container_manager_linux.go:320] [ContainerManager]: Updated QoS cgroup configuration
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.549933 27751 kubelet.go:1222] Container garbage collection succeeded
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.825834 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/u01-applicationSpace.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.825880 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/u01-applicationSpace.mount", but ignoring.
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.825894 27751 manager.go:901] ignoring container "/system.slice/u01-applicationSpace.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.825903 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/run-docker-netns-default.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.825911 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/run-docker-netns-default.mount", but ignoring.
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.825921 27751 manager.go:901] ignoring container "/system.slice/run-docker-netns-default.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.825928 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-9d97c593c8a0b1ef2442048619ba4466fad52c408e686c9285a48450f3a6669f.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.825938 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-9d97c593c8a0b1ef2442048619ba4466fad52c408e686c9285a48450f3a6669f.mount", but ignoring.
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.825948 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-9d97c593c8a0b1ef2442048619ba4466fad52c408e686c9285a48450f3a6669f.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.825957 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/boot.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.825964 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/boot.mount", but ignoring.
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.825972 27751 manager.go:901] ignoring container "/system.slice/boot.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.825979 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-cd27b64716d8032aa3ea595ec121a222dc749f2f903149a780eb711106fb0c53.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.825988 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-cd27b64716d8032aa3ea595ec121a222dc749f2f903149a780eb711106fb0c53.mount", but ignoring.
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.825998 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-cd27b64716d8032aa3ea595ec121a222dc749f2f903149a780eb711106fb0c53.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826006 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-kubelet-pods-9729c03a\\x2dc9a8\\x2d11e7\\x2d89f4\\x2dc6b053eac242-volumes-kubernetes.io\\x7esecret-kube\\x2dproxy\\x2dtoken\\x2dgqhfs.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826026 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-kubelet-pods-9729c03a\\x2dc9a8\\x2d11e7\\x2d89f4\\x2dc6b053eac242-volumes-kubernetes.io\\x7esecret-kube\\x2dproxy\\x2dtoken\\x2dgqhfs.mount", but ignoring.
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826040 27751 manager.go:901] ignoring container "/system.slice/var-lib-kubelet-pods-9729c03a\\x2dc9a8\\x2d11e7\\x2d89f4\\x2dc6b053eac242-volumes-kubernetes.io\\x7esecret-kube\\x2dproxy\\x2dtoken\\x2dgqhfs.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826051 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/-.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826058 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/-.mount", but ignoring.
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826067 27751 manager.go:901] ignoring container "/system.slice/-.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826073 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-439f518d46ec0d596518b7c17c503c52a85dd4da21d8a56ec174d5b6b3b98e03-shm.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826082 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-439f518d46ec0d596518b7c17c503c52a85dd4da21d8a56ec174d5b6b3b98e03-shm.mount", but ignoring.
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826092 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-containers-439f518d46ec0d596518b7c17c503c52a85dd4da21d8a56ec174d5b6b3b98e03-shm.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826100 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-fe18d337ce42bd3a4d2aacb1349be028274681d8073fbe5d6377946e815fa810-shm.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826109 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-fe18d337ce42bd3a4d2aacb1349be028274681d8073fbe5d6377946e815fa810-shm.mount", but ignoring.
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826120 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-containers-fe18d337ce42bd3a4d2aacb1349be028274681d8073fbe5d6377946e815fa810-shm.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826128 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-d5076ad0e9fb270d1b8c4ff7cdadbf32d6b1e30dc42ae24dfbc5cb01bb5aa934-shm.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826137 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-d5076ad0e9fb270d1b8c4ff7cdadbf32d6b1e30dc42ae24dfbc5cb01bb5aa934-shm.mount", but ignoring.
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826147 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-containers-d5076ad0e9fb270d1b8c4ff7cdadbf32d6b1e30dc42ae24dfbc5cb01bb5aa934-shm.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826155 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-82408ec054eef4c0bce3519663bc7e5cbaff7a4fea1d8b7490fa3cc5a74bac60.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826361 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-82408ec054eef4c0bce3519663bc7e5cbaff7a4fea1d8b7490fa3cc5a74bac60.mount", but ignoring.
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826384 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-82408ec054eef4c0bce3519663bc7e5cbaff7a4fea1d8b7490fa3cc5a74bac60.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826394 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/run-user-1000.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826402 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/run-user-1000.mount", but ignoring.
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826411 27751 manager.go:901] ignoring container "/system.slice/run-user-1000.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826417 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-984a35c7cf6d04c0af5fcd260872a7b6556aa9cd8b38bc1a1cf635972c1d6511.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826426 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-984a35c7cf6d04c0af5fcd260872a7b6556aa9cd8b38bc1a1cf635972c1d6511.mount", but ignoring.
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826437 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-984a35c7cf6d04c0af5fcd260872a7b6556aa9cd8b38bc1a1cf635972c1d6511.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826445 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/sys-kernel-debug.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826452 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/sys-kernel-debug.mount", but ignoring.
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826461 27751 manager.go:901] ignoring container "/system.slice/sys-kernel-debug.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826468 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/dev-hugepages.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826476 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/dev-hugepages.mount", but ignoring.
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826484 27751 manager.go:901] ignoring container "/system.slice/dev-hugepages.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826490 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-23d8892c53722cfb4d3aa6c5b126ad2674a7b8041b4873d30b98faefad6f5b44.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826499 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-23d8892c53722cfb4d3aa6c5b126ad2674a7b8041b4873d30b98faefad6f5b44.mount", but ignoring.
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826510 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-23d8892c53722cfb4d3aa6c5b126ad2674a7b8041b4873d30b98faefad6f5b44.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826518 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-8c430fd9b74591760bf92ca717db0989293743473d027f75b1dfded4d0661504.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826528 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-8c430fd9b74591760bf92ca717db0989293743473d027f75b1dfded4d0661504.mount", but ignoring.
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826539 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-8c430fd9b74591760bf92ca717db0989293743473d027f75b1dfded4d0661504.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826547 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826555 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper.mount", but ignoring.
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826563 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826570 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/dev-mqueue.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826577 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/dev-mqueue.mount", but ignoring.
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826586 27751 manager.go:901] ignoring container "/system.slice/dev-mqueue.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826594 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-91510886f3beb621e5d04309d502352ff78392e46405f21633802da4f7047069-shm.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826603 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-91510886f3beb621e5d04309d502352ff78392e46405f21633802da4f7047069-shm.mount", but ignoring.
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826614 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-containers-91510886f3beb621e5d04309d502352ff78392e46405f21633802da4f7047069-shm.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826622 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-7d2ebdadb8c541a4839058dd46b491ef9e14209821b47e7d86b539de402fbef5.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826632 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-7d2ebdadb8c541a4839058dd46b491ef9e14209821b47e7d86b539de402fbef5.mount", but ignoring.
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826673 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-7d2ebdadb8c541a4839058dd46b491ef9e14209821b47e7d86b539de402fbef5.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826683 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-29c1400e918b28b9982788b4c8b07c46b1767efcd8b96000e44bc78da639994b.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826692 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-29c1400e918b28b9982788b4c8b07c46b1767efcd8b96000e44bc78da639994b.mount", but ignoring.
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826702 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-29c1400e918b28b9982788b4c8b07c46b1767efcd8b96000e44bc78da639994b.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826727 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-6ddc96a63fb65ccd1743152170b39274e0eb3e4ccc4948bf2c0f5a6130ee9ce7.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826738 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-6ddc96a63fb65ccd1743152170b39274e0eb3e4ccc4948bf2c0f5a6130ee9ce7.mount", but ignoring.
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826748 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-6ddc96a63fb65ccd1743152170b39274e0eb3e4ccc4948bf2c0f5a6130ee9ce7.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826757 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/sys-kernel-config.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826764 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/sys-kernel-config.mount", but ignoring.
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826772 27751 manager.go:901] ignoring container "/system.slice/sys-kernel-config.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826779 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-faffc5ce0cf0da39c68377dff0a611eb436ee74dfce11692aecd1e723311ef80.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826788 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-faffc5ce0cf0da39c68377dff0a611eb436ee74dfce11692aecd1e723311ef80.mount", but ignoring.
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826798 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-faffc5ce0cf0da39c68377dff0a611eb436ee74dfce11692aecd1e723311ef80.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826807 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/proc-xen.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826814 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/proc-xen.mount", but ignoring.
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826823 27751 manager.go:901] ignoring container "/system.slice/proc-xen.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826830 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-19b06ca14052c17f92bf03d5316fd7642fe4fee6bf193755575355be3e87a13d-shm.mount"
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826839 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-19b06ca14052c17f92bf03d5316fd7642fe4fee6bf193755575355be3e87a13d-shm.mount", but ignoring.
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.826849 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-containers-19b06ca14052c17f92bf03d5316fd7642fe4fee6bf193755575355be3e87a13d-shm.mount"
Nov 15 02:00:50 af867b kubelet[27751]: W1115 02:00:50.850448 27751 cni.go:196] Unable to update cni config: No networks found in /etc/cni/net.d
Nov 15 02:00:50 af867b kubelet[27751]: I1115 02:00:50.850612 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 02:00:50 af867b kubelet[27751]: E1115 02:00:50.850635 27751 kubelet.go:2095] Container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 02:00:51 af867b kubelet[27751]: I1115 02:00:51.336801 27751 eviction_manager.go:221] eviction manager: synchronize housekeeping
Nov 15 02:00:51 af867b kubelet[27751]: I1115 02:00:51.402631 27751 helpers.go:871] eviction manager: observations: signal=memory.available, available: 6593816Ki, capacity: 7393360Ki, time: 2017-11-15 02:00:47.870483701 +0000 UTC
Nov 15 02:00:51 af867b kubelet[27751]: I1115 02:00:51.402672 27751 helpers.go:871] eviction manager: observations: signal=nodefs.available, available: 7411340Ki, capacity: 10198Mi, time: 2017-11-15 02:00:47.870483701 +0000 UTC
Nov 15 02:00:51 af867b kubelet[27751]: I1115 02:00:51.402693 27751 helpers.go:871] eviction manager: observations: signal=nodefs.inodesFree, available: 10384744, capacity: 10208Ki, time: 2017-11-15 02:00:47.870483701 +0000 UTC
Nov 15 02:00:51 af867b kubelet[27751]: I1115 02:00:51.402702 27751 helpers.go:871] eviction manager: observations: signal=imagefs.available, available: 39981Mi, capacity: 45Gi, time: 2017-11-15 02:00:47.870483701 +0000 UTC
Nov 15 02:00:51 af867b kubelet[27751]: I1115 02:00:51.402774 27751 helpers.go:873] eviction manager: observations: signal=allocatableMemory.available, available: 7037856Ki, capacity: 7393360Ki
Nov 15 02:00:51 af867b kubelet[27751]: I1115 02:00:51.402783 27751 helpers.go:873] eviction manager: observations: signal=allocatableNodeFs.available, available: 9624040228, capacity: 10198Mi
Nov 15 02:00:51 af867b kubelet[27751]: I1115 02:00:51.402805 27751 eviction_manager.go:325] eviction manager: no resources are starved
Nov 15 02:00:52 af867b kubelet[27751]: I1115 02:00:52.523542 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:00:52 af867b kubelet[27751]: I1115 02:00:52.959978 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 2379, Path: /health
Nov 15 02:00:52 af867b kubelet[27751]: I1115 02:00:52.960009 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:00:53 af867b kubelet[27751]: I1115 02:00:53.461856 27751 http.go:96] Probe succeeded for http://127.0.0.1:2379/health, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:00:53 GMT] Content-Length:[18] Content-Type:[text/plain; charset=utf-8]] 0xc4209c4ee0 18 [] true false map[] 0xc4200dd500 <nil>}
Nov 15 02:00:53 af867b kubelet[27751]: I1115 02:00:53.461970 27751 prober.go:113] Liveness probe for "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250):etcd" succeeded
Nov 15 02:00:53 af867b kubelet[27751]: I1115 02:00:53.617356 27751 prober.go:160] HTTP-Probe Host: https://127.0.0.1, Port: 6443, Path: /healthz
Nov 15 02:00:53 af867b kubelet[27751]: I1115 02:00:53.617390 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:00:53 af867b kubelet[27751]: I1115 02:00:53.624871 27751 http.go:96] Probe succeeded for https://127.0.0.1:6443/healthz, Response: {200 OK 200 HTTP/2.0 2 0 map[Content-Type:[text/plain; charset=utf-8] Content-Length:[2] Date:[Wed, 15 Nov 2017 02:00:53 GMT]] 0xc420aa7380 2 [] false false map[] 0xc42110b400 0xc420cc6f20}
Nov 15 02:00:53 af867b kubelet[27751]: I1115 02:00:53.624917 27751 prober.go:113] Liveness probe for "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d):kube-apiserver" succeeded
Nov 15 02:00:54 af867b kubelet[27751]: I1115 02:00:54.523438 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:00:54 af867b kubelet[27751]: I1115 02:00:54.790402 27751 kube_docker_client.go:330] Pulling image "weaveworks/weave-kube:2.0.5": "35577841e8d1: Downloading [====================================> ] 7.437MB/10.09MB"
Nov 15 02:00:55 af867b kubelet[27751]: I1115 02:00:55.362734 27751 worker.go:164] Probe target container not found: weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242) - weave
Nov 15 02:00:55 af867b kubelet[27751]: W1115 02:00:55.851696 27751 cni.go:196] Unable to update cni config: No networks found in /etc/cni/net.d
Nov 15 02:00:55 af867b kubelet[27751]: I1115 02:00:55.851903 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 02:00:55 af867b kubelet[27751]: E1115 02:00:55.851927 27751 kubelet.go:2095] Container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 02:00:56 af867b kubelet[27751]: I1115 02:00:56.523491 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:00:57 af867b kubelet[27751]: I1115 02:00:57.506133 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10252, Path: /healthz
Nov 15 02:00:57 af867b kubelet[27751]: I1115 02:00:57.506169 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:00:57 af867b kubelet[27751]: I1115 02:00:57.507077 27751 http.go:96] Probe succeeded for http://127.0.0.1:10252/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:00:57 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc4219866a0 2 [] true false map[] 0xc421178600 <nil>}
Nov 15 02:00:57 af867b kubelet[27751]: I1115 02:00:57.507121 27751 prober.go:113] Liveness probe for "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9):kube-controller-manager" succeeded
Nov 15 02:00:57 af867b kubelet[27751]: I1115 02:00:57.681659 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10251, Path: /healthz
Nov 15 02:00:57 af867b kubelet[27751]: I1115 02:00:57.681697 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:00:57 af867b kubelet[27751]: I1115 02:00:57.683085 27751 http.go:96] Probe succeeded for http://127.0.0.1:10251/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:00:57 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc421986900 2 [] true false map[] 0xc421178800 <nil>}
Nov 15 02:00:57 af867b kubelet[27751]: I1115 02:00:57.683142 27751 prober.go:113] Liveness probe for "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373):kube-scheduler" succeeded
Nov 15 02:00:58 af867b kubelet[27751]: I1115 02:00:58.523549 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:01:00 af867b kubelet[27751]: I1115 02:01:00.524535 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:01:00 af867b kubelet[27751]: I1115 02:01:00.854771 27751 kube_docker_client.go:333] Stop pulling image "weaveworks/weave-kube:2.0.5": "Status: Downloaded newer image for weaveworks/weave-kube:2.0.5"
Nov 15 02:01:00 af867b kubelet[27751]: W1115 02:01:00.857891 27751 cni.go:196] Unable to update cni config: No networks found in /etc/cni/net.d
Nov 15 02:01:00 af867b kubelet[27751]: I1115 02:01:00.858016 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 02:01:00 af867b kubelet[27751]: E1115 02:01:00.858044 27751 kubelet.go:2095] Container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
Nov 15 02:01:00 af867b kubelet[27751]: I1115 02:01:00.858225 27751 kuberuntime_container.go:100] Generating ref for container weave: &v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"weave-net-rg7fn", UID:"b77b0858-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"381", FieldPath:"spec.containers{weave}"}
Nov 15 02:01:00 af867b kubelet[27751]: I1115 02:01:00.858263 27751 container_manager_linux.go:634] Calling devicePluginHandler AllocateDevices
Nov 15 02:01:00 af867b kubelet[27751]: I1115 02:01:00.858326 27751 kubelet_pods.go:123] container: kube-system/weave-net-rg7fn/weave podIP: "10.196.65.210" creating hosts mount: true
Nov 15 02:01:00 af867b kubelet[27751]: I1115 02:01:00.861488 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"weave-net-rg7fn", UID:"b77b0858-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"381", FieldPath:"spec.containers{weave}"}): type: 'Normal' reason: 'Pulled' Successfully pulled image "weaveworks/weave-kube:2.0.5"
Nov 15 02:01:00 af867b kubelet[27751]: I1115 02:01:00.864452 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/burstable/podb77b0858-c9a8-11e7-89f4-c6b053eac242"
Nov 15 02:01:01 af867b kubelet[27751]: I1115 02:01:01.147880 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"weave-net-rg7fn", UID:"b77b0858-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"381", FieldPath:"spec.containers{weave}"}): type: 'Normal' reason: 'Created' Created container
Nov 15 02:01:01 af867b kubelet[27751]: I1115 02:01:01.341481 27751 factory.go:112] Using factory "docker" for container "/kubepods/burstable/podb77b0858-c9a8-11e7-89f4-c6b053eac242/f4fa9d4e3be52b17cedfe3b19a153e4584737c95ada75d64ea13225dde740f40"
Nov 15 02:01:01 af867b kubelet[27751]: I1115 02:01:01.344400 27751 kuberuntime_manager.go:705] Creating container &Container{Name:weave-npc,Image:weaveworks/weave-npc:2.0.5,Command:[],Args:[],WorkingDir:,Ports:[],Env:[{HOSTNAME EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,}}],Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {<nil>} 10m DecimalSI},},},VolumeMounts:[{xtables-lock false /run/xtables.lock <nil>} {weave-net-token-rn6j7 true /var/run/secrets/kubernetes.io/serviceaccount <nil>}],LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[],TerminationMessagePolicy:File,} in pod weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242)
Nov 15 02:01:01 af867b kubelet[27751]: I1115 02:01:01.344601 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"weave-net-rg7fn", UID:"b77b0858-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"381", FieldPath:"spec.containers{weave}"}): type: 'Normal' reason: 'Started' Started container
Nov 15 02:01:01 af867b kubelet[27751]: I1115 02:01:01.348553 27751 kuberuntime_image.go:46] Pulling image "weaveworks/weave-npc:2.0.5" without credentials
Nov 15 02:01:01 af867b kubelet[27751]: I1115 02:01:01.348644 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"weave-net-rg7fn", UID:"b77b0858-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"381", FieldPath:"spec.containers{weave-npc}"}): type: 'Normal' reason: 'Pulling' pulling image "weaveworks/weave-npc:2.0.5"
Nov 15 02:01:01 af867b kubelet[27751]: I1115 02:01:01.350246 27751 manager.go:932] Added container: "/kubepods/burstable/podb77b0858-c9a8-11e7-89f4-c6b053eac242/f4fa9d4e3be52b17cedfe3b19a153e4584737c95ada75d64ea13225dde740f40" (aliases: [k8s_weave_weave-net-rg7fn_kube-system_b77b0858-c9a8-11e7-89f4-c6b053eac242_0 f4fa9d4e3be52b17cedfe3b19a153e4584737c95ada75d64ea13225dde740f40], namespace: "docker")
Nov 15 02:01:01 af867b kubelet[27751]: I1115 02:01:01.350399 27751 handler.go:325] Added event &{/kubepods/burstable/podb77b0858-c9a8-11e7-89f4-c6b053eac242/f4fa9d4e3be52b17cedfe3b19a153e4584737c95ada75d64ea13225dde740f40 2017-11-15 02:01:01.256194711 +0000 UTC containerCreation {<nil>}}
Nov 15 02:01:01 af867b kubelet[27751]: I1115 02:01:01.350437 27751 container.go:409] Start housekeeping for container "/kubepods/burstable/podb77b0858-c9a8-11e7-89f4-c6b053eac242/f4fa9d4e3be52b17cedfe3b19a153e4584737c95ada75d64ea13225dde740f40"
Nov 15 02:01:01 af867b kubelet[27751]: I1115 02:01:01.403018 27751 eviction_manager.go:221] eviction manager: synchronize housekeeping
Nov 15 02:01:01 af867b kubelet[27751]: I1115 02:01:01.451478 27751 helpers.go:871] eviction manager: observations: signal=nodefs.inodesFree, available: 10384744, capacity: 10208Ki, time: 2017-11-15 02:00:47.870483701 +0000 UTC
Nov 15 02:01:01 af867b kubelet[27751]: I1115 02:01:01.451515 27751 helpers.go:871] eviction manager: observations: signal=imagefs.available, available: 39981Mi, capacity: 45Gi, time: 2017-11-15 02:00:47.870483701 +0000 UTC
Nov 15 02:01:01 af867b kubelet[27751]: I1115 02:01:01.451525 27751 helpers.go:873] eviction manager: observations: signal=allocatableMemory.available, available: 7035680Ki, capacity: 7393360Ki
Nov 15 02:01:01 af867b kubelet[27751]: I1115 02:01:01.451532 27751 helpers.go:873] eviction manager: observations: signal=allocatableNodeFs.available, available: 9624040228, capacity: 10198Mi
Nov 15 02:01:01 af867b kubelet[27751]: I1115 02:01:01.451539 27751 helpers.go:871] eviction manager: observations: signal=memory.available, available: 6593816Ki, capacity: 7393360Ki, time: 2017-11-15 02:00:47.870483701 +0000 UTC
Nov 15 02:01:01 af867b kubelet[27751]: I1115 02:01:01.451546 27751 helpers.go:871] eviction manager: observations: signal=nodefs.available, available: 7411340Ki, capacity: 10198Mi, time: 2017-11-15 02:00:47.870483701 +0000 UTC
Nov 15 02:01:01 af867b kubelet[27751]: I1115 02:01:01.451567 27751 eviction_manager.go:325] eviction manager: no resources are starved
Nov 15 02:01:01 af867b kubelet[27751]: I1115 02:01:01.923699 27751 generic.go:146] GenericPLEG: b77b0858-c9a8-11e7-89f4-c6b053eac242/f4fa9d4e3be52b17cedfe3b19a153e4584737c95ada75d64ea13225dde740f40: non-existent -> running
Nov 15 02:01:01 af867b kubelet[27751]: I1115 02:01:01.924574 27751 kuberuntime_manager.go:833] getSandboxIDByPodUID got sandbox IDs ["be14b0552100f929052c8e94e1b7cf587c4218e321d811314e58f2cf312f1c0f"] for pod "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:01:01 af867b kubelet[27751]: I1115 02:01:01.929035 27751 generic.go:345] PLEG: Write status for weave-net-rg7fn/kube-system: &container.PodStatus{ID:"b77b0858-c9a8-11e7-89f4-c6b053eac242", Name:"weave-net-rg7fn", Namespace:"kube-system", IP:"", ContainerStatuses:[]*container.ContainerStatus{(*container.ContainerStatus)(0xc42066eee0)}, SandboxStatuses:[]*runtime.PodSandboxStatus{(*runtime.PodSandboxStatus)(0xc42126f1d0)}} (err: <nil>)
Nov 15 02:01:01 af867b kubelet[27751]: I1115 02:01:01.929097 27751 kubelet.go:1871] SyncLoop (PLEG): "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242)", event: &pleg.PodLifecycleEvent{ID:"b77b0858-c9a8-11e7-89f4-c6b053eac242", Type:"ContainerStarted", Data:"f4fa9d4e3be52b17cedfe3b19a153e4584737c95ada75d64ea13225dde740f40"}
Nov 15 02:01:02 af867b kubelet[27751]: I1115 02:01:02.523539 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:01:02 af867b kubelet[27751]: I1115 02:01:02.959992 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 2379, Path: /health
Nov 15 02:01:02 af867b kubelet[27751]: I1115 02:01:02.960032 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:01:03 af867b kubelet[27751]: I1115 02:01:03.211105 27751 http.go:96] Probe succeeded for http://127.0.0.1:2379/health, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:01:03 GMT] Content-Length:[18] Content-Type:[text/plain; charset=utf-8]] 0xc420bb0400 18 [] true false map[] 0xc420a31900 <nil>}
Nov 15 02:01:03 af867b kubelet[27751]: I1115 02:01:03.211175 27751 prober.go:113] Liveness probe for "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250):etcd" succeeded
Nov 15 02:01:03 af867b kubelet[27751]: I1115 02:01:03.617417 27751 prober.go:160] HTTP-Probe Host: https://127.0.0.1, Port: 6443, Path: /healthz
Nov 15 02:01:03 af867b kubelet[27751]: I1115 02:01:03.617482 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:01:03 af867b kubelet[27751]: I1115 02:01:03.632130 27751 http.go:96] Probe succeeded for https://127.0.0.1:6443/healthz, Response: {200 OK 200 HTTP/2.0 2 0 map[Content-Type:[text/plain; charset=utf-8] Content-Length:[2] Date:[Wed, 15 Nov 2017 02:01:03 GMT]] 0xc420bd0100 2 [] false false map[] 0xc420a31b00 0xc421055ad0}
Nov 15 02:01:03 af867b kubelet[27751]: I1115 02:01:03.632179 27751 prober.go:113] Liveness probe for "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d):kube-apiserver" succeeded
Nov 15 02:01:04 af867b kubelet[27751]: I1115 02:01:04.523448 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:01:05 af867b kubelet[27751]: I1115 02:01:05.362811 27751 worker.go:164] Probe target container not found: weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242) - weave
Nov 15 02:01:05 af867b kubelet[27751]: I1115 02:01:05.859700 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=true reason: message:
Nov 15 02:01:06 af867b kubelet[27751]: I1115 02:01:06.523580 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:01:07 af867b kubelet[27751]: I1115 02:01:07.467368 27751 kubelet_node_status.go:443] Recording NodeReady event message for node af867b
Nov 15 02:01:07 af867b kubelet[27751]: I1115 02:01:07.471348 27751 server.go:227] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"af867b", UID:"af867b", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NodeReady' Node af867b status is now: NodeReady
Nov 15 02:01:07 af867b kubelet[27751]: I1115 02:01:07.506146 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10252, Path: /healthz
Nov 15 02:01:07 af867b kubelet[27751]: I1115 02:01:07.506164 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:01:07 af867b kubelet[27751]: I1115 02:01:07.507102 27751 http.go:96] Probe succeeded for http://127.0.0.1:10252/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:01:07 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc421db46a0 2 [] true false map[] 0xc421179a00 <nil>}
Nov 15 02:01:07 af867b kubelet[27751]: I1115 02:01:07.507141 27751 prober.go:113] Liveness probe for "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9):kube-controller-manager" succeeded
Nov 15 02:01:07 af867b kubelet[27751]: I1115 02:01:07.681617 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10251, Path: /healthz
Nov 15 02:01:07 af867b kubelet[27751]: I1115 02:01:07.681652 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:01:07 af867b kubelet[27751]: I1115 02:01:07.684314 27751 http.go:96] Probe succeeded for http://127.0.0.1:10251/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:01:07 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc421cff660 2 [] true false map[] 0xc421179c00 <nil>}
Nov 15 02:01:07 af867b kubelet[27751]: I1115 02:01:07.684356 27751 prober.go:113] Liveness probe for "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373):kube-scheduler" succeeded
Nov 15 02:01:08 af867b kubelet[27751]: I1115 02:01:08.523515 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:01:10 af867b kubelet[27751]: I1115 02:01:10.523531 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:01:10 af867b kubelet[27751]: I1115 02:01:10.861176 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=true reason: message:
Nov 15 02:01:11 af867b kubelet[27751]: I1115 02:01:11.451816 27751 eviction_manager.go:221] eviction manager: synchronize housekeeping
Nov 15 02:01:11 af867b kubelet[27751]: I1115 02:01:11.508453 27751 helpers.go:871] eviction manager: observations: signal=memory.available, available: 6533196Ki, capacity: 7393360Ki, time: 2017-11-15 02:01:06.674507567 +0000 UTC
Nov 15 02:01:11 af867b kubelet[27751]: I1115 02:01:11.508502 27751 helpers.go:871] eviction manager: observations: signal=nodefs.available, available: 7405248Ki, capacity: 10198Mi, time: 2017-11-15 02:01:06.674507567 +0000 UTC
Nov 15 02:01:11 af867b kubelet[27751]: I1115 02:01:11.508520 27751 helpers.go:871] eviction manager: observations: signal=nodefs.inodesFree, available: 10384695, capacity: 10208Ki, time: 2017-11-15 02:01:06.674507567 +0000 UTC
Nov 15 02:01:11 af867b kubelet[27751]: I1115 02:01:11.508531 27751 helpers.go:871] eviction manager: observations: signal=imagefs.available, available: 40868352Ki, capacity: 45Gi, time: 2017-11-15 02:01:06.674507567 +0000 UTC
Nov 15 02:01:11 af867b kubelet[27751]: I1115 02:01:11.508540 27751 helpers.go:873] eviction manager: observations: signal=allocatableMemory.available, available: 7035136Ki, capacity: 7393360Ki
Nov 15 02:01:11 af867b kubelet[27751]: I1115 02:01:11.508547 27751 helpers.go:873] eviction manager: observations: signal=allocatableNodeFs.available, available: 9624040228, capacity: 10198Mi
Nov 15 02:01:11 af867b kubelet[27751]: I1115 02:01:11.508570 27751 eviction_manager.go:325] eviction manager: no resources are starved
Nov 15 02:01:12 af867b kubelet[27751]: I1115 02:01:12.523551 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:01:12 af867b kubelet[27751]: I1115 02:01:12.709475 27751 kube_docker_client.go:330] Pulling image "weaveworks/weave-npc:2.0.5": "a2592a033c5d: Downloading [=======================> ] 5.291MB/11.27MB"
Nov 15 02:01:12 af867b kubelet[27751]: I1115 02:01:12.960044 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 2379, Path: /health
Nov 15 02:01:12 af867b kubelet[27751]: I1115 02:01:12.960101 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:01:13 af867b kubelet[27751]: I1115 02:01:13.213292 27751 http.go:96] Probe succeeded for http://127.0.0.1:2379/health, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:01:13 GMT] Content-Length:[18] Content-Type:[text/plain; charset=utf-8]] 0xc4212e0b60 18 [] true false map[] 0xc421178f00 <nil>}
Nov 15 02:01:13 af867b kubelet[27751]: I1115 02:01:13.213390 27751 prober.go:113] Liveness probe for "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250):etcd" succeeded
Nov 15 02:01:13 af867b kubelet[27751]: I1115 02:01:13.617400 27751 prober.go:160] HTTP-Probe Host: https://127.0.0.1, Port: 6443, Path: /healthz
Nov 15 02:01:13 af867b kubelet[27751]: I1115 02:01:13.617462 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:01:13 af867b kubelet[27751]: I1115 02:01:13.631317 27751 http.go:96] Probe succeeded for https://127.0.0.1:6443/healthz, Response: {200 OK 200 HTTP/2.0 2 0 map[Content-Type:[text/plain; charset=utf-8] Content-Length:[2] Date:[Wed, 15 Nov 2017 02:01:13 GMT]] 0xc420f64940 2 [] false false map[] 0xc420ee2300 0xc4213f0b00}
Nov 15 02:01:13 af867b kubelet[27751]: I1115 02:01:13.631374 27751 prober.go:113] Liveness probe for "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d):kube-apiserver" succeeded
Nov 15 02:01:14 af867b kubelet[27751]: I1115 02:01:14.523544 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:01:15 af867b kubelet[27751]: I1115 02:01:15.362759 27751 worker.go:164] Probe target container not found: weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242) - weave
Nov 15 02:01:15 af867b kubelet[27751]: I1115 02:01:15.862921 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=true reason: message:
Nov 15 02:01:16 af867b kubelet[27751]: I1115 02:01:16.524862 27751 kubelet.go:1890] SyncLoop (SYNC): 1 pods; kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)
Nov 15 02:01:16 af867b kubelet[27751]: I1115 02:01:16.524927 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:01:16 af867b kubelet[27751]: I1115 02:01:16.528595 27751 kubelet_pods.go:1284] Generating status for "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)"
Nov 15 02:01:16 af867b kubelet[27751]: I1115 02:01:16.528816 27751 status_manager.go:325] Ignoring same status for pod "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)", status: {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:58 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP:10.196.65.210 StartTime:2017-11-15 01:58:55 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:kube-controller-manager State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 01:58:57 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:gcr.io/google_containers/kube-controller-manager-amd64:v1.8.3 ImageID:docker-pullable://gcr.io/google_containers/kube-controller-manager-amd64@sha256:b6b633e3e107761d38fceb200f01bf552c51f65e3524b0aafc1a7710afff07be ContainerID:docker://272af7e3b2b1a203250d154349fdf77f296d7b7f65ce2c77b6b3a94e53dba356}] QOSClass:Burstable}
Nov 15 02:01:16 af867b kubelet[27751]: I1115 02:01:16.528971 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)"
Nov 15 02:01:16 af867b kubelet[27751]: I1115 02:01:16.829285 27751 volume_manager.go:366] All volumes are attached and mounted for pod "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)"
Nov 15 02:01:16 af867b kubelet[27751]: I1115 02:01:16.829499 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:false CreateSandbox:false SandboxID:439f518d46ec0d596518b7c17c503c52a85dd4da21d8a56ec174d5b6b3b98e03 Attempt:0 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)"
Nov 15 02:01:17 af867b kubelet[27751]: I1115 02:01:17.506098 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10252, Path: /healthz
Nov 15 02:01:17 af867b kubelet[27751]: I1115 02:01:17.506126 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:01:17 af867b kubelet[27751]: I1115 02:01:17.507335 27751 http.go:96] Probe succeeded for http://127.0.0.1:10252/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:01:17 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc421deeee0 2 [] true false map[] 0xc420afd400 <nil>}
Nov 15 02:01:17 af867b kubelet[27751]: I1115 02:01:17.507377 27751 prober.go:113] Liveness probe for "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9):kube-controller-manager" succeeded
Nov 15 02:01:17 af867b kubelet[27751]: I1115 02:01:17.681635 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10251, Path: /healthz
Nov 15 02:01:17 af867b kubelet[27751]: I1115 02:01:17.681667 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:01:17 af867b kubelet[27751]: I1115 02:01:17.682906 27751 http.go:96] Probe succeeded for http://127.0.0.1:10251/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:01:17 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc421c7b960 2 [] true false map[] 0xc420afd600 <nil>}
Nov 15 02:01:17 af867b kubelet[27751]: I1115 02:01:17.682960 27751 prober.go:113] Liveness probe for "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373):kube-scheduler" succeeded
Nov 15 02:01:18 af867b kubelet[27751]: I1115 02:01:18.523481 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:01:20 af867b kubelet[27751]: I1115 02:01:20.523458 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:01:20 af867b kubelet[27751]: I1115 02:01:20.866415 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=true reason: message:
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.508856 27751 eviction_manager.go:221] eviction manager: synchronize housekeeping
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.554574 27751 helpers.go:871] eviction manager: observations: signal=nodefs.available, available: 7401080Ki, capacity: 10198Mi, time: 2017-11-15 02:01:19.237424468 +0000 UTC
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.554620 27751 helpers.go:871] eviction manager: observations: signal=nodefs.inodesFree, available: 10384695, capacity: 10208Ki, time: 2017-11-15 02:01:19.237424468 +0000 UTC
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.554632 27751 helpers.go:871] eviction manager: observations: signal=imagefs.available, available: 40868352Ki, capacity: 45Gi, time: 2017-11-15 02:01:19.237424468 +0000 UTC
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.554643 27751 helpers.go:873] eviction manager: observations: signal=allocatableMemory.available, available: 6976916Ki, capacity: 7393360Ki
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.554651 27751 helpers.go:873] eviction manager: observations: signal=allocatableNodeFs.available, available: 9624040228, capacity: 10198Mi
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.554658 27751 helpers.go:871] eviction manager: observations: signal=memory.available, available: 6532960Ki, capacity: 7393360Ki, time: 2017-11-15 02:01:19.237424468 +0000 UTC
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.554680 27751 eviction_manager.go:325] eviction manager: no resources are starved
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.609382 27751 config.go:282] Setting pods for source api
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.610592 27751 config.go:404] Receiving a new pod "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.610905 27751 kubelet.go:1837] SyncLoop (ADD, "api"): "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.611102 27751 kubelet_pods.go:1284] Generating status for "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.611478 27751 qos_container_manager_linux.go:320] [ContainerManager]: Updated QoS cgroup configuration
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.613917 27751 factory.go:116] Factory "docker" was unable to handle container "/kubepods/burstable/pod97270c63-c9a8-11e7-89f4-c6b053eac242"
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.613940 27751 factory.go:105] Error trying to work out if we can handle /kubepods/burstable/pod97270c63-c9a8-11e7-89f4-c6b053eac242: /kubepods/burstable/pod97270c63-c9a8-11e7-89f4-c6b053eac242 not handled by systemd handler
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.613948 27751 factory.go:116] Factory "systemd" was unable to handle container "/kubepods/burstable/pod97270c63-c9a8-11e7-89f4-c6b053eac242"
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.613956 27751 factory.go:112] Using factory "raw" for container "/kubepods/burstable/pod97270c63-c9a8-11e7-89f4-c6b053eac242"
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.615190 27751 manager.go:932] Added container: "/kubepods/burstable/pod97270c63-c9a8-11e7-89f4-c6b053eac242" (aliases: [], namespace: "")
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.615318 27751 handler.go:325] Added event &{/kubepods/burstable/pod97270c63-c9a8-11e7-89f4-c6b053eac242 2017-11-15 02:01:21.613194711 +0000 UTC containerCreation {<nil>}}
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.615355 27751 container.go:409] Start housekeeping for container "/kubepods/burstable/pod97270c63-c9a8-11e7-89f4-c6b053eac242"
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.618187 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.624415 27751 status_manager.go:451] Status for pod "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242)" updated successfully: (1, {Phase:Pending Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:01:21 +0000 UTC Reason: Message:} {Type:Ready Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:01:21 +0000 UTC Reason:ContainersNotReady Message:containers with unready status: [kubedns dnsmasq sidecar]} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:01:21 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP: StartTime:2017-11-15 02:01:21 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:dnsmasq State:{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,} Running:nil Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:false RestartCount:0 Image:gcr.io/google_containers/k8s-dns-dnsmasq-nanny-amd64:1.14.5 ImageID: ContainerID:} {Name:kubedns State:{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,} Running:nil Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:false RestartCount:0 Image:gcr.io/google_containers/k8s-dns-kube-dns-amd64:1.14.5 ImageID: ContainerID:} {Name:sidecar State:{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,} Running:nil Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:false RestartCount:0 Image:gcr.io/google_containers/k8s-dns-sidecar-amd64:1.14.5 ImageID: ContainerID:}] QOSClass:Burstable})
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.624667 27751 config.go:282] Setting pods for source api
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.627670 27751 kubelet.go:1850] SyncLoop (RECONCILE, "api"): "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.786392 27751 reconciler.go:212] operationExecutor.VerifyControllerAttachedVolume started for volume "kube-dns-config" (UniqueName: "kubernetes.io/configmap/97270c63-c9a8-11e7-89f4-c6b053eac242-kube-dns-config") pod "kube-dns-545bc4bfd4-zvfqd" (UID: "97270c63-c9a8-11e7-89f4-c6b053eac242")
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.786445 27751 reconciler.go:212] operationExecutor.VerifyControllerAttachedVolume started for volume "kube-dns-token-987zv" (UniqueName: "kubernetes.io/secret/97270c63-c9a8-11e7-89f4-c6b053eac242-kube-dns-token-987zv") pod "kube-dns-545bc4bfd4-zvfqd" (UID: "97270c63-c9a8-11e7-89f4-c6b053eac242")
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.886781 27751 reconciler.go:257] operationExecutor.MountVolume started for volume "kube-dns-config" (UniqueName: "kubernetes.io/configmap/97270c63-c9a8-11e7-89f4-c6b053eac242-kube-dns-config") pod "kube-dns-545bc4bfd4-zvfqd" (UID: "97270c63-c9a8-11e7-89f4-c6b053eac242")
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.886852 27751 reconciler.go:257] operationExecutor.MountVolume started for volume "kube-dns-token-987zv" (UniqueName: "kubernetes.io/secret/97270c63-c9a8-11e7-89f4-c6b053eac242-kube-dns-token-987zv") pod "kube-dns-545bc4bfd4-zvfqd" (UID: "97270c63-c9a8-11e7-89f4-c6b053eac242")
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.886906 27751 secret.go:186] Setting up volume kube-dns-token-987zv for pod 97270c63-c9a8-11e7-89f4-c6b053eac242 at /var/lib/kubelet/pods/97270c63-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/kube-dns-token-987zv
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.887109 27751 empty_dir.go:264] pod 97270c63-c9a8-11e7-89f4-c6b053eac242: mounting tmpfs for volume wrapped_kube-dns-token-987zv
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.887129 27751 mount_linux.go:135] Mounting cmd (systemd-run) with arguments ([--description=Kubernetes transient mount for /var/lib/kubelet/pods/97270c63-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/kube-dns-token-987zv --scope -- mount -t tmpfs tmpfs /var/lib/kubelet/pods/97270c63-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/kube-dns-token-987zv])
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.889747 27751 configmap.go:187] Setting up volume kube-dns-config for pod 97270c63-c9a8-11e7-89f4-c6b053eac242 at /var/lib/kubelet/pods/97270c63-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~configmap/kube-dns-config
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.895372 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/run-30165.scope"
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.895389 27751 factory.go:105] Error trying to work out if we can handle /system.slice/run-30165.scope: /system.slice/run-30165.scope not handled by systemd handler
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.895396 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/run-30165.scope"
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.895403 27751 factory.go:112] Using factory "raw" for container "/system.slice/run-30165.scope"
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.895560 27751 manager.go:932] Added container: "/system.slice/run-30165.scope" (aliases: [], namespace: "")
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.895667 27751 handler.go:325] Added event &{/system.slice/run-30165.scope 2017-11-15 02:01:21.893194711 +0000 UTC containerCreation {<nil>}}
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.895700 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-9ba72245089e9b36e901a816a7446b004869c57dbf880db7daeb3edac0f81e51.mount"
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.895726 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-9ba72245089e9b36e901a816a7446b004869c57dbf880db7daeb3edac0f81e51.mount", but ignoring.
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.895737 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-9ba72245089e9b36e901a816a7446b004869c57dbf880db7daeb3edac0f81e51.mount"
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.895759 27751 container.go:409] Start housekeeping for container "/system.slice/run-30165.scope"
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.901861 27751 configmap.go:218] Received configMap kube-system/kube-dns containing (0) pieces of data, 0 total bytes
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.901928 27751 atomic_writer.go:142] pod kube-system/kube-dns-545bc4bfd4-zvfqd volume kube-dns-config: no update required for target directory /var/lib/kubelet/pods/97270c63-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~configmap/kube-dns-config
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.901965 27751 operation_generator.go:484] MountVolume.SetUp succeeded for volume "kube-dns-config" (UniqueName: "kubernetes.io/configmap/97270c63-c9a8-11e7-89f4-c6b053eac242-kube-dns-config") pod "kube-dns-545bc4bfd4-zvfqd" (UID: "97270c63-c9a8-11e7-89f4-c6b053eac242")
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.902243 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-dns-545bc4bfd4-zvfqd", UID:"97270c63-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"477", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "kube-dns-config"
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.902449 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-be14b0552100f929052c8e94e1b7cf587c4218e321d811314e58f2cf312f1c0f-shm.mount"
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.902463 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-be14b0552100f929052c8e94e1b7cf587c4218e321d811314e58f2cf312f1c0f-shm.mount", but ignoring.
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.902480 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-containers-be14b0552100f929052c8e94e1b7cf587c4218e321d811314e58f2cf312f1c0f-shm.mount"
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.902504 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-kubelet-pods-b77b0858\\x2dc9a8\\x2d11e7\\x2d89f4\\x2dc6b053eac242-volumes-kubernetes.io\\x7esecret-weave\\x2dnet\\x2dtoken\\x2drn6j7.mount"
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.902516 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-kubelet-pods-b77b0858\\x2dc9a8\\x2d11e7\\x2d89f4\\x2dc6b053eac242-volumes-kubernetes.io\\x7esecret-weave\\x2dnet\\x2dtoken\\x2drn6j7.mount", but ignoring.
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.902531 27751 manager.go:901] ignoring container "/system.slice/var-lib-kubelet-pods-b77b0858\\x2dc9a8\\x2d11e7\\x2d89f4\\x2dc6b053eac242-volumes-kubernetes.io\\x7esecret-weave\\x2dnet\\x2dtoken\\x2drn6j7.mount"
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.902543 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-ffde707e35c082af5481df3bb46e12f0eaf2bf20d3f28b499b6739ec31319af8.mount"
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.902551 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-ffde707e35c082af5481df3bb46e12f0eaf2bf20d3f28b499b6739ec31319af8.mount", but ignoring.
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.902560 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-ffde707e35c082af5481df3bb46e12f0eaf2bf20d3f28b499b6739ec31319af8.mount"
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.903278 27751 manager.go:989] Destroyed container: "/system.slice/run-30165.scope" (aliases: [], namespace: "")
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.903294 27751 handler.go:325] Added event &{/system.slice/run-30165.scope 2017-11-15 02:01:21.903289511 +0000 UTC containerDeletion {<nil>}}
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.906816 27751 secret.go:217] Received secret kube-system/kube-dns-token-987zv containing (3) pieces of data, 1896 total bytes
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.906873 27751 atomic_writer.go:145] pod kube-system/kube-dns-545bc4bfd4-zvfqd volume kube-dns-token-987zv: write required for target directory /var/lib/kubelet/pods/97270c63-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/kube-dns-token-987zv
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.906956 27751 atomic_writer.go:160] pod kube-system/kube-dns-545bc4bfd4-zvfqd volume kube-dns-token-987zv: performed write of new data to ts data directory: /var/lib/kubelet/pods/97270c63-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/kube-dns-token-987zv/..119811_15_11_02_01_21.803270174
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.907035 27751 operation_generator.go:484] MountVolume.SetUp succeeded for volume "kube-dns-token-987zv" (UniqueName: "kubernetes.io/secret/97270c63-c9a8-11e7-89f4-c6b053eac242-kube-dns-token-987zv") pod "kube-dns-545bc4bfd4-zvfqd" (UID: "97270c63-c9a8-11e7-89f4-c6b053eac242")
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.907059 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-dns-545bc4bfd4-zvfqd", UID:"97270c63-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"477", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "kube-dns-token-987zv"
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.918450 27751 volume_manager.go:366] All volumes are attached and mounted for pod "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.918475 27751 kuberuntime_manager.go:370] No sandbox for pod "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242)" can be found. Need to start a new one
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.918488 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:true CreateSandbox:true SandboxID: Attempt:0 NextInitContainerToStart:nil ContainersToStart:[0 1 2] ContainersToKill:map[]} for pod "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.918527 27751 kuberuntime_manager.go:565] SyncPod received new pod "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242)", will create a sandbox for it
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.918536 27751 kuberuntime_manager.go:574] Stopping PodSandbox for "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242)", will start new one
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.918569 27751 kuberuntime_manager.go:626] Creating sandbox for pod "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.921048 27751 expiration_cache.go:98] Entry version: {key:version obj:0xc420f30500} has expired
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.921810 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/burstable/pod97270c63-c9a8-11e7-89f4-c6b053eac242"
Nov 15 02:01:21 af867b kubelet[27751]: I1115 02:01:21.921828 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/burstable/pod97270c63-c9a8-11e7-89f4-c6b053eac242"
Nov 15 02:01:22 af867b kubelet[27751]: I1115 02:01:22.270358 27751 factory.go:112] Using factory "docker" for container "/kubepods/burstable/pod97270c63-c9a8-11e7-89f4-c6b053eac242/57f84a53f69dda718423e2a72b069d129afa4d226103816e6fa21ee5fab8b82b"
Nov 15 02:01:22 af867b kubelet[27751]: I1115 02:01:22.272386 27751 manager.go:932] Added container: "/kubepods/burstable/pod97270c63-c9a8-11e7-89f4-c6b053eac242/57f84a53f69dda718423e2a72b069d129afa4d226103816e6fa21ee5fab8b82b" (aliases: [k8s_POD_kube-dns-545bc4bfd4-zvfqd_kube-system_97270c63-c9a8-11e7-89f4-c6b053eac242_0 57f84a53f69dda718423e2a72b069d129afa4d226103816e6fa21ee5fab8b82b], namespace: "docker")
Nov 15 02:01:22 af867b kubelet[27751]: I1115 02:01:22.272525 27751 handler.go:325] Added event &{/kubepods/burstable/pod97270c63-c9a8-11e7-89f4-c6b053eac242/57f84a53f69dda718423e2a72b069d129afa4d226103816e6fa21ee5fab8b82b 2017-11-15 02:01:22.197194711 +0000 UTC containerCreation {<nil>}}
Nov 15 02:01:22 af867b kubelet[27751]: I1115 02:01:22.272564 27751 container.go:409] Start housekeeping for container "/kubepods/burstable/pod97270c63-c9a8-11e7-89f4-c6b053eac242/57f84a53f69dda718423e2a72b069d129afa4d226103816e6fa21ee5fab8b82b"
Nov 15 02:01:22 af867b kubelet[27751]: I1115 02:01:22.275968 27751 docker_sandbox.go:691] Will attempt to re-write config file /var/lib/docker/containers/57f84a53f69dda718423e2a72b069d129afa4d226103816e6fa21ee5fab8b82b/resolv.conf with:
Nov 15 02:01:22 af867b kubelet[27751]: [nameserver 10.196.65.209 search opcwlaas.oraclecloud.internal. opcwlaas.oraclecloud.internal.]
Nov 15 02:01:22 af867b kubelet[27751]: I1115 02:01:22.276073 27751 plugins.go:392] Calling network plugin cni to set up pod "kube-dns-545bc4bfd4-zvfqd_kube-system"
Nov 15 02:01:22 af867b kubelet[27751]: I1115 02:01:22.277634 27751 cni.go:326] Got netns path /proc/30231/ns/net
Nov 15 02:01:22 af867b kubelet[27751]: I1115 02:01:22.277645 27751 cni.go:327] Using netns path kube-system
Nov 15 02:01:22 af867b kubelet[27751]: I1115 02:01:22.277803 27751 cni.go:298] About to add CNI network cni-loopback (type=loopback)
Nov 15 02:01:22 af867b kubelet[27751]: I1115 02:01:22.304782 27751 cni.go:326] Got netns path /proc/30231/ns/net
Nov 15 02:01:22 af867b kubelet[27751]: I1115 02:01:22.304796 27751 cni.go:327] Using netns path kube-system
Nov 15 02:01:22 af867b kubelet[27751]: I1115 02:01:22.304935 27751 cni.go:298] About to add CNI network weave (type=weave-net)
Nov 15 02:01:22 af867b kubelet[27751]: I1115 02:01:22.407498 27751 kuberuntime_manager.go:640] Created PodSandbox "57f84a53f69dda718423e2a72b069d129afa4d226103816e6fa21ee5fab8b82b" for pod "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:01:22 af867b kubelet[27751]: I1115 02:01:22.413095 27751 kuberuntime_manager.go:654] Determined the ip "10.32.0.2" for pod "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242)" after sandbox changed
Nov 15 02:01:22 af867b kubelet[27751]: I1115 02:01:22.413230 27751 kuberuntime_manager.go:705] Creating container &Container{Name:kubedns,Image:gcr.io/google_containers/k8s-dns-kube-dns-amd64:1.14.5,Command:[],Args:[--domain=cluster.local. --dns-port=10053 --config-dir=/kube-dns-config --v=2],WorkingDir:,Ports:[{dns-local 0 10053 UDP } {dns-tcp-local 0 10053 TCP } {metrics 0 10055 TCP }],Env:[{PROMETHEUS_PORT 10055 nil}],Resources:ResourceRequirements{Limits:ResourceList{memory: {{178257920 0} {<nil>} 170Mi BinarySI},},Requests:ResourceList{cpu: {{100 -3} {<nil>} 100m DecimalSI},memory: {{73400320 0} {<nil>} 70Mi BinarySI},},},VolumeMounts:[{kube-dns-config false /kube-dns-config <nil>} {kube-dns-token-987zv true /var/run/secrets/kubernetes.io/serviceaccount <nil>}],LivenessProbe:&Probe{Handler:Handler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthcheck/kubedns,Port:10054,Host:,Scheme:HTTP,HTTPHeaders:[],},TCPSocket:nil,},InitialDelaySeconds:60,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:5,},ReadinessProbe:&Probe{Handler:Handler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readiness,Port:8081,Host:,Scheme:HTTP,HTTPHeaders:[],},TCPSocket:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[],TerminationMessagePolicy:File,} in pod kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242)
Nov 15 02:01:22 af867b kubelet[27751]: I1115 02:01:22.415191 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-dns-545bc4bfd4-zvfqd", UID:"97270c63-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"477", FieldPath:"spec.containers{kubedns}"}): type: 'Normal' reason: 'Pulling' pulling image "gcr.io/google_containers/k8s-dns-kube-dns-amd64:1.14.5"
Nov 15 02:01:22 af867b kubelet[27751]: I1115 02:01:22.523488 27751 kubelet.go:1890] SyncLoop (SYNC): 1 pods; kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)
Nov 15 02:01:22 af867b kubelet[27751]: I1115 02:01:22.523560 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:01:22 af867b kubelet[27751]: I1115 02:01:22.524917 27751 kubelet_pods.go:1284] Generating status for "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)"
Nov 15 02:01:22 af867b kubelet[27751]: I1115 02:01:22.525144 27751 status_manager.go:325] Ignoring same status for pod "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)", status: {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:58 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP:10.196.65.210 StartTime:2017-11-15 01:58:55 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:kube-apiserver State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 01:58:58 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:gcr.io/google_containers/kube-apiserver-amd64:v1.8.3 ImageID:docker-pullable://gcr.io/google_containers/kube-apiserver-amd64@sha256:872e3d4286a8ef4338df59945cb0d64c2622268ceb3e8a2ce7b52243279b02d0 ContainerID:docker://8da2c70a27f08a2f062af80b5708e01ac34ce76b42ab4a6eaa0288e2daf8a043}] QOSClass:Burstable}
Nov 15 02:01:22 af867b kubelet[27751]: I1115 02:01:22.525327 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)"
Nov 15 02:01:22 af867b kubelet[27751]: I1115 02:01:22.709402 27751 kube_docker_client.go:330] Pulling image "weaveworks/weave-npc:2.0.5": "a2592a033c5d: Downloading [====================================> ] 8.273MB/11.27MB"
Nov 15 02:01:22 af867b kubelet[27751]: I1115 02:01:22.825573 27751 volume_manager.go:366] All volumes are attached and mounted for pod "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)"
Nov 15 02:01:22 af867b kubelet[27751]: I1115 02:01:22.825763 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:false CreateSandbox:false SandboxID:19b06ca14052c17f92bf03d5316fd7642fe4fee6bf193755575355be3e87a13d Attempt:0 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d)"
Nov 15 02:01:22 af867b kubelet[27751]: I1115 02:01:22.959960 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 2379, Path: /health
Nov 15 02:01:22 af867b kubelet[27751]: I1115 02:01:22.960001 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:01:23 af867b kubelet[27751]: I1115 02:01:23.047567 27751 generic.go:146] GenericPLEG: 97270c63-c9a8-11e7-89f4-c6b053eac242/57f84a53f69dda718423e2a72b069d129afa4d226103816e6fa21ee5fab8b82b: non-existent -> running
Nov 15 02:01:23 af867b kubelet[27751]: I1115 02:01:23.049122 27751 kuberuntime_manager.go:833] getSandboxIDByPodUID got sandbox IDs ["57f84a53f69dda718423e2a72b069d129afa4d226103816e6fa21ee5fab8b82b"] for pod "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:01:23 af867b kubelet[27751]: I1115 02:01:23.055005 27751 generic.go:345] PLEG: Write status for kube-dns-545bc4bfd4-zvfqd/kube-system: &container.PodStatus{ID:"97270c63-c9a8-11e7-89f4-c6b053eac242", Name:"kube-dns-545bc4bfd4-zvfqd", Namespace:"kube-system", IP:"10.32.0.2", ContainerStatuses:[]*container.ContainerStatus{}, SandboxStatuses:[]*runtime.PodSandboxStatus{(*runtime.PodSandboxStatus)(0xc42139c1e0)}} (err: <nil>)
Nov 15 02:01:23 af867b kubelet[27751]: I1115 02:01:23.055065 27751 kubelet.go:1871] SyncLoop (PLEG): "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242)", event: &pleg.PodLifecycleEvent{ID:"97270c63-c9a8-11e7-89f4-c6b053eac242", Type:"ContainerStarted", Data:"57f84a53f69dda718423e2a72b069d129afa4d226103816e6fa21ee5fab8b82b"}
Nov 15 02:01:23 af867b kubelet[27751]: I1115 02:01:23.197471 27751 worker.go:164] Probe target container not found: kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242) - kubedns
Nov 15 02:01:23 af867b kubelet[27751]: I1115 02:01:23.211212 27751 http.go:96] Probe succeeded for http://127.0.0.1:2379/health, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Type:[text/plain; charset=utf-8] Date:[Wed, 15 Nov 2017 02:01:23 GMT] Content-Length:[18]] 0xc420e097a0 18 [] true false map[] 0xc420a31b00 <nil>}
Nov 15 02:01:23 af867b kubelet[27751]: I1115 02:01:23.211263 27751 prober.go:113] Liveness probe for "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250):etcd" succeeded
Nov 15 02:01:23 af867b kubelet[27751]: I1115 02:01:23.523446 27751 kubelet.go:1890] SyncLoop (SYNC): 1 pods; kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)
Nov 15 02:01:23 af867b kubelet[27751]: I1115 02:01:23.523541 27751 kubelet_pods.go:1284] Generating status for "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)"
Nov 15 02:01:23 af867b kubelet[27751]: I1115 02:01:23.523762 27751 status_manager.go:325] Ignoring same status for pod "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)", status: {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:58 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP:10.196.65.210 StartTime:2017-11-15 01:58:55 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:kube-scheduler State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 01:58:57 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:gcr.io/google_containers/kube-scheduler-amd64:v1.8.3 ImageID:docker-pullable://gcr.io/google_containers/kube-scheduler-amd64@sha256:c47b2438bbab28d58e8cbf64b37b7f66d26b000f5c3a31626ee829a4be8fb91e ContainerID:docker://413566b22305750f9a9aa46fbe256c11e75293e80f6e0d4afb7ec9e6afcdee05}] QOSClass:Burstable}
Nov 15 02:01:23 af867b kubelet[27751]: I1115 02:01:23.523936 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)"
Nov 15 02:01:23 af867b kubelet[27751]: I1115 02:01:23.617346 27751 prober.go:160] HTTP-Probe Host: https://127.0.0.1, Port: 6443, Path: /healthz
Nov 15 02:01:23 af867b kubelet[27751]: I1115 02:01:23.617390 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:01:23 af867b kubelet[27751]: I1115 02:01:23.623968 27751 http.go:96] Probe succeeded for https://127.0.0.1:6443/healthz, Response: {200 OK 200 HTTP/2.0 2 0 map[Content-Type:[text/plain; charset=utf-8] Content-Length:[2] Date:[Wed, 15 Nov 2017 02:01:23 GMT]] 0xc420e09b00 2 [] false false map[] 0xc42110ba00 0xc42104ef20}
Nov 15 02:01:23 af867b kubelet[27751]: I1115 02:01:23.624019 27751 prober.go:113] Liveness probe for "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d):kube-apiserver" succeeded
Nov 15 02:01:23 af867b kubelet[27751]: I1115 02:01:23.824224 27751 volume_manager.go:366] All volumes are attached and mounted for pod "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)"
Nov 15 02:01:23 af867b kubelet[27751]: I1115 02:01:23.824389 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:false CreateSandbox:false SandboxID:fe18d337ce42bd3a4d2aacb1349be028274681d8073fbe5d6377946e815fa810 Attempt:0 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)"
Nov 15 02:01:23 af867b kubelet[27751]: I1115 02:01:23.834330 27751 worker.go:164] Probe target container not found: kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242) - dnsmasq
Nov 15 02:01:24 af867b kubelet[27751]: I1115 02:01:24.504443 27751 worker.go:164] Probe target container not found: kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242) - sidecar
Nov 15 02:01:24 af867b kubelet[27751]: I1115 02:01:24.523439 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:01:25 af867b kubelet[27751]: I1115 02:01:25.362915 27751 worker.go:164] Probe target container not found: weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242) - weave
Nov 15 02:01:25 af867b kubelet[27751]: I1115 02:01:25.867814 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=true reason: message:
Nov 15 02:01:26 af867b kubelet[27751]: I1115 02:01:26.523447 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:01:27 af867b kubelet[27751]: I1115 02:01:27.506182 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10252, Path: /healthz
Nov 15 02:01:27 af867b kubelet[27751]: I1115 02:01:27.506208 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:01:27 af867b kubelet[27751]: I1115 02:01:27.507658 27751 http.go:96] Probe succeeded for http://127.0.0.1:10252/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:01:27 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc421347440 2 [] true false map[] 0xc420a31700 <nil>}
Nov 15 02:01:27 af867b kubelet[27751]: I1115 02:01:27.507706 27751 prober.go:113] Liveness probe for "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9):kube-controller-manager" succeeded
Nov 15 02:01:27 af867b kubelet[27751]: I1115 02:01:27.599386 27751 worker.go:164] Probe target container not found: kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242) - kubedns
Nov 15 02:01:27 af867b kubelet[27751]: I1115 02:01:27.681661 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10251, Path: /healthz
Nov 15 02:01:27 af867b kubelet[27751]: I1115 02:01:27.681697 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:01:27 af867b kubelet[27751]: I1115 02:01:27.682889 27751 http.go:96] Probe succeeded for http://127.0.0.1:10251/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:01:27 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc42099e380 2 [] true false map[] 0xc420d4be00 <nil>}
Nov 15 02:01:27 af867b kubelet[27751]: I1115 02:01:27.682936 27751 prober.go:113] Liveness probe for "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373):kube-scheduler" succeeded
Nov 15 02:01:28 af867b kubelet[27751]: I1115 02:01:28.523458 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:01:30 af867b kubelet[27751]: I1115 02:01:30.523951 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:01:30 af867b kubelet[27751]: I1115 02:01:30.869258 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=true reason: message:
Nov 15 02:01:31 af867b kubelet[27751]: I1115 02:01:31.554950 27751 eviction_manager.go:221] eviction manager: synchronize housekeeping
Nov 15 02:01:31 af867b kubelet[27751]: I1115 02:01:31.624888 27751 helper.go:148] Missing default interface "eth0" for pod:kube-system_kube-dns-545bc4bfd4-zvfqd
Nov 15 02:01:31 af867b kubelet[27751]: I1115 02:01:31.625012 27751 helpers.go:873] eviction manager: observations: signal=allocatableMemory.available, available: 6970772Ki, capacity: 7393360Ki
Nov 15 02:01:31 af867b kubelet[27751]: I1115 02:01:31.625032 27751 helpers.go:873] eviction manager: observations: signal=allocatableNodeFs.available, available: 9624040228, capacity: 10198Mi
Nov 15 02:01:31 af867b kubelet[27751]: I1115 02:01:31.625040 27751 helpers.go:871] eviction manager: observations: signal=memory.available, available: 6532960Ki, capacity: 7393360Ki, time: 2017-11-15 02:01:19.237424468 +0000 UTC
Nov 15 02:01:31 af867b kubelet[27751]: I1115 02:01:31.625059 27751 helpers.go:871] eviction manager: observations: signal=nodefs.available, available: 7401080Ki, capacity: 10198Mi, time: 2017-11-15 02:01:19.237424468 +0000 UTC
Nov 15 02:01:31 af867b kubelet[27751]: I1115 02:01:31.625070 27751 helpers.go:871] eviction manager: observations: signal=nodefs.inodesFree, available: 10384695, capacity: 10208Ki, time: 2017-11-15 02:01:19.237424468 +0000 UTC
Nov 15 02:01:31 af867b kubelet[27751]: I1115 02:01:31.625079 27751 helpers.go:871] eviction manager: observations: signal=imagefs.available, available: 40868352Ki, capacity: 45Gi, time: 2017-11-15 02:01:19.237424468 +0000 UTC
Nov 15 02:01:31 af867b kubelet[27751]: I1115 02:01:31.625100 27751 eviction_manager.go:325] eviction manager: no resources are starved
Nov 15 02:01:32 af867b kubelet[27751]: I1115 02:01:32.523482 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:01:32 af867b kubelet[27751]: I1115 02:01:32.556200 27751 kube_docker_client.go:333] Stop pulling image "weaveworks/weave-npc:2.0.5": "Status: Downloaded newer image for weaveworks/weave-npc:2.0.5"
Nov 15 02:01:32 af867b kubelet[27751]: I1115 02:01:32.558552 27751 kuberuntime_image.go:46] Pulling image "gcr.io/google_containers/k8s-dns-kube-dns-amd64:1.14.5" without credentials
Nov 15 02:01:32 af867b kubelet[27751]: I1115 02:01:32.558672 27751 kuberuntime_container.go:100] Generating ref for container weave-npc: &v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"weave-net-rg7fn", UID:"b77b0858-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"381", FieldPath:"spec.containers{weave-npc}"}
Nov 15 02:01:32 af867b kubelet[27751]: I1115 02:01:32.558739 27751 container_manager_linux.go:634] Calling devicePluginHandler AllocateDevices
Nov 15 02:01:32 af867b kubelet[27751]: I1115 02:01:32.558805 27751 kubelet_pods.go:123] container: kube-system/weave-net-rg7fn/weave-npc podIP: "10.196.65.210" creating hosts mount: true
Nov 15 02:01:32 af867b kubelet[27751]: I1115 02:01:32.559506 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"weave-net-rg7fn", UID:"b77b0858-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"381", FieldPath:"spec.containers{weave-npc}"}): type: 'Normal' reason: 'Pulled' Successfully pulled image "weaveworks/weave-npc:2.0.5"
Nov 15 02:01:32 af867b kubelet[27751]: I1115 02:01:32.566219 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/burstable/podb77b0858-c9a8-11e7-89f4-c6b053eac242"
Nov 15 02:01:32 af867b kubelet[27751]: I1115 02:01:32.821540 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"weave-net-rg7fn", UID:"b77b0858-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"381", FieldPath:"spec.containers{weave-npc}"}): type: 'Normal' reason: 'Created' Created container
Nov 15 02:01:32 af867b kubelet[27751]: I1115 02:01:32.960017 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 2379, Path: /health
Nov 15 02:01:32 af867b kubelet[27751]: I1115 02:01:32.960058 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:01:32 af867b kubelet[27751]: I1115 02:01:32.981077 27751 factory.go:112] Using factory "docker" for container "/kubepods/burstable/podb77b0858-c9a8-11e7-89f4-c6b053eac242/43fe4d3eda8b4078f76541a5e9dca827819ff4e80e9a043f89f50752401bf06c"
Nov 15 02:01:32 af867b kubelet[27751]: I1115 02:01:32.984077 27751 manager.go:932] Added container: "/kubepods/burstable/podb77b0858-c9a8-11e7-89f4-c6b053eac242/43fe4d3eda8b4078f76541a5e9dca827819ff4e80e9a043f89f50752401bf06c" (aliases: [k8s_weave-npc_weave-net-rg7fn_kube-system_b77b0858-c9a8-11e7-89f4-c6b053eac242_0 43fe4d3eda8b4078f76541a5e9dca827819ff4e80e9a043f89f50752401bf06c], namespace: "docker")
Nov 15 02:01:32 af867b kubelet[27751]: I1115 02:01:32.984254 27751 handler.go:325] Added event &{/kubepods/burstable/podb77b0858-c9a8-11e7-89f4-c6b053eac242/43fe4d3eda8b4078f76541a5e9dca827819ff4e80e9a043f89f50752401bf06c 2017-11-15 02:01:32.896194711 +0000 UTC containerCreation {<nil>}}
Nov 15 02:01:32 af867b kubelet[27751]: I1115 02:01:32.984304 27751 container.go:409] Start housekeeping for container "/kubepods/burstable/podb77b0858-c9a8-11e7-89f4-c6b053eac242/43fe4d3eda8b4078f76541a5e9dca827819ff4e80e9a043f89f50752401bf06c"
Nov 15 02:01:32 af867b kubelet[27751]: I1115 02:01:32.988919 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"weave-net-rg7fn", UID:"b77b0858-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"381", FieldPath:"spec.containers{weave-npc}"}): type: 'Normal' reason: 'Started' Started container
Nov 15 02:01:33 af867b kubelet[27751]: I1115 02:01:33.100121 27751 generic.go:146] GenericPLEG: b77b0858-c9a8-11e7-89f4-c6b053eac242/43fe4d3eda8b4078f76541a5e9dca827819ff4e80e9a043f89f50752401bf06c: non-existent -> running
Nov 15 02:01:33 af867b kubelet[27751]: I1115 02:01:33.106441 27751 kuberuntime_manager.go:833] getSandboxIDByPodUID got sandbox IDs ["be14b0552100f929052c8e94e1b7cf587c4218e321d811314e58f2cf312f1c0f"] for pod "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:01:33 af867b kubelet[27751]: I1115 02:01:33.113056 27751 generic.go:345] PLEG: Write status for weave-net-rg7fn/kube-system: &container.PodStatus{ID:"b77b0858-c9a8-11e7-89f4-c6b053eac242", Name:"weave-net-rg7fn", Namespace:"kube-system", IP:"", ContainerStatuses:[]*container.ContainerStatus{(*container.ContainerStatus)(0xc421cc70a0), (*container.ContainerStatus)(0xc421cc7180)}, SandboxStatuses:[]*runtime.PodSandboxStatus{(*runtime.PodSandboxStatus)(0xc4219240f0)}} (err: <nil>)
Nov 15 02:01:33 af867b kubelet[27751]: I1115 02:01:33.113129 27751 kubelet.go:1871] SyncLoop (PLEG): "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242)", event: &pleg.PodLifecycleEvent{ID:"b77b0858-c9a8-11e7-89f4-c6b053eac242", Type:"ContainerStarted", Data:"43fe4d3eda8b4078f76541a5e9dca827819ff4e80e9a043f89f50752401bf06c"}
Nov 15 02:01:33 af867b kubelet[27751]: I1115 02:01:33.113176 27751 kubelet_pods.go:1284] Generating status for "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:01:33 af867b kubelet[27751]: I1115 02:01:33.113434 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:01:33 af867b kubelet[27751]: I1115 02:01:33.134602 27751 secret.go:186] Setting up volume weave-net-token-rn6j7 for pod b77b0858-c9a8-11e7-89f4-c6b053eac242 at /var/lib/kubelet/pods/b77b0858-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/weave-net-token-rn6j7
Nov 15 02:01:33 af867b kubelet[27751]: I1115 02:01:33.135151 27751 status_manager.go:451] Status for pod "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242)" updated successfully: (2, {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:00:12 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:01:33 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP:10.196.65.210 StartTime:2017-11-15 02:00:12 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:weave State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 02:01:01 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:weaveworks/weave-kube:2.0.5 ImageID:docker-pullable://weaveworks/weave-kube@sha256:1af289ad3cf6ddaa7bb6cc31ad32f64adf2728635c971e4c54399a291c7aeb96 ContainerID:docker://f4fa9d4e3be52b17cedfe3b19a153e4584737c95ada75d64ea13225dde740f40} {Name:weave-npc State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 02:01:32 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:weaveworks/weave-npc:2.0.5 ImageID:docker-pullable://weaveworks/weave-npc@sha256:da936be1a2bd3f1c05cc80ab21e3282d15dd7d95223479fd563b6d1ae8c54ef3 ContainerID:docker://43fe4d3eda8b4078f76541a5e9dca827819ff4e80e9a043f89f50752401bf06c}] QOSClass:Burstable})
Nov 15 02:01:33 af867b kubelet[27751]: I1115 02:01:33.135385 27751 config.go:282] Setting pods for source api
Nov 15 02:01:33 af867b kubelet[27751]: I1115 02:01:33.145071 27751 kubelet.go:1850] SyncLoop (RECONCILE, "api"): "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:01:33 af867b kubelet[27751]: I1115 02:01:33.146091 27751 secret.go:217] Received secret kube-system/weave-net-token-rn6j7 containing (3) pieces of data, 1900 total bytes
Nov 15 02:01:33 af867b kubelet[27751]: I1115 02:01:33.147900 27751 atomic_writer.go:142] pod kube-system/weave-net-rg7fn volume weave-net-token-rn6j7: no update required for target directory /var/lib/kubelet/pods/b77b0858-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/weave-net-token-rn6j7
Nov 15 02:01:33 af867b kubelet[27751]: I1115 02:01:33.197614 27751 worker.go:164] Probe target container not found: kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242) - kubedns
Nov 15 02:01:33 af867b kubelet[27751]: I1115 02:01:33.218183 27751 http.go:96] Probe succeeded for http://127.0.0.1:2379/health, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Length:[18] Content-Type:[text/plain; charset=utf-8] Date:[Wed, 15 Nov 2017 02:01:33 GMT]] 0xc421dee0a0 18 [] true false map[] 0xc4211a8200 <nil>}
Nov 15 02:01:33 af867b kubelet[27751]: I1115 02:01:33.218248 27751 prober.go:113] Liveness probe for "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250):etcd" succeeded
Nov 15 02:01:33 af867b kubelet[27751]: I1115 02:01:33.413807 27751 volume_manager.go:366] All volumes are attached and mounted for pod "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:01:33 af867b kubelet[27751]: I1115 02:01:33.414060 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:false CreateSandbox:false SandboxID:be14b0552100f929052c8e94e1b7cf587c4218e321d811314e58f2cf312f1c0f Attempt:0 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:01:33 af867b kubelet[27751]: I1115 02:01:33.617369 27751 prober.go:160] HTTP-Probe Host: https://127.0.0.1, Port: 6443, Path: /healthz
Nov 15 02:01:33 af867b kubelet[27751]: I1115 02:01:33.617424 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:01:33 af867b kubelet[27751]: I1115 02:01:33.624691 27751 http.go:96] Probe succeeded for https://127.0.0.1:6443/healthz, Response: {200 OK 200 HTTP/2.0 2 0 map[Content-Type:[text/plain; charset=utf-8] Content-Length:[2] Date:[Wed, 15 Nov 2017 02:01:33 GMT]] 0xc421def160 2 [] false false map[] 0xc42110a100 0xc42196a580}
Nov 15 02:01:33 af867b kubelet[27751]: I1115 02:01:33.624767 27751 prober.go:113] Liveness probe for "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d):kube-apiserver" succeeded
Nov 15 02:01:33 af867b kubelet[27751]: I1115 02:01:33.834499 27751 worker.go:164] Probe target container not found: kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242) - dnsmasq
Nov 15 02:01:34 af867b kubelet[27751]: I1115 02:01:34.117271 27751 kubelet_pods.go:1284] Generating status for "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:01:34 af867b kubelet[27751]: I1115 02:01:34.117555 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:01:34 af867b kubelet[27751]: I1115 02:01:34.131983 27751 status_manager.go:451] Status for pod "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242)" updated successfully: (3, {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:00:12 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:01:33 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:01:33 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP:10.196.65.210 StartTime:2017-11-15 02:00:12 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:weave State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 02:01:01 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:weaveworks/weave-kube:2.0.5 ImageID:docker-pullable://weaveworks/weave-kube@sha256:1af289ad3cf6ddaa7bb6cc31ad32f64adf2728635c971e4c54399a291c7aeb96 ContainerID:docker://f4fa9d4e3be52b17cedfe3b19a153e4584737c95ada75d64ea13225dde740f40} {Name:weave-npc State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 02:01:32 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:weaveworks/weave-npc:2.0.5 ImageID:docker-pullable://weaveworks/weave-npc@sha256:da936be1a2bd3f1c05cc80ab21e3282d15dd7d95223479fd563b6d1ae8c54ef3 ContainerID:docker://43fe4d3eda8b4078f76541a5e9dca827819ff4e80e9a043f89f50752401bf06c}] QOSClass:Burstable})
Nov 15 02:01:34 af867b kubelet[27751]: I1115 02:01:34.132296 27751 config.go:282] Setting pods for source api
Nov 15 02:01:34 af867b kubelet[27751]: I1115 02:01:34.135423 27751 kubelet.go:1850] SyncLoop (RECONCILE, "api"): "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:01:34 af867b kubelet[27751]: I1115 02:01:34.141676 27751 secret.go:186] Setting up volume weave-net-token-rn6j7 for pod b77b0858-c9a8-11e7-89f4-c6b053eac242 at /var/lib/kubelet/pods/b77b0858-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/weave-net-token-rn6j7
Nov 15 02:01:34 af867b kubelet[27751]: I1115 02:01:34.155481 27751 secret.go:217] Received secret kube-system/weave-net-token-rn6j7 containing (3) pieces of data, 1900 total bytes
Nov 15 02:01:34 af867b kubelet[27751]: I1115 02:01:34.155646 27751 atomic_writer.go:142] pod kube-system/weave-net-rg7fn volume weave-net-token-rn6j7: no update required for target directory /var/lib/kubelet/pods/b77b0858-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/weave-net-token-rn6j7
Nov 15 02:01:34 af867b kubelet[27751]: I1115 02:01:34.417869 27751 volume_manager.go:366] All volumes are attached and mounted for pod "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:01:34 af867b kubelet[27751]: I1115 02:01:34.418105 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:false CreateSandbox:false SandboxID:be14b0552100f929052c8e94e1b7cf587c4218e321d811314e58f2cf312f1c0f Attempt:0 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:01:34 af867b kubelet[27751]: I1115 02:01:34.504658 27751 worker.go:164] Probe target container not found: kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242) - sidecar
Nov 15 02:01:34 af867b kubelet[27751]: I1115 02:01:34.523520 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:01:35 af867b kubelet[27751]: I1115 02:01:35.362724 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 6784, Path: /status
Nov 15 02:01:35 af867b kubelet[27751]: I1115 02:01:35.362793 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:01:35 af867b kubelet[27751]: I1115 02:01:35.366525 27751 http.go:96] Probe succeeded for http://127.0.0.1:6784/status, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:01:35 GMT] Content-Length:[445] Content-Type:[text/plain; charset=utf-8]] 0xc420d043e0 445 [] true false map[] 0xc420afd500 <nil>}
Nov 15 02:01:35 af867b kubelet[27751]: I1115 02:01:35.366585 27751 prober.go:113] Liveness probe for "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242):weave" succeeded
Nov 15 02:01:35 af867b kubelet[27751]: I1115 02:01:35.870791 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=true reason: message:
Nov 15 02:01:36 af867b kubelet[27751]: I1115 02:01:36.523464 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:01:37 af867b kubelet[27751]: I1115 02:01:37.506131 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10252, Path: /healthz
Nov 15 02:01:37 af867b kubelet[27751]: I1115 02:01:37.506171 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:01:37 af867b kubelet[27751]: I1115 02:01:37.508005 27751 http.go:96] Probe succeeded for http://127.0.0.1:10252/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:01:37 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc4211457a0 2 [] true false map[] 0xc420ee2c00 <nil>}
Nov 15 02:01:37 af867b kubelet[27751]: I1115 02:01:37.508058 27751 prober.go:113] Liveness probe for "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9):kube-controller-manager" succeeded
Nov 15 02:01:37 af867b kubelet[27751]: I1115 02:01:37.523453 27751 kubelet.go:1890] SyncLoop (SYNC): 1 pods; etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)
Nov 15 02:01:37 af867b kubelet[27751]: I1115 02:01:37.523521 27751 kubelet_pods.go:1284] Generating status for "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)"
Nov 15 02:01:37 af867b kubelet[27751]: I1115 02:01:37.523698 27751 status_manager.go:325] Ignoring same status for pod "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)", status: {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:59 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP:10.196.65.210 StartTime:2017-11-15 01:58:55 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:etcd State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 01:58:58 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:gcr.io/google_containers/etcd-amd64:3.0.17 ImageID:docker-pullable://gcr.io/google_containers/etcd-amd64@sha256:d83d3545e06fb035db8512e33bd44afb55dea007a3abd7b17742d3ac6d235940 ContainerID:docker://ab147bb1b65b2001333417bb7654896e6aadb25ce71a8c48c94ae802a2e0197f}] QOSClass:BestEffort}
Nov 15 02:01:37 af867b kubelet[27751]: I1115 02:01:37.523909 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)"
Nov 15 02:01:37 af867b kubelet[27751]: I1115 02:01:37.599553 27751 worker.go:164] Probe target container not found: kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242) - kubedns
Nov 15 02:01:37 af867b kubelet[27751]: I1115 02:01:37.681769 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10251, Path: /healthz
Nov 15 02:01:37 af867b kubelet[27751]: I1115 02:01:37.681802 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:01:37 af867b kubelet[27751]: I1115 02:01:37.686121 27751 http.go:96] Probe succeeded for http://127.0.0.1:10251/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:01:37 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc420bb1500 2 [] true false map[] 0xc420ee3700 <nil>}
Nov 15 02:01:37 af867b kubelet[27751]: I1115 02:01:37.686171 27751 prober.go:113] Liveness probe for "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373):kube-scheduler" succeeded
Nov 15 02:01:37 af867b kubelet[27751]: I1115 02:01:37.824154 27751 volume_manager.go:366] All volumes are attached and mounted for pod "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)"
Nov 15 02:01:37 af867b kubelet[27751]: I1115 02:01:37.824281 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:false CreateSandbox:false SandboxID:d5076ad0e9fb270d1b8c4ff7cdadbf32d6b1e30dc42ae24dfbc5cb01bb5aa934 Attempt:0 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)"
Nov 15 02:01:38 af867b kubelet[27751]: I1115 02:01:38.523526 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:01:40 af867b kubelet[27751]: I1115 02:01:40.523904 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:01:40 af867b kubelet[27751]: I1115 02:01:40.872299 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=true reason: message:
Nov 15 02:01:41 af867b kubelet[27751]: I1115 02:01:41.625267 27751 eviction_manager.go:221] eviction manager: synchronize housekeeping
Nov 15 02:01:41 af867b kubelet[27751]: I1115 02:01:41.683527 27751 helpers.go:873] eviction manager: observations: signal=allocatableNodeFs.available, available: 9624040228, capacity: 10198Mi
Nov 15 02:01:41 af867b kubelet[27751]: I1115 02:01:41.683572 27751 helpers.go:871] eviction manager: observations: signal=memory.available, available: 6485644Ki, capacity: 7393360Ki, time: 2017-11-15 02:01:34.10605089 +0000 UTC
Nov 15 02:01:41 af867b kubelet[27751]: I1115 02:01:41.683593 27751 helpers.go:871] eviction manager: observations: signal=nodefs.available, available: 7408308Ki, capacity: 10198Mi, time: 2017-11-15 02:01:34.10605089 +0000 UTC
Nov 15 02:01:41 af867b kubelet[27751]: I1115 02:01:41.683605 27751 helpers.go:871] eviction manager: observations: signal=nodefs.inodesFree, available: 10384627, capacity: 10208Ki, time: 2017-11-15 02:01:34.10605089 +0000 UTC
Nov 15 02:01:41 af867b kubelet[27751]: I1115 02:01:41.683617 27751 helpers.go:871] eviction manager: observations: signal=imagefs.available, available: 39839Mi, capacity: 45Gi, time: 2017-11-15 02:01:34.10605089 +0000 UTC
Nov 15 02:01:41 af867b kubelet[27751]: I1115 02:01:41.683629 27751 helpers.go:873] eviction manager: observations: signal=allocatableMemory.available, available: 6967980Ki, capacity: 7393360Ki
Nov 15 02:01:41 af867b kubelet[27751]: I1115 02:01:41.683651 27751 eviction_manager.go:325] eviction manager: no resources are starved
Nov 15 02:01:42 af867b kubelet[27751]: I1115 02:01:42.523476 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:01:42 af867b kubelet[27751]: I1115 02:01:42.960000 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 2379, Path: /health
Nov 15 02:01:42 af867b kubelet[27751]: I1115 02:01:42.960037 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:01:43 af867b kubelet[27751]: I1115 02:01:43.197626 27751 worker.go:164] Probe target container not found: kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242) - kubedns
Nov 15 02:01:43 af867b kubelet[27751]: I1115 02:01:43.212432 27751 http.go:96] Probe succeeded for http://127.0.0.1:2379/health, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:01:43 GMT] Content-Length:[18] Content-Type:[text/plain; charset=utf-8]] 0xc420eb7980 18 [] true false map[] 0xc421178100 <nil>}
Nov 15 02:01:43 af867b kubelet[27751]: I1115 02:01:43.212484 27751 prober.go:113] Liveness probe for "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250):etcd" succeeded
Nov 15 02:01:43 af867b kubelet[27751]: I1115 02:01:43.617311 27751 prober.go:160] HTTP-Probe Host: https://127.0.0.1, Port: 6443, Path: /healthz
Nov 15 02:01:43 af867b kubelet[27751]: I1115 02:01:43.617347 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:01:43 af867b kubelet[27751]: I1115 02:01:43.624135 27751 http.go:96] Probe succeeded for https://127.0.0.1:6443/healthz, Response: {200 OK 200 HTTP/2.0 2 0 map[Content-Type:[text/plain; charset=utf-8] Content-Length:[2] Date:[Wed, 15 Nov 2017 02:01:43 GMT]] 0xc4216c55a0 2 [] false false map[] 0xc420c91000 0xc4210d4630}
Nov 15 02:01:43 af867b kubelet[27751]: I1115 02:01:43.624181 27751 prober.go:113] Liveness probe for "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d):kube-apiserver" succeeded
Nov 15 02:01:43 af867b kubelet[27751]: I1115 02:01:43.834482 27751 worker.go:164] Probe target container not found: kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242) - dnsmasq
Nov 15 02:01:44 af867b kubelet[27751]: I1115 02:01:44.159387 27751 kube_docker_client.go:330] Pulling image "gcr.io/google_containers/k8s-dns-kube-dns-amd64:1.14.5": "a46f95c56b32: Downloading [==================================> ] 7.623MB/11.16MB"
Nov 15 02:01:44 af867b kubelet[27751]: I1115 02:01:44.504602 27751 worker.go:164] Probe target container not found: kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242) - sidecar
Nov 15 02:01:44 af867b kubelet[27751]: I1115 02:01:44.523442 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:01:45 af867b kubelet[27751]: I1115 02:01:45.362708 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 6784, Path: /status
Nov 15 02:01:45 af867b kubelet[27751]: I1115 02:01:45.362744 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:01:45 af867b kubelet[27751]: I1115 02:01:45.364147 27751 http.go:96] Probe succeeded for http://127.0.0.1:6784/status, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:01:45 GMT] Content-Length:[445] Content-Type:[text/plain; charset=utf-8]] 0xc420c99680 445 [] true false map[] 0xc421179300 <nil>}
Nov 15 02:01:45 af867b kubelet[27751]: I1115 02:01:45.364204 27751 prober.go:113] Liveness probe for "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242):weave" succeeded
Nov 15 02:01:45 af867b kubelet[27751]: I1115 02:01:45.873728 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=true reason: message:
Nov 15 02:01:46 af867b kubelet[27751]: I1115 02:01:46.523448 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:01:47 af867b kubelet[27751]: I1115 02:01:47.506092 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10252, Path: /healthz
Nov 15 02:01:47 af867b kubelet[27751]: I1115 02:01:47.506128 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:01:47 af867b kubelet[27751]: I1115 02:01:47.507756 27751 http.go:96] Probe succeeded for http://127.0.0.1:10252/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:01:47 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc42147df20 2 [] true false map[] 0xc421179d00 <nil>}
Nov 15 02:01:47 af867b kubelet[27751]: I1115 02:01:47.507796 27751 prober.go:113] Liveness probe for "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9):kube-controller-manager" succeeded
Nov 15 02:01:47 af867b kubelet[27751]: I1115 02:01:47.600800 27751 worker.go:164] Probe target container not found: kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242) - kubedns
Nov 15 02:01:47 af867b kubelet[27751]: I1115 02:01:47.681591 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10251, Path: /healthz
Nov 15 02:01:47 af867b kubelet[27751]: I1115 02:01:47.681617 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:01:47 af867b kubelet[27751]: I1115 02:01:47.683280 27751 http.go:96] Probe succeeded for http://127.0.0.1:10251/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:01:47 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc420f4be20 2 [] true false map[] 0xc420afcd00 <nil>}
Nov 15 02:01:47 af867b kubelet[27751]: I1115 02:01:47.683333 27751 prober.go:113] Liveness probe for "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373):kube-scheduler" succeeded
Nov 15 02:01:48 af867b kubelet[27751]: I1115 02:01:48.488111 27751 kube_docker_client.go:333] Stop pulling image "gcr.io/google_containers/k8s-dns-kube-dns-amd64:1.14.5": "Status: Downloaded newer image for gcr.io/google_containers/k8s-dns-kube-dns-amd64:1.14.5"
Nov 15 02:01:48 af867b kubelet[27751]: I1115 02:01:48.489776 27751 kuberuntime_container.go:100] Generating ref for container kubedns: &v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-dns-545bc4bfd4-zvfqd", UID:"97270c63-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"477", FieldPath:"spec.containers{kubedns}"}
Nov 15 02:01:48 af867b kubelet[27751]: I1115 02:01:48.489814 27751 container_manager_linux.go:634] Calling devicePluginHandler AllocateDevices
Nov 15 02:01:48 af867b kubelet[27751]: I1115 02:01:48.489873 27751 kubelet_pods.go:123] container: kube-system/kube-dns-545bc4bfd4-zvfqd/kubedns podIP: "10.32.0.2" creating hosts mount: true
Nov 15 02:01:48 af867b kubelet[27751]: I1115 02:01:48.490449 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-dns-545bc4bfd4-zvfqd", UID:"97270c63-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"477", FieldPath:"spec.containers{kubedns}"}): type: 'Normal' reason: 'Pulled' Successfully pulled image "gcr.io/google_containers/k8s-dns-kube-dns-amd64:1.14.5"
Nov 15 02:01:48 af867b kubelet[27751]: I1115 02:01:48.493581 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/burstable/pod97270c63-c9a8-11e7-89f4-c6b053eac242"
Nov 15 02:01:48 af867b kubelet[27751]: I1115 02:01:48.524520 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:01:48 af867b kubelet[27751]: I1115 02:01:48.661615 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-dns-545bc4bfd4-zvfqd", UID:"97270c63-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"477", FieldPath:"spec.containers{kubedns}"}): type: 'Normal' reason: 'Created' Created container
Nov 15 02:01:48 af867b kubelet[27751]: I1115 02:01:48.783830 27751 factory.go:112] Using factory "docker" for container "/kubepods/burstable/pod97270c63-c9a8-11e7-89f4-c6b053eac242/e430c4a0a025c95f870505c4ec608f56ad8cd6517f0e425139fb82fb5bfd6211"
Nov 15 02:01:48 af867b kubelet[27751]: I1115 02:01:48.786161 27751 kuberuntime_manager.go:705] Creating container &Container{Name:dnsmasq,Image:gcr.io/google_containers/k8s-dns-dnsmasq-nanny-amd64:1.14.5,Command:[],Args:[-v=2 -logtostderr -configDir=/etc/k8s/dns/dnsmasq-nanny -restartDnsmasq=true -- -k --cache-size=1000 --log-facility=- --server=/cluster.local/127.0.0.1#10053 --server=/in-addr.arpa/127.0.0.1#10053 --server=/ip6.arpa/127.0.0.1#10053],WorkingDir:,Ports:[{dns 0 53 UDP } {dns-tcp 0 53 TCP }],Env:[],Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{150 -3} {<nil>} 150m DecimalSI},memory: {{20971520 0} {<nil>} 20Mi BinarySI},},},VolumeMounts:[{kube-dns-config false /etc/k8s/dns/dnsmasq-nanny <nil>} {kube-dns-token-987zv true /var/run/secrets/kubernetes.io/serviceaccount <nil>}],LivenessProbe:&Probe{Handler:Handler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthcheck/dnsmasq,Port:10054,Host:,Scheme:HTTP,HTTPHeaders:[],},TCPSocket:nil,},InitialDelaySeconds:60,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:5,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[],TerminationMessagePolicy:File,} in pod kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242)
Nov 15 02:01:48 af867b kubelet[27751]: I1115 02:01:48.786608 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-dns-545bc4bfd4-zvfqd", UID:"97270c63-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"477", FieldPath:"spec.containers{kubedns}"}): type: 'Normal' reason: 'Started' Started container
Nov 15 02:01:48 af867b kubelet[27751]: I1115 02:01:48.789551 27751 manager.go:932] Added container: "/kubepods/burstable/pod97270c63-c9a8-11e7-89f4-c6b053eac242/e430c4a0a025c95f870505c4ec608f56ad8cd6517f0e425139fb82fb5bfd6211" (aliases: [k8s_kubedns_kube-dns-545bc4bfd4-zvfqd_kube-system_97270c63-c9a8-11e7-89f4-c6b053eac242_0 e430c4a0a025c95f870505c4ec608f56ad8cd6517f0e425139fb82fb5bfd6211], namespace: "docker")
Nov 15 02:01:48 af867b kubelet[27751]: I1115 02:01:48.789724 27751 handler.go:325] Added event &{/kubepods/burstable/pod97270c63-c9a8-11e7-89f4-c6b053eac242/e430c4a0a025c95f870505c4ec608f56ad8cd6517f0e425139fb82fb5bfd6211 2017-11-15 02:01:48.723194711 +0000 UTC containerCreation {<nil>}}
Nov 15 02:01:48 af867b kubelet[27751]: I1115 02:01:48.789763 27751 container.go:409] Start housekeeping for container "/kubepods/burstable/pod97270c63-c9a8-11e7-89f4-c6b053eac242/e430c4a0a025c95f870505c4ec608f56ad8cd6517f0e425139fb82fb5bfd6211"
Nov 15 02:01:48 af867b kubelet[27751]: I1115 02:01:48.799920 27751 kuberuntime_image.go:46] Pulling image "gcr.io/google_containers/k8s-dns-dnsmasq-nanny-amd64:1.14.5" without credentials
Nov 15 02:01:48 af867b kubelet[27751]: I1115 02:01:48.800007 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-dns-545bc4bfd4-zvfqd", UID:"97270c63-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"477", FieldPath:"spec.containers{dnsmasq}"}): type: 'Normal' reason: 'Pulling' pulling image "gcr.io/google_containers/k8s-dns-dnsmasq-nanny-amd64:1.14.5"
Nov 15 02:01:49 af867b kubelet[27751]: I1115 02:01:49.185023 27751 generic.go:146] GenericPLEG: 97270c63-c9a8-11e7-89f4-c6b053eac242/e430c4a0a025c95f870505c4ec608f56ad8cd6517f0e425139fb82fb5bfd6211: non-existent -> running
Nov 15 02:01:49 af867b kubelet[27751]: I1115 02:01:49.186274 27751 kuberuntime_manager.go:833] getSandboxIDByPodUID got sandbox IDs ["57f84a53f69dda718423e2a72b069d129afa4d226103816e6fa21ee5fab8b82b"] for pod "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:01:49 af867b kubelet[27751]: I1115 02:01:49.197224 27751 generic.go:345] PLEG: Write status for kube-dns-545bc4bfd4-zvfqd/kube-system: &container.PodStatus{ID:"97270c63-c9a8-11e7-89f4-c6b053eac242", Name:"kube-dns-545bc4bfd4-zvfqd", Namespace:"kube-system", IP:"10.32.0.2", ContainerStatuses:[]*container.ContainerStatus{(*container.ContainerStatus)(0xc4217d6a80)}, SandboxStatuses:[]*runtime.PodSandboxStatus{(*runtime.PodSandboxStatus)(0xc421a515e0)}} (err: <nil>)
Nov 15 02:01:49 af867b kubelet[27751]: I1115 02:01:49.197288 27751 kubelet.go:1871] SyncLoop (PLEG): "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242)", event: &pleg.PodLifecycleEvent{ID:"97270c63-c9a8-11e7-89f4-c6b053eac242", Type:"ContainerStarted", Data:"e430c4a0a025c95f870505c4ec608f56ad8cd6517f0e425139fb82fb5bfd6211"}
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.523854 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.531924 27751 qos_container_manager_linux.go:320] [ContainerManager]: Updated QoS cgroup configuration
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.562647 27751 kubelet.go:1222] Container garbage collection succeeded
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827105 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/run-docker-netns-default.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827142 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/run-docker-netns-default.mount", but ignoring.
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827152 27751 manager.go:901] ignoring container "/system.slice/run-docker-netns-default.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827159 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827166 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper.mount", but ignoring.
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827173 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827179 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-faffc5ce0cf0da39c68377dff0a611eb436ee74dfce11692aecd1e723311ef80.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827189 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-faffc5ce0cf0da39c68377dff0a611eb436ee74dfce11692aecd1e723311ef80.mount", but ignoring.
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827198 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-faffc5ce0cf0da39c68377dff0a611eb436ee74dfce11692aecd1e723311ef80.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827205 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/dev-hugepages.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827211 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/dev-hugepages.mount", but ignoring.
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827218 27751 manager.go:901] ignoring container "/system.slice/dev-hugepages.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827224 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-fe18d337ce42bd3a4d2aacb1349be028274681d8073fbe5d6377946e815fa810-shm.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827231 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-fe18d337ce42bd3a4d2aacb1349be028274681d8073fbe5d6377946e815fa810-shm.mount", but ignoring.
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827239 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-containers-fe18d337ce42bd3a4d2aacb1349be028274681d8073fbe5d6377946e815fa810-shm.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827247 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-9d97c593c8a0b1ef2442048619ba4466fad52c408e686c9285a48450f3a6669f.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827254 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-9d97c593c8a0b1ef2442048619ba4466fad52c408e686c9285a48450f3a6669f.mount", but ignoring.
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827263 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-9d97c593c8a0b1ef2442048619ba4466fad52c408e686c9285a48450f3a6669f.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827270 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-82408ec054eef4c0bce3519663bc7e5cbaff7a4fea1d8b7490fa3cc5a74bac60.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827277 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-82408ec054eef4c0bce3519663bc7e5cbaff7a4fea1d8b7490fa3cc5a74bac60.mount", but ignoring.
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827285 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-82408ec054eef4c0bce3519663bc7e5cbaff7a4fea1d8b7490fa3cc5a74bac60.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827292 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/proc-xen.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827297 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/proc-xen.mount", but ignoring.
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827304 27751 manager.go:901] ignoring container "/system.slice/proc-xen.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827309 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-439f518d46ec0d596518b7c17c503c52a85dd4da21d8a56ec174d5b6b3b98e03-shm.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827316 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-439f518d46ec0d596518b7c17c503c52a85dd4da21d8a56ec174d5b6b3b98e03-shm.mount", but ignoring.
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827325 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-containers-439f518d46ec0d596518b7c17c503c52a85dd4da21d8a56ec174d5b6b3b98e03-shm.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827332 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-ffde707e35c082af5481df3bb46e12f0eaf2bf20d3f28b499b6739ec31319af8.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827339 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-ffde707e35c082af5481df3bb46e12f0eaf2bf20d3f28b499b6739ec31319af8.mount", but ignoring.
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827347 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-ffde707e35c082af5481df3bb46e12f0eaf2bf20d3f28b499b6739ec31319af8.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827354 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-d5076ad0e9fb270d1b8c4ff7cdadbf32d6b1e30dc42ae24dfbc5cb01bb5aa934-shm.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827361 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-d5076ad0e9fb270d1b8c4ff7cdadbf32d6b1e30dc42ae24dfbc5cb01bb5aa934-shm.mount", but ignoring.
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827369 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-containers-d5076ad0e9fb270d1b8c4ff7cdadbf32d6b1e30dc42ae24dfbc5cb01bb5aa934-shm.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827375 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-19b06ca14052c17f92bf03d5316fd7642fe4fee6bf193755575355be3e87a13d-shm.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827382 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-19b06ca14052c17f92bf03d5316fd7642fe4fee6bf193755575355be3e87a13d-shm.mount", but ignoring.
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827391 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-containers-19b06ca14052c17f92bf03d5316fd7642fe4fee6bf193755575355be3e87a13d-shm.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827397 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-8c430fd9b74591760bf92ca717db0989293743473d027f75b1dfded4d0661504.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827404 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-8c430fd9b74591760bf92ca717db0989293743473d027f75b1dfded4d0661504.mount", but ignoring.
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827413 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-8c430fd9b74591760bf92ca717db0989293743473d027f75b1dfded4d0661504.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827420 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/dev-mqueue.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827425 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/dev-mqueue.mount", but ignoring.
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827432 27751 manager.go:901] ignoring container "/system.slice/dev-mqueue.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827436 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-29c1400e918b28b9982788b4c8b07c46b1767efcd8b96000e44bc78da639994b.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827444 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-29c1400e918b28b9982788b4c8b07c46b1767efcd8b96000e44bc78da639994b.mount", but ignoring.
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827452 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-29c1400e918b28b9982788b4c8b07c46b1767efcd8b96000e44bc78da639994b.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827459 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/-.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827465 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/-.mount", but ignoring.
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827471 27751 manager.go:901] ignoring container "/system.slice/-.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827476 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-cd27b64716d8032aa3ea595ec121a222dc749f2f903149a780eb711106fb0c53.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827483 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-cd27b64716d8032aa3ea595ec121a222dc749f2f903149a780eb711106fb0c53.mount", but ignoring.
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827492 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-cd27b64716d8032aa3ea595ec121a222dc749f2f903149a780eb711106fb0c53.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827498 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/sys-kernel-debug.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827504 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/sys-kernel-debug.mount", but ignoring.
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827511 27751 manager.go:901] ignoring container "/system.slice/sys-kernel-debug.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827516 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-23d8892c53722cfb4d3aa6c5b126ad2674a7b8041b4873d30b98faefad6f5b44.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827523 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-23d8892c53722cfb4d3aa6c5b126ad2674a7b8041b4873d30b98faefad6f5b44.mount", but ignoring.
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827531 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-23d8892c53722cfb4d3aa6c5b126ad2674a7b8041b4873d30b98faefad6f5b44.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827541 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-7d2ebdadb8c541a4839058dd46b491ef9e14209821b47e7d86b539de402fbef5.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827549 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-7d2ebdadb8c541a4839058dd46b491ef9e14209821b47e7d86b539de402fbef5.mount", but ignoring.
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827558 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-7d2ebdadb8c541a4839058dd46b491ef9e14209821b47e7d86b539de402fbef5.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827565 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/boot.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827571 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/boot.mount", but ignoring.
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827577 27751 manager.go:901] ignoring container "/system.slice/boot.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827582 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/sys-kernel-config.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827588 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/sys-kernel-config.mount", but ignoring.
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827595 27751 manager.go:901] ignoring container "/system.slice/sys-kernel-config.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827600 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-984a35c7cf6d04c0af5fcd260872a7b6556aa9cd8b38bc1a1cf635972c1d6511.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827607 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-984a35c7cf6d04c0af5fcd260872a7b6556aa9cd8b38bc1a1cf635972c1d6511.mount", but ignoring.
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827616 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-984a35c7cf6d04c0af5fcd260872a7b6556aa9cd8b38bc1a1cf635972c1d6511.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827623 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-kubelet-pods-b77b0858\\x2dc9a8\\x2d11e7\\x2d89f4\\x2dc6b053eac242-volumes-kubernetes.io\\x7esecret-weave\\x2dnet\\x2dtoken\\x2drn6j7.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827631 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-kubelet-pods-b77b0858\\x2dc9a8\\x2d11e7\\x2d89f4\\x2dc6b053eac242-volumes-kubernetes.io\\x7esecret-weave\\x2dnet\\x2dtoken\\x2drn6j7.mount", but ignoring.
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827642 27751 manager.go:901] ignoring container "/system.slice/var-lib-kubelet-pods-b77b0858\\x2dc9a8\\x2d11e7\\x2d89f4\\x2dc6b053eac242-volumes-kubernetes.io\\x7esecret-weave\\x2dnet\\x2dtoken\\x2drn6j7.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827650 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-91510886f3beb621e5d04309d502352ff78392e46405f21633802da4f7047069-shm.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827657 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-91510886f3beb621e5d04309d502352ff78392e46405f21633802da4f7047069-shm.mount", but ignoring.
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827666 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-containers-91510886f3beb621e5d04309d502352ff78392e46405f21633802da4f7047069-shm.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827672 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/u01-applicationSpace.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827678 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/u01-applicationSpace.mount", but ignoring.
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827685 27751 manager.go:901] ignoring container "/system.slice/u01-applicationSpace.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827690 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-9ba72245089e9b36e901a816a7446b004869c57dbf880db7daeb3edac0f81e51.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827697 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-9ba72245089e9b36e901a816a7446b004869c57dbf880db7daeb3edac0f81e51.mount", but ignoring.
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827705 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-9ba72245089e9b36e901a816a7446b004869c57dbf880db7daeb3edac0f81e51.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827728 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-6ddc96a63fb65ccd1743152170b39274e0eb3e4ccc4948bf2c0f5a6130ee9ce7.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827737 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-6ddc96a63fb65ccd1743152170b39274e0eb3e4ccc4948bf2c0f5a6130ee9ce7.mount", but ignoring.
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827745 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-6ddc96a63fb65ccd1743152170b39274e0eb3e4ccc4948bf2c0f5a6130ee9ce7.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827752 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/run-user-1000.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827758 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/run-user-1000.mount", but ignoring.
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827764 27751 manager.go:901] ignoring container "/system.slice/run-user-1000.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827769 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-be14b0552100f929052c8e94e1b7cf587c4218e321d811314e58f2cf312f1c0f-shm.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827777 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-be14b0552100f929052c8e94e1b7cf587c4218e321d811314e58f2cf312f1c0f-shm.mount", but ignoring.
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827785 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-containers-be14b0552100f929052c8e94e1b7cf587c4218e321d811314e58f2cf312f1c0f-shm.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827792 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-kubelet-pods-9729c03a\\x2dc9a8\\x2d11e7\\x2d89f4\\x2dc6b053eac242-volumes-kubernetes.io\\x7esecret-kube\\x2dproxy\\x2dtoken\\x2dgqhfs.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827800 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-kubelet-pods-9729c03a\\x2dc9a8\\x2d11e7\\x2d89f4\\x2dc6b053eac242-volumes-kubernetes.io\\x7esecret-kube\\x2dproxy\\x2dtoken\\x2dgqhfs.mount", but ignoring.
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.827810 27751 manager.go:901] ignoring container "/system.slice/var-lib-kubelet-pods-9729c03a\\x2dc9a8\\x2d11e7\\x2d89f4\\x2dc6b053eac242-volumes-kubernetes.io\\x7esecret-kube\\x2dproxy\\x2dtoken\\x2dgqhfs.mount"
Nov 15 02:01:50 af867b kubelet[27751]: I1115 02:01:50.892425 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=true reason: message:
Nov 15 02:01:51 af867b kubelet[27751]: I1115 02:01:51.683977 27751 eviction_manager.go:221] eviction manager: synchronize housekeeping
Nov 15 02:01:51 af867b kubelet[27751]: I1115 02:01:51.731864 27751 helpers.go:871] eviction manager: observations: signal=memory.available, available: 6483764Ki, capacity: 7393360Ki, time: 2017-11-15 02:01:46.953790411 +0000 UTC
Nov 15 02:01:51 af867b kubelet[27751]: I1115 02:01:51.731907 27751 helpers.go:871] eviction manager: observations: signal=nodefs.available, available: 7392020Ki, capacity: 10198Mi, time: 2017-11-15 02:01:46.953790411 +0000 UTC
Nov 15 02:01:51 af867b kubelet[27751]: I1115 02:01:51.731918 27751 helpers.go:871] eviction manager: observations: signal=nodefs.inodesFree, available: 10384617, capacity: 10208Ki, time: 2017-11-15 02:01:46.953790411 +0000 UTC
Nov 15 02:01:51 af867b kubelet[27751]: I1115 02:01:51.731926 27751 helpers.go:871] eviction manager: observations: signal=imagefs.available, available: 40779264Ki, capacity: 45Gi, time: 2017-11-15 02:01:46.953790411 +0000 UTC
Nov 15 02:01:51 af867b kubelet[27751]: I1115 02:01:51.731935 27751 helpers.go:873] eviction manager: observations: signal=allocatableMemory.available, available: 6934588Ki, capacity: 7393360Ki
Nov 15 02:01:51 af867b kubelet[27751]: I1115 02:01:51.731941 27751 helpers.go:873] eviction manager: observations: signal=allocatableNodeFs.available, available: 9624040228, capacity: 10198Mi
Nov 15 02:01:51 af867b kubelet[27751]: I1115 02:01:51.731958 27751 eviction_manager.go:325] eviction manager: no resources are starved
Nov 15 02:01:52 af867b kubelet[27751]: I1115 02:01:52.523462 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:01:52 af867b kubelet[27751]: I1115 02:01:52.959942 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 2379, Path: /health
Nov 15 02:01:52 af867b kubelet[27751]: I1115 02:01:52.959975 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:01:53 af867b kubelet[27751]: I1115 02:01:53.197611 27751 worker.go:164] Probe target container not found: kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242) - kubedns
Nov 15 02:01:53 af867b kubelet[27751]: I1115 02:01:53.211104 27751 http.go:96] Probe succeeded for http://127.0.0.1:2379/health, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:01:53 GMT] Content-Length:[18] Content-Type:[text/plain; charset=utf-8]] 0xc421a68ae0 18 [] true false map[] 0xc420b09900 <nil>}
Nov 15 02:01:53 af867b kubelet[27751]: I1115 02:01:53.211154 27751 prober.go:113] Liveness probe for "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250):etcd" succeeded
Nov 15 02:01:53 af867b kubelet[27751]: I1115 02:01:53.617384 27751 prober.go:160] HTTP-Probe Host: https://127.0.0.1, Port: 6443, Path: /healthz
Nov 15 02:01:53 af867b kubelet[27751]: I1115 02:01:53.617421 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:01:53 af867b kubelet[27751]: I1115 02:01:53.623935 27751 http.go:96] Probe succeeded for https://127.0.0.1:6443/healthz, Response: {200 OK 200 HTTP/2.0 2 0 map[Content-Type:[text/plain; charset=utf-8] Content-Length:[2] Date:[Wed, 15 Nov 2017 02:01:53 GMT]] 0xc421c300c0 2 [] false false map[] 0xc420ee2100 0xc420e78580}
Nov 15 02:01:53 af867b kubelet[27751]: I1115 02:01:53.623983 27751 prober.go:113] Liveness probe for "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d):kube-apiserver" succeeded
Nov 15 02:01:53 af867b kubelet[27751]: I1115 02:01:53.834530 27751 worker.go:164] Probe target container not found: kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242) - dnsmasq
Nov 15 02:01:54 af867b kubelet[27751]: I1115 02:01:54.504581 27751 worker.go:164] Probe target container not found: kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242) - sidecar
Nov 15 02:01:54 af867b kubelet[27751]: I1115 02:01:54.523430 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:01:55 af867b kubelet[27751]: I1115 02:01:55.362751 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 6784, Path: /status
Nov 15 02:01:55 af867b kubelet[27751]: I1115 02:01:55.362783 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:01:55 af867b kubelet[27751]: I1115 02:01:55.364544 27751 http.go:96] Probe succeeded for http://127.0.0.1:6784/status, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:01:55 GMT] Content-Length:[445] Content-Type:[text/plain; charset=utf-8]] 0xc42014e320 445 [] true false map[] 0xc420ee2e00 <nil>}
Nov 15 02:01:55 af867b kubelet[27751]: I1115 02:01:55.364583 27751 prober.go:113] Liveness probe for "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242):weave" succeeded
Nov 15 02:01:55 af867b kubelet[27751]: I1115 02:01:55.894203 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=true reason: message:
Nov 15 02:01:56 af867b kubelet[27751]: I1115 02:01:56.523448 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:01:57 af867b kubelet[27751]: I1115 02:01:57.506117 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10252, Path: /healthz
Nov 15 02:01:57 af867b kubelet[27751]: I1115 02:01:57.506155 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:01:57 af867b kubelet[27751]: I1115 02:01:57.507008 27751 http.go:96] Probe succeeded for http://127.0.0.1:10252/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Type:[text/plain; charset=utf-8] Date:[Wed, 15 Nov 2017 02:01:57 GMT] Content-Length:[2]] 0xc420f3e6c0 2 [] true false map[] 0xc4211a8100 <nil>}
Nov 15 02:01:57 af867b kubelet[27751]: I1115 02:01:57.507051 27751 prober.go:113] Liveness probe for "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9):kube-controller-manager" succeeded
Nov 15 02:01:57 af867b kubelet[27751]: I1115 02:01:57.599531 27751 worker.go:164] Probe target container not found: kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242) - kubedns
Nov 15 02:01:57 af867b kubelet[27751]: I1115 02:01:57.681641 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10251, Path: /healthz
Nov 15 02:01:57 af867b kubelet[27751]: I1115 02:01:57.681665 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:01:57 af867b kubelet[27751]: I1115 02:01:57.683216 27751 http.go:96] Probe succeeded for http://127.0.0.1:10251/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:01:57 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc4214f7140 2 [] true false map[] 0xc4211a8e00 <nil>}
Nov 15 02:01:57 af867b kubelet[27751]: I1115 02:01:57.683274 27751 prober.go:113] Liveness probe for "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373):kube-scheduler" succeeded
Nov 15 02:01:58 af867b kubelet[27751]: I1115 02:01:58.523446 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:02:00 af867b kubelet[27751]: I1115 02:02:00.498824 27751 kube_docker_client.go:330] Pulling image "gcr.io/google_containers/k8s-dns-dnsmasq-nanny-amd64:1.14.5": "62106dacdb76: Extracting [====================================> ] 6.685MB/9.193MB"
Nov 15 02:02:00 af867b kubelet[27751]: I1115 02:02:00.528253 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:02:00 af867b kubelet[27751]: I1115 02:02:00.758862 27751 kube_docker_client.go:333] Stop pulling image "gcr.io/google_containers/k8s-dns-dnsmasq-nanny-amd64:1.14.5": "Status: Downloaded newer image for gcr.io/google_containers/k8s-dns-dnsmasq-nanny-amd64:1.14.5"
Nov 15 02:02:00 af867b kubelet[27751]: I1115 02:02:00.760121 27751 kuberuntime_container.go:100] Generating ref for container dnsmasq: &v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-dns-545bc4bfd4-zvfqd", UID:"97270c63-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"477", FieldPath:"spec.containers{dnsmasq}"}
Nov 15 02:02:00 af867b kubelet[27751]: I1115 02:02:00.760164 27751 container_manager_linux.go:634] Calling devicePluginHandler AllocateDevices
Nov 15 02:02:00 af867b kubelet[27751]: I1115 02:02:00.760221 27751 kubelet_pods.go:123] container: kube-system/kube-dns-545bc4bfd4-zvfqd/dnsmasq podIP: "10.32.0.2" creating hosts mount: true
Nov 15 02:02:00 af867b kubelet[27751]: I1115 02:02:00.760671 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-dns-545bc4bfd4-zvfqd", UID:"97270c63-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"477", FieldPath:"spec.containers{dnsmasq}"}): type: 'Normal' reason: 'Pulled' Successfully pulled image "gcr.io/google_containers/k8s-dns-dnsmasq-nanny-amd64:1.14.5"
Nov 15 02:02:00 af867b kubelet[27751]: I1115 02:02:00.762040 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/burstable/pod97270c63-c9a8-11e7-89f4-c6b053eac242"
Nov 15 02:02:00 af867b kubelet[27751]: I1115 02:02:00.896997 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=true reason: message:
Nov 15 02:02:00 af867b kubelet[27751]: I1115 02:02:00.917696 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-dns-545bc4bfd4-zvfqd", UID:"97270c63-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"477", FieldPath:"spec.containers{dnsmasq}"}): type: 'Normal' reason: 'Created' Created container
Nov 15 02:02:01 af867b kubelet[27751]: I1115 02:02:01.072690 27751 factory.go:112] Using factory "docker" for container "/kubepods/burstable/pod97270c63-c9a8-11e7-89f4-c6b053eac242/49d5926d73b0ca6b1f1b8b2f76e9e3623ecaecb67d89db69f0f94b5dfe890a5a"
Nov 15 02:02:01 af867b kubelet[27751]: I1115 02:02:01.073665 27751 kuberuntime_manager.go:705] Creating container &Container{Name:sidecar,Image:gcr.io/google_containers/k8s-dns-sidecar-amd64:1.14.5,Command:[],Args:[--v=2 --logtostderr --probe=kubedns,127.0.0.1:10053,kubernetes.default.svc.cluster.local,5,A --probe=dnsmasq,127.0.0.1:53,kubernetes.default.svc.cluster.local,5,A],WorkingDir:,Ports:[{metrics 0 10054 TCP }],Env:[],Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {<nil>} 10m DecimalSI},memory: {{20971520 0} {<nil>} 20Mi BinarySI},},},VolumeMounts:[{kube-dns-token-987zv true /var/run/secrets/kubernetes.io/serviceaccount <nil>}],LivenessProbe:&Probe{Handler:Handler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/metrics,Port:10054,Host:,Scheme:HTTP,HTTPHeaders:[],},TCPSocket:nil,},InitialDelaySeconds:60,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:5,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[],TerminationMessagePolicy:File,} in pod kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242)
Nov 15 02:02:01 af867b kubelet[27751]: I1115 02:02:01.073943 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-dns-545bc4bfd4-zvfqd", UID:"97270c63-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"477", FieldPath:"spec.containers{dnsmasq}"}): type: 'Normal' reason: 'Started' Started container
Nov 15 02:02:01 af867b kubelet[27751]: I1115 02:02:01.075623 27751 kuberuntime_image.go:46] Pulling image "gcr.io/google_containers/k8s-dns-sidecar-amd64:1.14.5" without credentials
Nov 15 02:02:01 af867b kubelet[27751]: I1115 02:02:01.075698 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-dns-545bc4bfd4-zvfqd", UID:"97270c63-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"477", FieldPath:"spec.containers{sidecar}"}): type: 'Normal' reason: 'Pulling' pulling image "gcr.io/google_containers/k8s-dns-sidecar-amd64:1.14.5"
Nov 15 02:02:01 af867b kubelet[27751]: I1115 02:02:01.075749 27751 manager.go:932] Added container: "/kubepods/burstable/pod97270c63-c9a8-11e7-89f4-c6b053eac242/49d5926d73b0ca6b1f1b8b2f76e9e3623ecaecb67d89db69f0f94b5dfe890a5a" (aliases: [k8s_dnsmasq_kube-dns-545bc4bfd4-zvfqd_kube-system_97270c63-c9a8-11e7-89f4-c6b053eac242_0 49d5926d73b0ca6b1f1b8b2f76e9e3623ecaecb67d89db69f0f94b5dfe890a5a], namespace: "docker")
Nov 15 02:02:01 af867b kubelet[27751]: I1115 02:02:01.075878 27751 handler.go:325] Added event &{/kubepods/burstable/pod97270c63-c9a8-11e7-89f4-c6b053eac242/49d5926d73b0ca6b1f1b8b2f76e9e3623ecaecb67d89db69f0f94b5dfe890a5a 2017-11-15 02:02:01.000194711 +0000 UTC containerCreation {<nil>}}
Nov 15 02:02:01 af867b kubelet[27751]: I1115 02:02:01.075927 27751 container.go:409] Start housekeeping for container "/kubepods/burstable/pod97270c63-c9a8-11e7-89f4-c6b053eac242/49d5926d73b0ca6b1f1b8b2f76e9e3623ecaecb67d89db69f0f94b5dfe890a5a"
Nov 15 02:02:01 af867b kubelet[27751]: I1115 02:02:01.256577 27751 generic.go:146] GenericPLEG: 97270c63-c9a8-11e7-89f4-c6b053eac242/49d5926d73b0ca6b1f1b8b2f76e9e3623ecaecb67d89db69f0f94b5dfe890a5a: non-existent -> running
Nov 15 02:02:01 af867b kubelet[27751]: I1115 02:02:01.257422 27751 kuberuntime_manager.go:833] getSandboxIDByPodUID got sandbox IDs ["57f84a53f69dda718423e2a72b069d129afa4d226103816e6fa21ee5fab8b82b"] for pod "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:02:01 af867b kubelet[27751]: I1115 02:02:01.266524 27751 generic.go:345] PLEG: Write status for kube-dns-545bc4bfd4-zvfqd/kube-system: &container.PodStatus{ID:"97270c63-c9a8-11e7-89f4-c6b053eac242", Name:"kube-dns-545bc4bfd4-zvfqd", Namespace:"kube-system", IP:"10.32.0.2", ContainerStatuses:[]*container.ContainerStatus{(*container.ContainerStatus)(0xc421a07a40), (*container.ContainerStatus)(0xc421a07c00)}, SandboxStatuses:[]*runtime.PodSandboxStatus{(*runtime.PodSandboxStatus)(0xc4214ed4f0)}} (err: <nil>)
Nov 15 02:02:01 af867b kubelet[27751]: I1115 02:02:01.266599 27751 kubelet.go:1871] SyncLoop (PLEG): "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242)", event: &pleg.PodLifecycleEvent{ID:"97270c63-c9a8-11e7-89f4-c6b053eac242", Type:"ContainerStarted", Data:"49d5926d73b0ca6b1f1b8b2f76e9e3623ecaecb67d89db69f0f94b5dfe890a5a"}
Nov 15 02:02:01 af867b kubelet[27751]: I1115 02:02:01.732127 27751 eviction_manager.go:221] eviction manager: synchronize housekeeping
Nov 15 02:02:01 af867b kubelet[27751]: I1115 02:02:01.783017 27751 helpers.go:871] eviction manager: observations: signal=nodefs.available, available: 7392020Ki, capacity: 10198Mi, time: 2017-11-15 02:01:46.953790411 +0000 UTC
Nov 15 02:02:01 af867b kubelet[27751]: I1115 02:02:01.783071 27751 helpers.go:871] eviction manager: observations: signal=nodefs.inodesFree, available: 10384617, capacity: 10208Ki, time: 2017-11-15 02:01:46.953790411 +0000 UTC
Nov 15 02:02:01 af867b kubelet[27751]: I1115 02:02:01.783084 27751 helpers.go:871] eviction manager: observations: signal=imagefs.available, available: 40779264Ki, capacity: 45Gi, time: 2017-11-15 02:01:46.953790411 +0000 UTC
Nov 15 02:02:01 af867b kubelet[27751]: I1115 02:02:01.783095 27751 helpers.go:873] eviction manager: observations: signal=allocatableMemory.available, available: 6933156Ki, capacity: 7393360Ki
Nov 15 02:02:01 af867b kubelet[27751]: I1115 02:02:01.783104 27751 helpers.go:873] eviction manager: observations: signal=allocatableNodeFs.available, available: 9624040228, capacity: 10198Mi
Nov 15 02:02:01 af867b kubelet[27751]: I1115 02:02:01.783114 27751 helpers.go:871] eviction manager: observations: signal=memory.available, available: 6483764Ki, capacity: 7393360Ki, time: 2017-11-15 02:01:46.953790411 +0000 UTC
Nov 15 02:02:01 af867b kubelet[27751]: I1115 02:02:01.783139 27751 eviction_manager.go:325] eviction manager: no resources are starved
Nov 15 02:02:02 af867b kubelet[27751]: I1115 02:02:02.523464 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:02:02 af867b kubelet[27751]: I1115 02:02:02.959981 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 2379, Path: /health
Nov 15 02:02:02 af867b kubelet[27751]: I1115 02:02:02.960027 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:02:03 af867b kubelet[27751]: I1115 02:02:03.197632 27751 worker.go:164] Probe target container not found: kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242) - kubedns
Nov 15 02:02:03 af867b kubelet[27751]: I1115 02:02:03.211645 27751 http.go:96] Probe succeeded for http://127.0.0.1:2379/health, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:02:03 GMT] Content-Length:[18] Content-Type:[text/plain; charset=utf-8]] 0xc421914d60 18 [] true false map[] 0xc420afda00 <nil>}
Nov 15 02:02:03 af867b kubelet[27751]: I1115 02:02:03.211697 27751 prober.go:113] Liveness probe for "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250):etcd" succeeded
Nov 15 02:02:03 af867b kubelet[27751]: I1115 02:02:03.617264 27751 prober.go:160] HTTP-Probe Host: https://127.0.0.1, Port: 6443, Path: /healthz
Nov 15 02:02:03 af867b kubelet[27751]: I1115 02:02:03.617302 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:02:03 af867b kubelet[27751]: I1115 02:02:03.625346 27751 http.go:96] Probe succeeded for https://127.0.0.1:6443/healthz, Response: {200 OK 200 HTTP/2.0 2 0 map[Content-Length:[2] Date:[Wed, 15 Nov 2017 02:02:03 GMT] Content-Type:[text/plain; charset=utf-8]] 0xc42186a980 2 [] false false map[] 0xc42110a000 0xc4215dc9a0}
Nov 15 02:02:03 af867b kubelet[27751]: I1115 02:02:03.625386 27751 prober.go:113] Liveness probe for "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d):kube-apiserver" succeeded
Nov 15 02:02:03 af867b kubelet[27751]: I1115 02:02:03.834498 27751 worker.go:164] Probe target container not found: kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242) - dnsmasq
Nov 15 02:02:04 af867b kubelet[27751]: I1115 02:02:04.504601 27751 worker.go:164] Probe target container not found: kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242) - sidecar
Nov 15 02:02:04 af867b kubelet[27751]: I1115 02:02:04.523446 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:02:05 af867b kubelet[27751]: I1115 02:02:05.362708 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 6784, Path: /status
Nov 15 02:02:05 af867b kubelet[27751]: I1115 02:02:05.362737 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:02:05 af867b kubelet[27751]: I1115 02:02:05.364098 27751 http.go:96] Probe succeeded for http://127.0.0.1:6784/status, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:02:05 GMT] Content-Length:[445] Content-Type:[text/plain; charset=utf-8]] 0xc421101a80 445 [] true false map[] 0xc4211a9900 <nil>}
Nov 15 02:02:05 af867b kubelet[27751]: I1115 02:02:05.364144 27751 prober.go:113] Liveness probe for "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242):weave" succeeded
Nov 15 02:02:05 af867b kubelet[27751]: I1115 02:02:05.898259 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=true reason: message:
Nov 15 02:02:06 af867b kubelet[27751]: I1115 02:02:06.523454 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:02:07 af867b kubelet[27751]: I1115 02:02:07.506101 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10252, Path: /healthz
Nov 15 02:02:07 af867b kubelet[27751]: I1115 02:02:07.506135 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:02:07 af867b kubelet[27751]: I1115 02:02:07.507912 27751 http.go:96] Probe succeeded for http://127.0.0.1:10252/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:02:07 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc4216f1320 2 [] true false map[] 0xc420ee2600 <nil>}
Nov 15 02:02:07 af867b kubelet[27751]: I1115 02:02:07.507954 27751 prober.go:113] Liveness probe for "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9):kube-controller-manager" succeeded
Nov 15 02:02:07 af867b kubelet[27751]: I1115 02:02:07.599520 27751 worker.go:164] Probe target container not found: kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242) - kubedns
Nov 15 02:02:07 af867b kubelet[27751]: I1115 02:02:07.681619 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10251, Path: /healthz
Nov 15 02:02:07 af867b kubelet[27751]: I1115 02:02:07.681640 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:02:07 af867b kubelet[27751]: I1115 02:02:07.682651 27751 http.go:96] Probe succeeded for http://127.0.0.1:10251/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:02:07 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc421c300a0 2 [] true false map[] 0xc420ee2f00 <nil>}
Nov 15 02:02:07 af867b kubelet[27751]: I1115 02:02:07.682690 27751 prober.go:113] Liveness probe for "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373):kube-scheduler" succeeded
Nov 15 02:02:08 af867b kubelet[27751]: I1115 02:02:08.523449 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:02:10 af867b kubelet[27751]: I1115 02:02:10.198378 27751 kube_docker_client.go:333] Stop pulling image "gcr.io/google_containers/k8s-dns-sidecar-amd64:1.14.5": "Status: Downloaded newer image for gcr.io/google_containers/k8s-dns-sidecar-amd64:1.14.5"
Nov 15 02:02:10 af867b kubelet[27751]: I1115 02:02:10.199624 27751 kuberuntime_container.go:100] Generating ref for container sidecar: &v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-dns-545bc4bfd4-zvfqd", UID:"97270c63-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"477", FieldPath:"spec.containers{sidecar}"}
Nov 15 02:02:10 af867b kubelet[27751]: I1115 02:02:10.199671 27751 container_manager_linux.go:634] Calling devicePluginHandler AllocateDevices
Nov 15 02:02:10 af867b kubelet[27751]: I1115 02:02:10.199742 27751 kubelet_pods.go:123] container: kube-system/kube-dns-545bc4bfd4-zvfqd/sidecar podIP: "10.32.0.2" creating hosts mount: true
Nov 15 02:02:10 af867b kubelet[27751]: I1115 02:02:10.200380 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-dns-545bc4bfd4-zvfqd", UID:"97270c63-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"477", FieldPath:"spec.containers{sidecar}"}): type: 'Normal' reason: 'Pulled' Successfully pulled image "gcr.io/google_containers/k8s-dns-sidecar-amd64:1.14.5"
Nov 15 02:02:10 af867b kubelet[27751]: W1115 02:02:10.201773 27751 kuberuntime_container.go:191] Non-root verification doesn't support non-numeric user (nobody)
Nov 15 02:02:10 af867b kubelet[27751]: I1115 02:02:10.207112 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/burstable/pod97270c63-c9a8-11e7-89f4-c6b053eac242"
Nov 15 02:02:10 af867b kubelet[27751]: I1115 02:02:10.420911 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-dns-545bc4bfd4-zvfqd", UID:"97270c63-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"477", FieldPath:"spec.containers{sidecar}"}): type: 'Normal' reason: 'Created' Created container
Nov 15 02:02:10 af867b kubelet[27751]: I1115 02:02:10.523403 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:02:10 af867b kubelet[27751]: I1115 02:02:10.561246 27751 factory.go:112] Using factory "docker" for container "/kubepods/burstable/pod97270c63-c9a8-11e7-89f4-c6b053eac242/12935c093ed0d04807a97f6b78b64a7a87f39b36f15aad612e8e47dba47ac96f"
Nov 15 02:02:10 af867b kubelet[27751]: I1115 02:02:10.571800 27751 manager.go:932] Added container: "/kubepods/burstable/pod97270c63-c9a8-11e7-89f4-c6b053eac242/12935c093ed0d04807a97f6b78b64a7a87f39b36f15aad612e8e47dba47ac96f" (aliases: [k8s_sidecar_kube-dns-545bc4bfd4-zvfqd_kube-system_97270c63-c9a8-11e7-89f4-c6b053eac242_0 12935c093ed0d04807a97f6b78b64a7a87f39b36f15aad612e8e47dba47ac96f], namespace: "docker")
Nov 15 02:02:10 af867b kubelet[27751]: I1115 02:02:10.572406 27751 handler.go:325] Added event &{/kubepods/burstable/pod97270c63-c9a8-11e7-89f4-c6b053eac242/12935c093ed0d04807a97f6b78b64a7a87f39b36f15aad612e8e47dba47ac96f 2017-11-15 02:02:10.485194711 +0000 UTC containerCreation {<nil>}}
Nov 15 02:02:10 af867b kubelet[27751]: I1115 02:02:10.573602 27751 container.go:409] Start housekeeping for container "/kubepods/burstable/pod97270c63-c9a8-11e7-89f4-c6b053eac242/12935c093ed0d04807a97f6b78b64a7a87f39b36f15aad612e8e47dba47ac96f"
Nov 15 02:02:10 af867b kubelet[27751]: I1115 02:02:10.585585 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-dns-545bc4bfd4-zvfqd", UID:"97270c63-c9a8-11e7-89f4-c6b053eac242", APIVersion:"v1", ResourceVersion:"477", FieldPath:"spec.containers{sidecar}"}): type: 'Normal' reason: 'Started' Started container
Nov 15 02:02:10 af867b kubelet[27751]: I1115 02:02:10.899758 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=true reason: message:
Nov 15 02:02:11 af867b kubelet[27751]: I1115 02:02:11.312027 27751 generic.go:146] GenericPLEG: 97270c63-c9a8-11e7-89f4-c6b053eac242/12935c093ed0d04807a97f6b78b64a7a87f39b36f15aad612e8e47dba47ac96f: non-existent -> running
Nov 15 02:02:11 af867b kubelet[27751]: I1115 02:02:11.313617 27751 kuberuntime_manager.go:833] getSandboxIDByPodUID got sandbox IDs ["57f84a53f69dda718423e2a72b069d129afa4d226103816e6fa21ee5fab8b82b"] for pod "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:02:11 af867b kubelet[27751]: I1115 02:02:11.326550 27751 generic.go:345] PLEG: Write status for kube-dns-545bc4bfd4-zvfqd/kube-system: &container.PodStatus{ID:"97270c63-c9a8-11e7-89f4-c6b053eac242", Name:"kube-dns-545bc4bfd4-zvfqd", Namespace:"kube-system", IP:"10.32.0.2", ContainerStatuses:[]*container.ContainerStatus{(*container.ContainerStatus)(0xc421cc6700), (*container.ContainerStatus)(0xc421a06e00), (*container.ContainerStatus)(0xc421cc69a0)}, SandboxStatuses:[]*runtime.PodSandboxStatus{(*runtime.PodSandboxStatus)(0xc421ee1180)}} (err: <nil>)
Nov 15 02:02:11 af867b kubelet[27751]: I1115 02:02:11.326621 27751 kubelet.go:1871] SyncLoop (PLEG): "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242)", event: &pleg.PodLifecycleEvent{ID:"97270c63-c9a8-11e7-89f4-c6b053eac242", Type:"ContainerStarted", Data:"12935c093ed0d04807a97f6b78b64a7a87f39b36f15aad612e8e47dba47ac96f"}
Nov 15 02:02:11 af867b kubelet[27751]: I1115 02:02:11.326685 27751 kubelet_pods.go:1284] Generating status for "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:02:11 af867b kubelet[27751]: I1115 02:02:11.326992 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:02:11 af867b kubelet[27751]: I1115 02:02:11.336692 27751 config.go:282] Setting pods for source api
Nov 15 02:02:11 af867b kubelet[27751]: I1115 02:02:11.338509 27751 kubelet.go:1850] SyncLoop (RECONCILE, "api"): "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:02:11 af867b kubelet[27751]: I1115 02:02:11.336857 27751 status_manager.go:451] Status for pod "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242)" updated successfully: (2, {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:01:21 +0000 UTC Reason: Message:} {Type:Ready Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:01:21 +0000 UTC Reason:ContainersNotReady Message:containers with unready status: [kubedns]} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:01:21 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP:10.32.0.2 StartTime:2017-11-15 02:01:21 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:dnsmasq State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 02:02:01 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:gcr.io/google_containers/k8s-dns-dnsmasq-nanny-amd64:1.14.5 ImageID:docker-pullable://gcr.io/google_containers/k8s-dns-dnsmasq-nanny-amd64@sha256:46b933bb70270c8a02fa6b6f87d440f6f1fce1a5a2a719e164f83f7b109f7544 ContainerID:docker://49d5926d73b0ca6b1f1b8b2f76e9e3623ecaecb67d89db69f0f94b5dfe890a5a} {Name:kubedns State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 02:01:48 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:false RestartCount:0 Image:gcr.io/google_containers/k8s-dns-kube-dns-amd64:1.14.5 ImageID:docker-pullable://gcr.io/google_containers/k8s-dns-kube-dns-amd64@sha256:1a3fc069de481ae690188f6f1ba4664b5cc7760af37120f70c86505c79eea61d ContainerID:docker://e430c4a0a025c95f870505c4ec608f56ad8cd6517f0e425139fb82fb5bfd6211} {Name:sidecar State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 02:02:10 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Ima
Nov 15 02:02:11 af867b kubelet[27751]: ge:gcr.io/google_containers/k8s-dns-sidecar-amd64:1.14.5 ImageID:docker-pullable://gcr.io/google_containers/k8s-dns-sidecar-amd64@sha256:9aab42bf6a2a068b797fe7d91a5d8d915b10dbbc3d6f2b10492848debfba6044 ContainerID:docker://12935c093ed0d04807a97f6b78b64a7a87f39b36f15aad612e8e47dba47ac96f}] QOSClass:Burstable})
Nov 15 02:02:11 af867b kubelet[27751]: I1115 02:02:11.371074 27751 configmap.go:187] Setting up volume kube-dns-config for pod 97270c63-c9a8-11e7-89f4-c6b053eac242 at /var/lib/kubelet/pods/97270c63-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~configmap/kube-dns-config
Nov 15 02:02:11 af867b kubelet[27751]: I1115 02:02:11.371465 27751 secret.go:186] Setting up volume kube-dns-token-987zv for pod 97270c63-c9a8-11e7-89f4-c6b053eac242 at /var/lib/kubelet/pods/97270c63-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/kube-dns-token-987zv
Nov 15 02:02:11 af867b kubelet[27751]: I1115 02:02:11.373530 27751 configmap.go:218] Received configMap kube-system/kube-dns containing (0) pieces of data, 0 total bytes
Nov 15 02:02:11 af867b kubelet[27751]: I1115 02:02:11.373612 27751 atomic_writer.go:142] pod kube-system/kube-dns-545bc4bfd4-zvfqd volume kube-dns-config: no update required for target directory /var/lib/kubelet/pods/97270c63-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~configmap/kube-dns-config
Nov 15 02:02:11 af867b kubelet[27751]: I1115 02:02:11.374372 27751 secret.go:217] Received secret kube-system/kube-dns-token-987zv containing (3) pieces of data, 1896 total bytes
Nov 15 02:02:11 af867b kubelet[27751]: I1115 02:02:11.374519 27751 atomic_writer.go:142] pod kube-system/kube-dns-545bc4bfd4-zvfqd volume kube-dns-token-987zv: no update required for target directory /var/lib/kubelet/pods/97270c63-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/kube-dns-token-987zv
Nov 15 02:02:11 af867b kubelet[27751]: I1115 02:02:11.627250 27751 volume_manager.go:366] All volumes are attached and mounted for pod "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:02:11 af867b kubelet[27751]: I1115 02:02:11.627538 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:false CreateSandbox:false SandboxID:57f84a53f69dda718423e2a72b069d129afa4d226103816e6fa21ee5fab8b82b Attempt:0 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:02:11 af867b kubelet[27751]: I1115 02:02:11.783298 27751 eviction_manager.go:221] eviction manager: synchronize housekeeping
Nov 15 02:02:11 af867b kubelet[27751]: I1115 02:02:11.840972 27751 helpers.go:871] eviction manager: observations: signal=memory.available, available: 6470564Ki, capacity: 7393360Ki, time: 2017-11-15 02:02:06.156376344 +0000 UTC
Nov 15 02:02:11 af867b kubelet[27751]: I1115 02:02:11.841011 27751 helpers.go:871] eviction manager: observations: signal=nodefs.available, available: 7404548Ki, capacity: 10198Mi, time: 2017-11-15 02:02:06.156376344 +0000 UTC
Nov 15 02:02:11 af867b kubelet[27751]: I1115 02:02:11.841021 27751 helpers.go:871] eviction manager: observations: signal=nodefs.inodesFree, available: 10384529, capacity: 10208Ki, time: 2017-11-15 02:02:06.156376344 +0000 UTC
Nov 15 02:02:11 af867b kubelet[27751]: I1115 02:02:11.841029 27751 helpers.go:871] eviction manager: observations: signal=imagefs.available, available: 39696Mi, capacity: 45Gi, time: 2017-11-15 02:02:06.156376344 +0000 UTC
Nov 15 02:02:11 af867b kubelet[27751]: I1115 02:02:11.841036 27751 helpers.go:873] eviction manager: observations: signal=allocatableMemory.available, available: 6923584Ki, capacity: 7393360Ki
Nov 15 02:02:11 af867b kubelet[27751]: I1115 02:02:11.841042 27751 helpers.go:873] eviction manager: observations: signal=allocatableNodeFs.available, available: 9624040228, capacity: 10198Mi
Nov 15 02:02:11 af867b kubelet[27751]: I1115 02:02:11.841061 27751 eviction_manager.go:325] eviction manager: no resources are starved
Nov 15 02:02:12 af867b kubelet[27751]: I1115 02:02:12.331202 27751 kubelet_pods.go:1284] Generating status for "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:02:12 af867b kubelet[27751]: I1115 02:02:12.331420 27751 status_manager.go:325] Ignoring same status for pod "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242)", status: {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:01:21 +0000 UTC Reason: Message:} {Type:Ready Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:01:21 +0000 UTC Reason:ContainersNotReady Message:containers with unready status: [kubedns]} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:01:21 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP:10.32.0.2 StartTime:2017-11-15 02:01:21 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:dnsmasq State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 02:02:01 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:gcr.io/google_containers/k8s-dns-dnsmasq-nanny-amd64:1.14.5 ImageID:docker-pullable://gcr.io/google_containers/k8s-dns-dnsmasq-nanny-amd64@sha256:46b933bb70270c8a02fa6b6f87d440f6f1fce1a5a2a719e164f83f7b109f7544 ContainerID:docker://49d5926d73b0ca6b1f1b8b2f76e9e3623ecaecb67d89db69f0f94b5dfe890a5a} {Name:kubedns State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 02:01:48 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:false RestartCount:0 Image:gcr.io/google_containers/k8s-dns-kube-dns-amd64:1.14.5 ImageID:docker-pullable://gcr.io/google_containers/k8s-dns-kube-dns-amd64@sha256:1a3fc069de481ae690188f6f1ba4664b5cc7760af37120f70c86505c79eea61d ContainerID:docker://e430c4a0a025c95f870505c4ec608f56ad8cd6517f0e425139fb82fb5bfd6211} {Name:sidecar State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 02:02:10 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:
Nov 15 02:02:12 af867b kubelet[27751]: gcr.io/google_containers/k8s-dns-sidecar-amd64:1.14.5 ImageID:docker-pullable://gcr.io/google_containers/k8s-dns-sidecar-amd64@sha256:9aab42bf6a2a068b797fe7d91a5d8d915b10dbbc3d6f2b10492848debfba6044 ContainerID:docker://12935c093ed0d04807a97f6b78b64a7a87f39b36f15aad612e8e47dba47ac96f}] QOSClass:Burstable}
Nov 15 02:02:12 af867b kubelet[27751]: I1115 02:02:12.331616 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:02:12 af867b kubelet[27751]: I1115 02:02:12.373498 27751 configmap.go:187] Setting up volume kube-dns-config for pod 97270c63-c9a8-11e7-89f4-c6b053eac242 at /var/lib/kubelet/pods/97270c63-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~configmap/kube-dns-config
Nov 15 02:02:12 af867b kubelet[27751]: I1115 02:02:12.373499 27751 secret.go:186] Setting up volume kube-dns-token-987zv for pod 97270c63-c9a8-11e7-89f4-c6b053eac242 at /var/lib/kubelet/pods/97270c63-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/kube-dns-token-987zv
Nov 15 02:02:12 af867b kubelet[27751]: I1115 02:02:12.377133 27751 secret.go:217] Received secret kube-system/kube-dns-token-987zv containing (3) pieces of data, 1896 total bytes
Nov 15 02:02:12 af867b kubelet[27751]: I1115 02:02:12.377333 27751 atomic_writer.go:142] pod kube-system/kube-dns-545bc4bfd4-zvfqd volume kube-dns-token-987zv: no update required for target directory /var/lib/kubelet/pods/97270c63-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/kube-dns-token-987zv
Nov 15 02:02:12 af867b kubelet[27751]: I1115 02:02:12.377539 27751 configmap.go:218] Received configMap kube-system/kube-dns containing (0) pieces of data, 0 total bytes
Nov 15 02:02:12 af867b kubelet[27751]: I1115 02:02:12.377577 27751 atomic_writer.go:142] pod kube-system/kube-dns-545bc4bfd4-zvfqd volume kube-dns-config: no update required for target directory /var/lib/kubelet/pods/97270c63-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~configmap/kube-dns-config
Nov 15 02:02:12 af867b kubelet[27751]: I1115 02:02:12.523453 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:02:12 af867b kubelet[27751]: I1115 02:02:12.631843 27751 volume_manager.go:366] All volumes are attached and mounted for pod "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:02:12 af867b kubelet[27751]: I1115 02:02:12.632103 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:false CreateSandbox:false SandboxID:57f84a53f69dda718423e2a72b069d129afa4d226103816e6fa21ee5fab8b82b Attempt:0 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:02:12 af867b kubelet[27751]: I1115 02:02:12.959944 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 2379, Path: /health
Nov 15 02:02:12 af867b kubelet[27751]: I1115 02:02:12.959970 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:02:13 af867b kubelet[27751]: I1115 02:02:13.210831 27751 http.go:96] Probe succeeded for http://127.0.0.1:2379/health, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:02:13 GMT] Content-Length:[18] Content-Type:[text/plain; charset=utf-8]] 0xc421210fe0 18 [] true false map[] 0xc420afc300 <nil>}
Nov 15 02:02:13 af867b kubelet[27751]: I1115 02:02:13.210887 27751 prober.go:113] Liveness probe for "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250):etcd" succeeded
Nov 15 02:02:13 af867b kubelet[27751]: I1115 02:02:13.617255 27751 prober.go:160] HTTP-Probe Host: https://127.0.0.1, Port: 6443, Path: /healthz
Nov 15 02:02:13 af867b kubelet[27751]: I1115 02:02:13.617293 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:02:13 af867b kubelet[27751]: I1115 02:02:13.624277 27751 http.go:96] Probe succeeded for https://127.0.0.1:6443/healthz, Response: {200 OK 200 HTTP/2.0 2 0 map[Content-Type:[text/plain; charset=utf-8] Content-Length:[2] Date:[Wed, 15 Nov 2017 02:02:13 GMT]] 0xc421347aa0 2 [] false false map[] 0xc420afca00 0xc42132c580}
Nov 15 02:02:13 af867b kubelet[27751]: I1115 02:02:13.624322 27751 prober.go:113] Liveness probe for "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d):kube-apiserver" succeeded
Nov 15 02:02:14 af867b kubelet[27751]: I1115 02:02:14.523425 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:02:15 af867b kubelet[27751]: I1115 02:02:15.362686 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 6784, Path: /status
Nov 15 02:02:15 af867b kubelet[27751]: I1115 02:02:15.362708 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:02:15 af867b kubelet[27751]: I1115 02:02:15.364092 27751 http.go:96] Probe succeeded for http://127.0.0.1:6784/status, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Type:[text/plain; charset=utf-8] Date:[Wed, 15 Nov 2017 02:02:15 GMT] Content-Length:[445]] 0xc4211006c0 445 [] true false map[] 0xc4211a9500 <nil>}
Nov 15 02:02:15 af867b kubelet[27751]: I1115 02:02:15.364130 27751 prober.go:113] Liveness probe for "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242):weave" succeeded
Nov 15 02:02:15 af867b kubelet[27751]: I1115 02:02:15.523438 27751 kubelet.go:1890] SyncLoop (SYNC): 1 pods; kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)
Nov 15 02:02:15 af867b kubelet[27751]: I1115 02:02:15.523519 27751 kubelet_pods.go:1284] Generating status for "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:02:15 af867b kubelet[27751]: I1115 02:02:15.523674 27751 status_manager.go:325] Ignoring same status for pod "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)", status: {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:59:18 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:59:20 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:59:20 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP:10.196.65.210 StartTime:2017-11-15 01:59:18 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:kube-proxy State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 01:59:20 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:gcr.io/google_containers/kube-proxy-amd64:v1.8.3 ImageID:docker-pullable://gcr.io/google_containers/kube-proxy-amd64@sha256:63210bc9690144d41126a646caf03a3d76ddc6d06b8bad119d468193c3e90c24 ContainerID:docker://7ea397ec6048e25ce044b9edba43fe0ef2ed54803c9f3516c7b6e780d2c60a34}] QOSClass:BestEffort}
Nov 15 02:02:15 af867b kubelet[27751]: I1115 02:02:15.523852 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:02:15 af867b kubelet[27751]: I1115 02:02:15.580233 27751 secret.go:186] Setting up volume kube-proxy-token-gqhfs for pod 9729c03a-c9a8-11e7-89f4-c6b053eac242 at /var/lib/kubelet/pods/9729c03a-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/kube-proxy-token-gqhfs
Nov 15 02:02:15 af867b kubelet[27751]: I1115 02:02:15.580236 27751 configmap.go:187] Setting up volume kube-proxy for pod 9729c03a-c9a8-11e7-89f4-c6b053eac242 at /var/lib/kubelet/pods/9729c03a-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~configmap/kube-proxy
Nov 15 02:02:15 af867b kubelet[27751]: I1115 02:02:15.582739 27751 secret.go:217] Received secret kube-system/kube-proxy-token-gqhfs containing (3) pieces of data, 1904 total bytes
Nov 15 02:02:15 af867b kubelet[27751]: I1115 02:02:15.582807 27751 configmap.go:218] Received configMap kube-system/kube-proxy containing (1) pieces of data, 407 total bytes
Nov 15 02:02:15 af867b kubelet[27751]: I1115 02:02:15.582894 27751 atomic_writer.go:142] pod kube-system/kube-proxy-nnsjf volume kube-proxy-token-gqhfs: no update required for target directory /var/lib/kubelet/pods/9729c03a-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/kube-proxy-token-gqhfs
Nov 15 02:02:15 af867b kubelet[27751]: I1115 02:02:15.582903 27751 atomic_writer.go:142] pod kube-system/kube-proxy-nnsjf volume kube-proxy: no update required for target directory /var/lib/kubelet/pods/9729c03a-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~configmap/kube-proxy
Nov 15 02:02:15 af867b kubelet[27751]: I1115 02:02:15.825852 27751 volume_manager.go:366] All volumes are attached and mounted for pod "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:02:15 af867b kubelet[27751]: I1115 02:02:15.825972 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:false CreateSandbox:false SandboxID:91510886f3beb621e5d04309d502352ff78392e46405f21633802da4f7047069 Attempt:0 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:02:15 af867b kubelet[27751]: I1115 02:02:15.904598 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=true reason: message:
Nov 15 02:02:16 af867b kubelet[27751]: I1115 02:02:16.523453 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:02:17 af867b kubelet[27751]: I1115 02:02:17.506784 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10252, Path: /healthz
Nov 15 02:02:17 af867b kubelet[27751]: I1115 02:02:17.506824 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:02:17 af867b kubelet[27751]: I1115 02:02:17.508698 27751 http.go:96] Probe succeeded for http://127.0.0.1:10252/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:02:17 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc42192c060 2 [] true false map[] 0xc420c91b00 <nil>}
Nov 15 02:02:17 af867b kubelet[27751]: I1115 02:02:17.508763 27751 prober.go:113] Liveness probe for "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9):kube-controller-manager" succeeded
Nov 15 02:02:17 af867b kubelet[27751]: I1115 02:02:17.599548 27751 prober.go:160] HTTP-Probe Host: http://10.32.0.2, Port: 8081, Path: /readiness
Nov 15 02:02:17 af867b kubelet[27751]: I1115 02:02:17.599577 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:02:17 af867b kubelet[27751]: I1115 02:02:17.599911 27751 status_manager.go:203] Container readiness unchanged (false): "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242)" - "docker://e430c4a0a025c95f870505c4ec608f56ad8cd6517f0e425139fb82fb5bfd6211"
Nov 15 02:02:17 af867b kubelet[27751]: I1115 02:02:17.601188 27751 http.go:96] Probe succeeded for http://10.32.0.2:8081/readiness, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:02:17 GMT] Content-Length:[3] Content-Type:[text/plain; charset=utf-8]] 0xc421d02300 3 [] true false map[] 0xc420b09400 <nil>}
Nov 15 02:02:17 af867b kubelet[27751]: I1115 02:02:17.601244 27751 prober.go:113] Readiness probe for "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242):kubedns" succeeded
Nov 15 02:02:17 af867b kubelet[27751]: I1115 02:02:17.609030 27751 config.go:282] Setting pods for source api
Nov 15 02:02:17 af867b kubelet[27751]: I1115 02:02:17.610347 27751 kubelet.go:1850] SyncLoop (RECONCILE, "api"): "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:02:17 af867b kubelet[27751]: I1115 02:02:17.613936 27751 status_manager.go:451] Status for pod "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242)" updated successfully: (3, {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:01:21 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:02:17 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:01:21 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP:10.32.0.2 StartTime:2017-11-15 02:01:21 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:dnsmasq State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 02:02:01 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:gcr.io/google_containers/k8s-dns-dnsmasq-nanny-amd64:1.14.5 ImageID:docker-pullable://gcr.io/google_containers/k8s-dns-dnsmasq-nanny-amd64@sha256:46b933bb70270c8a02fa6b6f87d440f6f1fce1a5a2a719e164f83f7b109f7544 ContainerID:docker://49d5926d73b0ca6b1f1b8b2f76e9e3623ecaecb67d89db69f0f94b5dfe890a5a} {Name:kubedns State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 02:01:48 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:gcr.io/google_containers/k8s-dns-kube-dns-amd64:1.14.5 ImageID:docker-pullable://gcr.io/google_containers/k8s-dns-kube-dns-amd64@sha256:1a3fc069de481ae690188f6f1ba4664b5cc7760af37120f70c86505c79eea61d ContainerID:docker://e430c4a0a025c95f870505c4ec608f56ad8cd6517f0e425139fb82fb5bfd6211} {Name:sidecar State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 02:02:10 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Image:gcr.io/google_containers/k8s-dns-sidecar-amd64:1.14.5 Imag
Nov 15 02:02:17 af867b kubelet[27751]: eID:docker-pullable://gcr.io/google_containers/k8s-dns-sidecar-amd64@sha256:9aab42bf6a2a068b797fe7d91a5d8d915b10dbbc3d6f2b10492848debfba6044 ContainerID:docker://12935c093ed0d04807a97f6b78b64a7a87f39b36f15aad612e8e47dba47ac96f}] QOSClass:Burstable})
Nov 15 02:02:17 af867b kubelet[27751]: I1115 02:02:17.681660 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10251, Path: /healthz
Nov 15 02:02:17 af867b kubelet[27751]: I1115 02:02:17.681696 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:02:17 af867b kubelet[27751]: I1115 02:02:17.683472 27751 http.go:96] Probe succeeded for http://127.0.0.1:10251/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:02:17 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc421d22100 2 [] true false map[] 0xc420b09c00 <nil>}
Nov 15 02:02:17 af867b kubelet[27751]: I1115 02:02:17.683512 27751 prober.go:113] Liveness probe for "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373):kube-scheduler" succeeded
Nov 15 02:02:18 af867b kubelet[27751]: I1115 02:02:18.523431 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:02:20 af867b kubelet[27751]: I1115 02:02:20.523423 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:02:20 af867b kubelet[27751]: I1115 02:02:20.912239 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=true reason: message:
Nov 15 02:16:23 af867b kubelet[27751]: I1115 02:16:23.199466 27751 prober.go:113] Liveness probe for "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242):kubedns" succeede
Nov 15 02:16:23 af867b kubelet[27751]: I1115 02:16:23.463080 27751 http.go:96] Probe succeeded for http://127.0.0.1:2379/health, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:1:23 GMT] Content-Length:[18] Content-Type:[text/plain; charset=utf-8]] 0xc421528840 18 [] true false map[] 0xc420ee2e00 <nil>}
Nov 15 02:16:23 af867b kubelet[27751]: I1115 02:16:23.463189 27751 prober.go:113] Liveness probe for "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250):etcd" succeeded
Nov 15 02:16:23 af867b kubelet[27751]: I1115 02:16:23.617350 27751 prober.go:160] HTTP-Probe Host: https://127.0.0.1, Port: 6443, Path: /healthz
Nov 15 02:16:23 af867b kubelet[27751]: I1115 02:16:23.617410 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:16:23 af867b kubelet[27751]: I1115 02:16:23.630597 27751 http.go:96] Probe succeeded for https://127.0.0.1:6443/healthz, Response: {200 OK 200 HTTP/2.0 2 0 map[Content-Type:[text/plain;charset=utf-8] Content-Length:[2] Date:[Wed, 15 Nov 2017 02:16:23 GMT]] 0xc4215296a0 2 [] false false map[] 0xc420ee3500 0xc42140e840}
Nov 15 02:16:23 af867b kubelet[27751]: I1115 02:16:23.630643 27751 prober.go:113] Liveness probe for "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d):kube-apiserver" succeeded
Nov 15 02:16:23 af867b kubelet[27751]: I1115 02:16:23.834501 27751 prober.go:160] HTTP-Probe Host: http://10.32.0.2, Port: 10054, Path: /healthcheck/dnsmasq
Nov 15 02:16:23 af867b kubelet[27751]: I1115 02:16:23.834541 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:16:23 af867b kubelet[27751]: I1115 02:16:23.835543 27751 http.go:96] Probe succeeded for http://10.32.0.2:10054/healthcheck/dnsmasq, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Lengt:[51] Content-Type:[application/json] Date:[Wed, 15 Nov 2017 02:16:23 GMT]] 0xc4214f7fa0 51 [] true false map[] 0xc420d4b700 <nil>}
Nov 15 02:16:23 af867b kubelet[27751]: I1115 02:16:23.835592 27751 prober.go:113] Liveness probe for "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242):dnsmasq" succeede
Nov 15 02:16:24 af867b kubelet[27751]: I1115 02:16:24.504617 27751 prober.go:160] HTTP-Probe Host: http://10.32.0.2, Port: 10054, Path: /metrics
Nov 15 02:16:24 af867b kubelet[27751]: I1115 02:16:24.504654 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:16:24 af867b kubelet[27751]: I1115 02:16:24.510100 27751 http.go:96] Probe succeeded for http://10.32.0.2:10054/metrics, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Type:[text/plain;version=0.0.4] Date:[Wed, 15 Nov 2017 02:16:24 GMT]] 0xc4213588e0 -1 [] true true map[] 0xc420ee3b00 <nil>}
Nov 15 02:16:24 af867b kubelet[27751]: I1115 02:16:24.510163 27751 prober.go:113] Liveness probe for "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242):sidecar" succeede
Nov 15 02:16:24 af867b kubelet[27751]: I1115 02:16:24.523449 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:16:25 af867b kubelet[27751]: I1115 02:16:25.362734 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 6784, Path: /status
Nov 15 02:16:25 af867b kubelet[27751]: I1115 02:16:25.362783 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:16:25 af867b kubelet[27751]: I1115 02:16:25.365672 27751 http.go:96] Probe succeeded for http://127.0.0.1:6784/status, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Length:[445] Conten-Type:[text/plain; charset=utf-8] Date:[Wed, 15 Nov 2017 02:16:25 GMT]] 0xc420aa7dc0 445 [] true false map[] 0xc4211a8700 <nil>}
Nov 15 02:16:25 af867b kubelet[27751]: I1115 02:16:25.365740 27751 prober.go:113] Liveness probe for "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242):weave" succeeded
Nov 15 02:16:26 af867b kubelet[27751]: I1115 02:16:26.216297 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=true reason: mesage:
Nov 15 02:16:26 af867b kubelet[27751]: I1115 02:16:26.373008 27751 eviction_manager.go:221] eviction manager: synchronize housekeeping
Nov 15 02:16:26 af867b kubelet[27751]: I1115 02:16:26.431639 27751 helpers.go:871] eviction manager: observations: signal=nodefs.inodesFree, available: 10384476, capacity: 10208Ki, time: 2017-11-5 02:16:24.845571562 +0000 UTC
Nov 15 02:16:26 af867b kubelet[27751]: I1115 02:16:26.431677 27751 helpers.go:871] eviction manager: observations: signal=imagefs.available, available: 39632Mi, capacity: 45Gi, time: 2017-11-15 0:16:24.845571562 +0000 UTC
Nov 15 02:16:26 af867b kubelet[27751]: I1115 02:16:26.431689 27751 helpers.go:873] eviction manager: observations: signal=allocatableMemory.available, available: 6884156Ki, capacity: 7393360Ki
Nov 15 02:16:26 af867b kubelet[27751]: I1115 02:16:26.431697 27751 helpers.go:873] eviction manager: observations: signal=allocatableNodeFs.available, available: 9624040228, capacity: 10198Mi
Nov 15 02:16:26 af867b kubelet[27751]: I1115 02:16:26.431704 27751 helpers.go:871] eviction manager: observations: signal=memory.available, available: 5609352Ki, capacity: 7393360Ki, time: 2017-1-15 02:16:24.845571562 +0000 UTC
Nov 15 02:16:26 af867b kubelet[27751]: I1115 02:16:26.432749 27751 helpers.go:871] eviction manager: observations: signal=nodefs.available, available: 7399676Ki, capacity: 10198Mi, time: 2017-11-5 02:16:24.845571562 +0000 UTC
Nov 15 02:16:26 af867b kubelet[27751]: I1115 02:16:26.432776 27751 eviction_manager.go:325] eviction manager: no resources are starved
Nov 15 02:16:26 af867b kubelet[27751]: I1115 02:16:26.523494 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:16:27 af867b kubelet[27751]: I1115 02:16:27.599530 27751 prober.go:160] HTTP-Probe Host: http://10.32.0.2, Port: 8081, Path: /readiness
Nov 15 02:16:27 af867b kubelet[27751]: I1115 02:16:27.599566 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:16:27 af867b kubelet[27751]: I1115 02:16:27.601375 27751 http.go:96] Probe succeeded for http://10.32.0.2:8081/readiness, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 0:16:27 GMT] Content-Length:[3] Content-Type:[text/plain; charset=utf-8]] 0xc421bf2a80 3 [] true false map[] 0xc42110a600 <nil>}
Nov 15 02:16:27 af867b kubelet[27751]: I1115 02:16:27.601416 27751 prober.go:113] Readiness probe for "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242):kubedns" succeedd
Nov 15 02:16:27 af867b kubelet[27751]: I1115 02:16:27.681665 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10251, Path: /healthz
Nov 15 02:16:27 af867b kubelet[27751]: I1115 02:16:27.681703 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:16:27 af867b kubelet[27751]: I1115 02:16:27.683017 27751 http.go:96] Probe succeeded for http://127.0.0.1:10251/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 0216:27 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc421bf2e60 2 [] true false map[] 0xc42110ab00 <nil>}
Nov 15 02:16:27 af867b kubelet[27751]: I1115 02:16:27.683056 27751 prober.go:113] Liveness probe for "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373):kube-scheduler" succeeded
Nov 15 02:16:28 af867b kubelet[27751]: I1115 02:16:28.523439 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:16:29 af867b kubelet[27751]: I1115 02:16:29.214125 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10252, Path: /healthz
Nov 15 02:16:29 af867b kubelet[27751]: I1115 02:16:29.214182 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:16:29 af867b kubelet[27751]: I1115 02:16:29.217451 27751 http.go:96] Probe succeeded for http://127.0.0.1:10252/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Length:[2] Conten-Type:[text/plain; charset=utf-8] Date:[Wed, 15 Nov 2017 02:16:29 GMT]] 0xc42203fd80 2 [] true false map[] 0xc420b08700 <nil>}
Nov 15 02:16:29 af867b kubelet[27751]: I1115 02:16:29.217500 27751 prober.go:113] Liveness probe for "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9):kube-controller-anager" succeeded
Nov 15 02:16:30 af867b kubelet[27751]: I1115 02:16:30.523436 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:16:31 af867b kubelet[27751]: I1115 02:16:31.217902 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=true reason: mesage:
Nov 15 02:16:32 af867b kubelet[27751]: I1115 02:16:32.523520 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:16:32 af867b kubelet[27751]: I1115 02:16:32.959955 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 2379, Path: /health
Nov 15 02:16:32 af867b kubelet[27751]: I1115 02:16:32.960000 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:16:33 af867b kubelet[27751]: I1115 02:16:33.197613 27751 prober.go:160] HTTP-Probe Host: http://10.32.0.2, Port: 10054, Path: /healthcheck/kubedns
Nov 15 02:16:33 af867b kubelet[27751]: I1115 02:16:33.197659 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:16:33 af867b kubelet[27751]: I1115 02:16:33.198672 27751 http.go:96] Probe succeeded for http://10.32.0.2:10054/healthcheck/kubedns, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15Nov 2017 02:16:33 GMT] Content-Length:[51] Content-Type:[application/json]] 0xc421176480 51 [] true false map[] 0xc420c90b00 <nil>}
Nov 15 02:16:33 af867b kubelet[27751]: I1115 02:16:33.198745 27751 prober.go:113] Liveness probe for "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242):kubedns" succeede
Nov 15 02:16:33 af867b kubelet[27751]: I1115 02:16:33.462133 27751 http.go:96] Probe succeeded for http://127.0.0.1:2379/health, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:1:33 GMT] Content-Length:[18] Content-Type:[text/plain; charset=utf-8]] 0xc421fe56c0 18 [] true false map[] 0xc420c90400 <nil>}
Nov 15 02:16:33 af867b kubelet[27751]: I1115 02:16:33.462212 27751 prober.go:113] Liveness probe for "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250):etcd" succeeded
Nov 15 02:16:33 af867b kubelet[27751]: I1115 02:16:33.617411 27751 prober.go:160] HTTP-Probe Host: https://127.0.0.1, Port: 6443, Path: /healthz
Nov 15 02:16:33 af867b kubelet[27751]: I1115 02:16:33.617459 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:16:33 af867b kubelet[27751]: I1115 02:16:33.624589 27751 http.go:96] Probe succeeded for https://127.0.0.1:6443/healthz, Response: {200 OK 200 HTTP/2.0 2 0 map[Content-Length:[2] Date:[ed, 15 Nov 2017 02:16:33 GMT] Content-Type:[text/plain; charset=utf-8]] 0xc421177a60 2 [] false false map[] 0xc42110bf00 0xc421f1ed10}
Nov 15 02:16:33 af867b kubelet[27751]: I1115 02:16:33.624640 27751 prober.go:113] Liveness probe for "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d):kube-apiserver" succeeded
Nov 15 02:16:33 af867b kubelet[27751]: I1115 02:16:33.834514 27751 prober.go:160] HTTP-Probe Host: http://10.32.0.2, Port: 10054, Path: /healthcheck/dnsmasq
Nov 15 02:16:33 af867b kubelet[27751]: I1115 02:16:33.834562 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:16:33 af867b kubelet[27751]: I1115 02:16:33.836107 27751 http.go:96] Probe succeeded for http://10.32.0.2:10054/healthcheck/dnsmasq, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Lengt:[51] Content-Type:[application/json] Date:[Wed, 15 Nov 2017 02:16:33 GMT]] 0xc421fe59c0 51 [] true false map[] 0xc420c91100 <nil>}
Nov 15 02:16:33 af867b kubelet[27751]: I1115 02:16:33.836170 27751 prober.go:113] Liveness probe for "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242):dnsmasq" succeede
Nov 15 02:16:34 af867b kubelet[27751]: I1115 02:16:34.504612 27751 prober.go:160] HTTP-Probe Host: http://10.32.0.2, Port: 10054, Path: /metrics
Nov 15 02:16:34 af867b kubelet[27751]: I1115 02:16:34.504656 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:16:34 af867b kubelet[27751]: I1115 02:16:34.509868 27751 http.go:96] Probe succeeded for http://10.32.0.2:10054/metrics, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Type:[text/plain;version=0.0.4] Date:[Wed, 15 Nov 2017 02:16:34 GMT]] 0xc421a21300 -1 [] true true map[] 0xc420c91600 <nil>}
Nov 15 02:16:34 af867b kubelet[27751]: I1115 02:16:34.509915 27751 prober.go:113] Liveness probe for "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242):sidecar" succeede
Nov 15 02:16:34 af867b kubelet[27751]: I1115 02:16:34.523469 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:16:35 af867b kubelet[27751]: I1115 02:16:35.362813 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 6784, Path: /status
Nov 15 02:16:35 af867b kubelet[27751]: I1115 02:16:35.362889 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:16:35 af867b kubelet[27751]: I1115 02:16:35.364510 27751 http.go:96] Probe succeeded for http://127.0.0.1:6784/status, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:1:35 GMT] Content-Length:[445] Content-Type:[text/plain; charset=utf-8]] 0xc421a685a0 445 [] true false map[] 0xc420c91b00 <nil>}
Nov 15 02:16:35 af867b kubelet[27751]: I1115 02:16:35.364570 27751 prober.go:113] Liveness probe for "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242):weave" succeeded
Nov 15 02:16:36 af867b kubelet[27751]: I1115 02:16:36.219654 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=true reason: mesage:
Nov 15 02:16:36 af867b kubelet[27751]: I1115 02:16:36.432958 27751 eviction_manager.go:221] eviction manager: synchronize housekeeping
Nov 15 02:16:36 af867b kubelet[27751]: I1115 02:16:36.481409 27751 helpers.go:871] eviction manager: observations: signal=memory.available, available: 5609352Ki, capacity: 7393360Ki, time: 2017-1-15 02:16:24.845571562 +0000 UTC
Nov 15 02:16:36 af867b kubelet[27751]: I1115 02:16:36.481459 27751 helpers.go:871] eviction manager: observations: signal=nodefs.available, available: 7399676Ki, capacity: 10198Mi, time: 2017-11-5 02:16:24.845571562 +0000 UTC
Nov 15 02:16:36 af867b kubelet[27751]: I1115 02:16:36.481470 27751 helpers.go:871] eviction manager: observations: signal=nodefs.inodesFree, available: 10384476, capacity: 10208Ki, time: 2017-11-5 02:16:24.845571562 +0000 UTC
Nov 15 02:16:36 af867b kubelet[27751]: I1115 02:16:36.481479 27751 helpers.go:871] eviction manager: observations: signal=imagefs.available, available: 39632Mi, capacity: 45Gi, time: 2017-11-15 0:16:24.845571562 +0000 UTC
Nov 15 02:16:36 af867b kubelet[27751]: I1115 02:16:36.481488 27751 helpers.go:873] eviction manager: observations: signal=allocatableMemory.available, available: 6884104Ki, capacity: 7393360Ki
Nov 15 02:16:36 af867b kubelet[27751]: I1115 02:16:36.481496 27751 helpers.go:873] eviction manager: observations: signal=allocatableNodeFs.available, available: 9624040228, capacity: 10198Mi
Nov 15 02:16:36 af867b kubelet[27751]: I1115 02:16:36.481518 27751 eviction_manager.go:325] eviction manager: no resources are starved
Nov 15 02:16:36 af867b kubelet[27751]: I1115 02:16:36.523444 27751 kubelet.go:1890] SyncLoop (SYNC): 1 pods; etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)
Nov 15 02:16:36 af867b kubelet[27751]: I1115 02:16:36.523506 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:16:36 af867b kubelet[27751]: I1115 02:16:36.525996 27751 kubelet_pods.go:1284] Generating status for "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)"
Nov 15 02:16:36 af867b kubelet[27751]: I1115 02:16:36.526365 27751 status_manager.go:325] Ignoring same status for pod "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)", status: {Phase:unning Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:59 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-1-15 01:58:55 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP:10.196.65.210 StartTime:2017-11-15 01:58:55 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:etcd Stte:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 01:58:58 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestartCount:0 Imag:gcr.io/google_containers/etcd-amd64:3.0.17 ImageID:docker-pullable://gcr.io/google_containers/etcd-amd64@sha256:d83d3545e06fb035db8512e33bd44afb55dea007a3abd7b17742d3ac6d235940 ContainerID:docker:/ab147bb1b65b2001333417bb7654896e6aadb25ce71a8c48c94ae802a2e0197f}] QOSClass:BestEffort}
Nov 15 02:16:36 af867b kubelet[27751]: I1115 02:16:36.526694 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250"
Nov 15 02:16:36 af867b kubelet[27751]: I1115 02:16:36.827493 27751 volume_manager.go:366] All volumes are attached and mounted for pod "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)"
Nov 15 02:16:36 af867b kubelet[27751]: I1115 02:16:36.827767 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:false CreateSandbox:false SandboxID:d5076ad0e9fb270d1b8c4ff7cdadbf32db1e30dc42ae24dfbc5cb01bb5aa934 Attempt:0 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250)"
Nov 15 02:16:37 af867b kubelet[27751]: I1115 02:16:37.599607 27751 prober.go:160] HTTP-Probe Host: http://10.32.0.2, Port: 8081, Path: /readiness
Nov 15 02:16:37 af867b kubelet[27751]: I1115 02:16:37.599675 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:16:37 af867b kubelet[27751]: I1115 02:16:37.601397 27751 http.go:96] Probe succeeded for http://10.32.0.2:8081/readiness, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 0:16:37 GMT] Content-Length:[3] Content-Type:[text/plain; charset=utf-8]] 0xc4210a26e0 3 [] true false map[] 0xc421179600 <nil>}
Nov 15 02:16:37 af867b kubelet[27751]: I1115 02:16:37.601454 27751 prober.go:113] Readiness probe for "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242):kubedns" succeedd
Nov 15 02:16:37 af867b kubelet[27751]: I1115 02:16:37.681650 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10251, Path: /healthz
Nov 15 02:16:37 af867b kubelet[27751]: I1115 02:16:37.681687 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:16:37 af867b kubelet[27751]: I1115 02:16:37.682968 27751 http.go:96] Probe succeeded for http://127.0.0.1:10251/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 0216:37 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc4210a2840 2 [] true false map[] 0xc421179900 <nil>}
Nov 15 02:16:37 af867b kubelet[27751]: I1115 02:16:37.683016 27751 prober.go:113] Liveness probe for "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373):kube-scheduler" succeeded
Nov 15 02:16:38 af867b kubelet[27751]: I1115 02:16:38.523481 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:16:39 af867b kubelet[27751]: I1115 02:16:39.214109 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10252, Path: /healthz
Nov 15 02:16:39 af867b kubelet[27751]: I1115 02:16:39.214176 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:16:39 af867b kubelet[27751]: I1115 02:16:39.216105 27751 http.go:96] Probe succeeded for http://127.0.0.1:10252/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Length:[2] Conten-Type:[text/plain; charset=utf-8] Date:[Wed, 15 Nov 2017 02:16:39 GMT]] 0xc420db3d00 2 [] true false map[] 0xc420b09400 <nil>}
Nov 15 02:16:39 af867b kubelet[27751]: I1115 02:16:39.216194 27751 prober.go:113] Liveness probe for "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9):kube-controller-anager" succeeded
Nov 15 02:16:39 af867b kubelet[27751]: I1115 02:16:39.524766 27751 kubelet.go:1890] SyncLoop (SYNC): 1 pods; kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)
Nov 15 02:16:39 af867b kubelet[27751]: I1115 02:16:39.524874 27751 kubelet_pods.go:1284] Generating status for "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:16:39 af867b kubelet[27751]: I1115 02:16:39.525057 27751 status_manager.go:325] Ignoring same status for pod "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)", status {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:59:18 +0000 UTC Reason: Message:} {Type:Ready Status:True LastrobeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:59:20 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTie:2017-11-15 01:59:20 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP:10.196.65.210 StartTime:2017-11-15 01:59:18 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Nam:kube-proxy State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 01:59:20 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:true RestrtCount:0 Image:gcr.io/google_containers/kube-proxy-amd64:v1.8.3 ImageID:docker-pullable://gcr.io/google_containers/kube-proxy-amd64@sha256:63210bc9690144d41126a646caf03a3d76ddc6d06b8bad119d468193ce90c24 ContainerID:docker://7ea397ec6048e25ce044b9edba43fe0ef2ed54803c9f3516c7b6e780d2c60a34}] QOSClass:BestEffort}
Nov 15 02:16:39 af867b kubelet[27751]: I1115 02:16:39.525234 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b53eac242)"
Nov 15 02:16:39 af867b kubelet[27751]: I1115 02:16:39.541677 27751 configmap.go:187] Setting up volume kube-proxy for pod 9729c03a-c9a8-11e7-89f4-c6b053eac242 at /var/lib/kubelet/pods/9729c03a-c98-11e7-89f4-c6b053eac242/volumes/kubernetes.io~configmap/kube-proxy
Nov 15 02:16:39 af867b kubelet[27751]: I1115 02:16:39.541959 27751 secret.go:186] Setting up volume kube-proxy-token-gqhfs for pod 9729c03a-c9a8-11e7-89f4-c6b053eac242 at /var/lib/kubelet/pods/979c03a-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/kube-proxy-token-gqhfs
Nov 15 02:16:39 af867b kubelet[27751]: I1115 02:16:39.544543 27751 secret.go:217] Received secret kube-system/kube-proxy-token-gqhfs containing (3) pieces of data, 1904 total bytes
Nov 15 02:16:39 af867b kubelet[27751]: I1115 02:16:39.544605 27751 configmap.go:218] Received configMap kube-system/kube-proxy containing (1) pieces of data, 407 total bytes
Nov 15 02:16:39 af867b kubelet[27751]: I1115 02:16:39.544712 27751 atomic_writer.go:142] pod kube-system/kube-proxy-nnsjf volume kube-proxy-token-gqhfs: no update required for target directory /vr/lib/kubelet/pods/9729c03a-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/kube-proxy-token-gqhfs
Nov 15 02:16:39 af867b kubelet[27751]: I1115 02:16:39.544943 27751 atomic_writer.go:142] pod kube-system/kube-proxy-nnsjf volume kube-proxy: no update required for target directory /var/lib/kubelt/pods/9729c03a-c9a8-11e7-89f4-c6b053eac242/volumes/kubernetes.io~configmap/kube-proxy
Nov 15 02:16:39 af867b kubelet[27751]: I1115 02:16:39.825440 27751 volume_manager.go:366] All volumes are attached and mounted for pod "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053ac242)"
Nov 15 02:16:39 af867b kubelet[27751]: I1115 02:16:39.825571 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:false CreateSandbox:false SandboxID:91510886f3beb621e5d04309d502352ff8392e46405f21633802da4f7047069 Attempt:0 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "kube-proxy-nnsjf_kube-system(9729c03a-c9a8-11e7-89f4-c6b053eac242)"
Nov 15 02:16:40 af867b kubelet[27751]: I1115 02:16:40.523926 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:16:41 af867b kubelet[27751]: I1115 02:16:41.221358 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=true reason: mesage:
Nov 15 02:16:41 af867b kubelet[27751]: I1115 02:16:41.523502 27751 kubelet.go:1890] SyncLoop (SYNC): 1 pods; kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)
Nov 15 02:16:41 af867b kubelet[27751]: I1115 02:16:41.523664 27751 kubelet_pods.go:1284] Generating status for "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)"
Nov 15 02:16:41 af867b kubelet[27751]: I1115 02:16:41.524038 27751 status_manager.go:325] Ignoring same status for pod "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)", statu: {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:55 +0000 UTC Reason: Message:} {Type:Ready Status:True LasProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 01:58:58 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTme:2017-11-15 01:58:55 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP:10.196.65.210 StartTime:2017-11-15 01:58:55 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Nae:kube-scheduler State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 01:58:57 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:trueRestartCount:0 Image:gcr.io/google_containers/kube-scheduler-amd64:v1.8.3 ImageID:docker-pullable://gcr.io/google_containers/kube-scheduler-amd64@sha256:c47b2438bbab28d58e8cbf64b37b7f66d26b000f5c3a1626ee829a4be8fb91e ContainerID:docker://413566b22305750f9a9aa46fbe256c11e75293e80f6e0d4afb7ec9e6afcdee05}] QOSClass:Burstable}
Nov 15 02:16:41 af867b kubelet[27751]: I1115 02:16:41.524406 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217fd5c14373)"
Nov 15 02:16:41 af867b kubelet[27751]: I1115 02:16:41.824722 27751 volume_manager.go:366] All volumes are attached and mounted for pod "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfdc14373)"
Nov 15 02:16:41 af867b kubelet[27751]: I1115 02:16:41.824897 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:false CreateSandbox:false SandboxID:fe18d337ce42bd3a4d2aacb1349be02824681d8073fbe5d6377946e815fa810 Attempt:0 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373)"
Nov 15 02:16:42 af867b kubelet[27751]: I1115 02:16:42.523478 27751 kubelet.go:1890] SyncLoop (SYNC): 1 pods; kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)
Nov 15 02:16:42 af867b kubelet[27751]: I1115 02:16:42.523569 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:16:42 af867b kubelet[27751]: I1115 02:16:42.526100 27751 kubelet_pods.go:1284] Generating status for "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9)"
Nov 15 02:16:42 af867b kubelet[27751]: I1115 02:16:42.526503 27751 status_manager.go:325] Ignoring same status for pod "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9", status: {Phase:Running Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:12:51 +0000 UTC Reason: Message:} {Type:Ready StatusTrue LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:12:52 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTrnsitionTime:2017-11-15 02:12:51 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP:10.196.65.210 StartTime:2017-11-15 02:12:51 +0000 UTC InitContainerStatuses:[] ContainerStatses:[{Name:kube-controller-manager State:{Waiting:nil Running:&ContainerStateRunning{StartedAt:2017-11-15 02:12:52 +0000 UTC,} Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminatd:nil} Ready:true RestartCount:0 Image:gcr.io/google_containers/kube-controller-manager-amd64:v1.8.3 ImageID:docker-pullable://gcr.io/google_containers/kube-controller-manager-amd64@sha256:b6b633e3107761d38fceb200f01bf552c51f65e3524b0aafc1a7710afff07be ContainerID:docker://5b8a6b9d2792044cc30bacf05707077a8f7d5d3b7c1ff35931c96f933cf41f6e}] QOSClass:Burstable}
Nov 15 02:16:42 af867b kubelet[27751]: I1115 02:16:42.526886 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "kube-controller-manager-af867b_kube-system(f49ee4da5c66a63a0b4bcea4f69baf9)"
Nov 15 02:16:42 af867b kubelet[27751]: I1115 02:16:42.827260 27751 volume_manager.go:366] All volumes are attached and mounted for pod "kube-controller-manager-af867b_kube-system(f49ee4da5c66af630b4bcea4f69baf9)"
Nov 15 02:16:42 af867b kubelet[27751]: I1115 02:16:42.827446 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:false CreateSandbox:false SandboxID:a05e8c51434768693d26caf18b1a9774240b6197e992c2a2edcd1a9cb3b597f Attempt:0 NextInitContainerToStart:nil ContainersToStart:[] ContainersToKill:map[]} for pod "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69ba9)"
Nov 15 02:16:42 af867b kubelet[27751]: I1115 02:16:42.959972 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 2379, Path: /health
Nov 15 02:16:42 af867b kubelet[27751]: I1115 02:16:42.960021 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:16:43 af867b kubelet[27751]: I1115 02:16:43.197546 27751 prober.go:160] HTTP-Probe Host: http://10.32.0.2, Port: 10054, Path: /healthcheck/kubedns
Nov 15 02:16:43 af867b kubelet[27751]: I1115 02:16:43.197578 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:16:43 af867b kubelet[27751]: I1115 02:16:43.199539 27751 http.go:96] Probe succeeded for http://10.32.0.2:10054/healthcheck/kubedns, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Type:application/json] Date:[Wed, 15 Nov 2017 02:16:43 GMT] Content-Length:[51]] 0xc421438c80 51 [] true false map[] 0xc420ee3100 <nil>}
Nov 15 02:16:43 af867b kubelet[27751]: I1115 02:16:43.199632 27751 prober.go:113] Liveness probe for "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242):kubedns" succeede
Nov 15 02:16:43 af867b kubelet[27751]: I1115 02:16:43.461912 27751 http.go:96] Probe succeeded for http://127.0.0.1:2379/health, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:1:43 GMT] Content-Length:[18] Content-Type:[text/plain; charset=utf-8]] 0xc421439ca0 18 [] true false map[] 0xc420d4b600 <nil>}
Nov 15 02:16:43 af867b kubelet[27751]: I1115 02:16:43.461996 27751 prober.go:113] Liveness probe for "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250):etcd" succeeded
Nov 15 02:16:43 af867b kubelet[27751]: I1115 02:16:43.617390 27751 prober.go:160] HTTP-Probe Host: https://127.0.0.1, Port: 6443, Path: /healthz
Nov 15 02:16:43 af867b kubelet[27751]: I1115 02:16:43.617446 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:16:43 af867b kubelet[27751]: I1115 02:16:43.625673 27751 http.go:96] Probe succeeded for https://127.0.0.1:6443/healthz, Response: {200 OK 200 HTTP/2.0 2 0 map[Content-Type:[text/plain;charset=utf-8] Content-Length:[2] Date:[Wed, 15 Nov 2017 02:16:43 GMT]] 0xc42138a3a0 2 [] false false map[] 0xc420ee3400 0xc4218f0b00}
Nov 15 02:16:43 af867b kubelet[27751]: I1115 02:16:43.625737 27751 prober.go:113] Liveness probe for "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d):kube-apiserver" succeeded
Nov 15 02:16:43 af867b kubelet[27751]: I1115 02:16:43.834536 27751 prober.go:160] HTTP-Probe Host: http://10.32.0.2, Port: 10054, Path: /healthcheck/dnsmasq
Nov 15 02:16:43 af867b kubelet[27751]: I1115 02:16:43.834576 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:16:43 af867b kubelet[27751]: I1115 02:16:43.836154 27751 http.go:96] Probe succeeded for http://10.32.0.2:10054/healthcheck/dnsmasq, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Lengt:[51] Content-Type:[application/json] Date:[Wed, 15 Nov 2017 02:16:43 GMT]] 0xc42138a4e0 51 [] true false map[] 0xc420d4bd00 <nil>}
Nov 15 02:16:43 af867b kubelet[27751]: I1115 02:16:43.836204 27751 prober.go:113] Liveness probe for "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242):dnsmasq" succeede
Nov 15 02:16:44 af867b kubelet[27751]: I1115 02:16:44.504751 27751 prober.go:160] HTTP-Probe Host: http://10.32.0.2, Port: 10054, Path: /metrics
Nov 15 02:16:44 af867b kubelet[27751]: I1115 02:16:44.504800 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:16:44 af867b kubelet[27751]: I1115 02:16:44.516940 27751 http.go:96] Probe succeeded for http://10.32.0.2:10054/metrics, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Type:[text/plain;version=0.0.4] Date:[Wed, 15 Nov 2017 02:16:44 GMT]] 0xc42138bda0 -1 [] true true map[] 0xc420afc000 <nil>}
Nov 15 02:16:44 af867b kubelet[27751]: I1115 02:16:44.516987 27751 prober.go:113] Liveness probe for "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242):sidecar" succeede
Nov 15 02:16:44 af867b kubelet[27751]: I1115 02:16:44.523546 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:16:45 af867b kubelet[27751]: I1115 02:16:45.362705 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 6784, Path: /status
Nov 15 02:16:45 af867b kubelet[27751]: I1115 02:16:45.362737 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:16:45 af867b kubelet[27751]: I1115 02:16:45.364744 27751 http.go:96] Probe succeeded for http://127.0.0.1:6784/status, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Type:[text/plain; carset=utf-8] Date:[Wed, 15 Nov 2017 02:16:45 GMT] Content-Length:[445]] 0xc42014e320 445 [] true false map[] 0xc420afd300 <nil>}
Nov 15 02:16:45 af867b kubelet[27751]: I1115 02:16:45.364783 27751 prober.go:113] Liveness probe for "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242):weave" succeeded
Nov 15 02:16:46 af867b kubelet[27751]: I1115 02:16:46.223433 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=true reason: mesage:
Nov 15 02:16:46 af867b kubelet[27751]: I1115 02:16:46.481737 27751 eviction_manager.go:221] eviction manager: synchronize housekeeping
Nov 15 02:16:46 af867b kubelet[27751]: I1115 02:16:46.523809 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:16:46 af867b kubelet[27751]: I1115 02:16:46.547679 27751 helpers.go:871] eviction manager: observations: signal=memory.available, available: 5609296Ki, capacity: 7393360Ki, time: 2017-1-15 02:16:41.962872214 +0000 UTC
Nov 15 02:16:46 af867b kubelet[27751]: I1115 02:16:46.547712 27751 helpers.go:871] eviction manager: observations: signal=nodefs.available, available: 7399696Ki, capacity: 10198Mi, time: 2017-11-5 02:16:41.962872214 +0000 UTC
Nov 15 02:16:46 af867b kubelet[27751]: I1115 02:16:46.547722 27751 helpers.go:871] eviction manager: observations: signal=nodefs.inodesFree, available: 10384476, capacity: 10208Ki, time: 2017-11-5 02:16:41.962872214 +0000 UTC
Nov 15 02:16:46 af867b kubelet[27751]: I1115 02:16:46.547730 27751 helpers.go:871] eviction manager: observations: signal=imagefs.available, available: 39632Mi, capacity: 45Gi, time: 2017-11-15 0:16:41.962872214 +0000 UTC
Nov 15 02:16:46 af867b kubelet[27751]: I1115 02:16:46.547738 27751 helpers.go:873] eviction manager: observations: signal=allocatableMemory.available, available: 6884076Ki, capacity: 7393360Ki
Nov 15 02:16:46 af867b kubelet[27751]: I1115 02:16:46.547746 27751 helpers.go:873] eviction manager: observations: signal=allocatableNodeFs.available, available: 9624040228, capacity: 10198Mi
Nov 15 02:16:46 af867b kubelet[27751]: I1115 02:16:46.547763 27751 eviction_manager.go:325] eviction manager: no resources are starved
Nov 15 02:16:47 af867b kubelet[27751]: I1115 02:16:47.600763 27751 prober.go:160] HTTP-Probe Host: http://10.32.0.2, Port: 8081, Path: /readiness
Nov 15 02:16:47 af867b kubelet[27751]: I1115 02:16:47.600802 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:16:47 af867b kubelet[27751]: I1115 02:16:47.624474 27751 http.go:96] Probe succeeded for http://10.32.0.2:8081/readiness, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 0:16:47 GMT] Content-Length:[3] Content-Type:[text/plain; charset=utf-8]] 0xc421c3afe0 3 [] true false map[] 0xc420a30900 <nil>}
Nov 15 02:16:47 af867b kubelet[27751]: I1115 02:16:47.624533 27751 prober.go:113] Readiness probe for "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242):kubedns" succeedd
Nov 15 02:16:47 af867b kubelet[27751]: I1115 02:16:47.681909 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10251, Path: /healthz
Nov 15 02:16:47 af867b kubelet[27751]: I1115 02:16:47.681955 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:16:47 af867b kubelet[27751]: I1115 02:16:47.685904 27751 http.go:96] Probe succeeded for http://127.0.0.1:10251/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 0216:47 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc421c52e00 2 [] true false map[] 0xc420a30f00 <nil>}
Nov 15 02:16:47 af867b kubelet[27751]: I1115 02:16:47.685954 27751 prober.go:113] Liveness probe for "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373):kube-scheduler" succeeded
Nov 15 02:16:48 af867b kubelet[27751]: I1115 02:16:48.525895 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:16:49 af867b kubelet[27751]: I1115 02:16:49.216783 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10252, Path: /healthz
Nov 15 02:16:49 af867b kubelet[27751]: I1115 02:16:49.216815 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:16:49 af867b kubelet[27751]: I1115 02:16:49.219907 27751 http.go:96] Probe succeeded for http://127.0.0.1:10252/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Type:[text/plain;charset=utf-8] Date:[Wed, 15 Nov 2017 02:16:49 GMT] Content-Length:[2]] 0xc4221a18c0 2 [] true false map[] 0xc420b08f00 <nil>}
Nov 15 02:16:49 af867b kubelet[27751]: I1115 02:16:49.219947 27751 prober.go:113] Liveness probe for "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9):kube-controller-anager" succeeded
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.526372 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.553383 27751 qos_container_manager_linux.go:320] [ContainerManager]: Updated QoS cgroup configuration
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.750906 27751 kubelet.go:1222] Container garbage collection succeeded
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.831942 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-29c1400e918b28b992788b4c8b07c46b1767efcd8b96000e44bc78da639994b.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.831987 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-29c1400e918b28b9982788b4c807c46b1767efcd8b96000e44bc78da639994b.mount", but ignoring.
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832001 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-29c1400e918b28b9982788b4c8b07c46b1767efcd8b9600e44bc78da639994b.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832011 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-faffc5ce0cf0da39c8377dff0a611eb436ee74dfce11692aecd1e723311ef80.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832020 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-faffc5ce0cf0da39c68377dff0611eb436ee74dfce11692aecd1e723311ef80.mount", but ignoring.
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832029 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-faffc5ce0cf0da39c68377dff0a611eb436ee74dfce1162aecd1e723311ef80.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832120 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/sys-kernel-debug.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832137 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/sys-kernel-debug.mount", but ignoring.
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832150 27751 manager.go:901] ignoring container "/system.slice/sys-kernel-debug.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832160 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-ffde707e35c082af581df3bb46e12f0eaf2bf20d3f28b499b6739ec31319af8.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832175 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-ffde707e35c082af5481df3bb4e12f0eaf2bf20d3f28b499b6739ec31319af8.mount", but ignoring.
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832190 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-ffde707e35c082af5481df3bb46e12f0eaf2bf20d3f28b99b6739ec31319af8.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832202 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832212 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper.mount", but ignoring.
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832225 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832272 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/boot.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832283 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/boot.mount", but ignoring.
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832291 27751 manager.go:901] ignoring container "/system.slice/boot.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832297 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/proc-xen.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832304 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/proc-xen.mount", but ignoring.
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832311 27751 manager.go:901] ignoring container "/system.slice/proc-xen.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832317 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/run-user-1000.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832323 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/run-user-1000.mount", but ignoring.
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832330 27751 manager.go:901] ignoring container "/system.slice/run-user-1000.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832337 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/sys-kernel-config.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832343 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/sys-kernel-config.mount", but ignoring.
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832351 27751 manager.go:901] ignoring container "/system.slice/sys-kernel-config.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832357 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-7d2ebdadb8c541a489058dd46b491ef9e14209821b47e7d86b539de402fbef5.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832365 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-7d2ebdadb8c541a4839058dd46491ef9e14209821b47e7d86b539de402fbef5.mount", but ignoring.
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832375 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-7d2ebdadb8c541a4839058dd46b491ef9e14209821b47ed86b539de402fbef5.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832383 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-9d97c593c8a0b1ef242048619ba4466fad52c408e686c9285a48450f3a6669f.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832391 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-9d97c593c8a0b1ef2442048619a4466fad52c408e686c9285a48450f3a6669f.mount", but ignoring.
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832401 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-9d97c593c8a0b1ef2442048619ba4466fad52c408e686c285a48450f3a6669f.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832408 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-kubelet-pods-9729c03a\\x2dc9a8\\x2d11e7\\2d89f4\\x2dc6b053eac242-volumes-kubernetes.io\\x7esecret-kube\\x2dproxy\\x2dtoken\\x2dgqhfs.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832430 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-kubelet-pods-9729c03a\\x2dc9a8\\x2d11e7\\x2d89f4\\2dc6b053eac242-volumes-kubernetes.io\\x7esecret-kube\\x2dproxy\\x2dtoken\\x2dgqhfs.mount", but ignoring.
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832444 27751 manager.go:901] ignoring container "/system.slice/var-lib-kubelet-pods-9729c03a\\x2dc9a8\\x2d11e7\\x2d89f4\\x2dc6b053eac242-volues-kubernetes.io\\x7esecret-kube\\x2dproxy\\x2dtoken\\x2dgqhfs.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832454 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/dev-mqueue.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832460 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/dev-mqueue.mount", but ignoring.
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832468 27751 manager.go:901] ignoring container "/system.slice/dev-mqueue.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832473 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-cd27b64716d8032aaea595ec121a222dc749f2f903149a780eb711106fb0c53.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832482 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-cd27b64716d8032aa3ea595ec11a222dc749f2f903149a780eb711106fb0c53.mount", but ignoring.
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832491 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-cd27b64716d8032aa3ea595ec121a222dc749f2f903149780eb711106fb0c53.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832500 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/u01-applicationSpace.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832506 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/u01-applicationSpace.mount", but ignoring.
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832514 27751 manager.go:901] ignoring container "/system.slice/u01-applicationSpace.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832519 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-fe18d337ce42bd3a4d2aacb349be028274681d8073fbe5d6377946e815fa810-shm.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832527 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-fe18d337ce42bd3a4d2aacb1349be02874681d8073fbe5d6377946e815fa810-shm.mount", but ignoring.
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832538 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-containers-fe18d337ce42bd3a4d2aacb1349be028274681d8073fbe5d637746e815fa810-shm.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832545 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-984a35c7cf6d04c0a5fcd260872a7b6556aa9cd8b38bc1a1cf635972c1d6511.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832554 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-984a35c7cf6d04c0af5fcd26082a7b6556aa9cd8b38bc1a1cf635972c1d6511.mount", but ignoring.
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832564 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-984a35c7cf6d04c0af5fcd260872a7b6556aa9cd8b38bca1cf635972c1d6511.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832572 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/-.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832578 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/-.mount", but ignoring.
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832585 27751 manager.go:901] ignoring container "/system.slice/-.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832591 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/dev-hugepages.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832598 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/dev-hugepages.mount", but ignoring.
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832606 27751 manager.go:901] ignoring container "/system.slice/dev-hugepages.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832612 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-91510886f3beb621e5d0430d502352ff78392e46405f21633802da4f7047069-shm.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832620 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-91510886f3beb621e5d04309d502352f78392e46405f21633802da4f7047069-shm.mount", but ignoring.
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832630 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-containers-91510886f3beb621e5d04309d502352ff78392e46405f2163380da4f7047069-shm.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832638 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-19b06ca14052c17f92bf03d316fd7642fe4fee6bf193755575355be3e87a13d-shm.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832646 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-19b06ca14052c17f92bf03d5316fd764fe4fee6bf193755575355be3e87a13d-shm.mount", but ignoring.
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832655 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-containers-19b06ca14052c17f92bf03d5316fd7642fe4fee6bf19375557535be3e87a13d-shm.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832663 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-23d8892c53722cfb43aa6c5b126ad2674a7b8041b4873d30b98faefad6f5b44.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832671 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-23d8892c53722cfb4d3aa6c5b16ad2674a7b8041b4873d30b98faefad6f5b44.mount", but ignoring.
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832680 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-23d8892c53722cfb4d3aa6c5b126ad2674a7b8041b487330b98faefad6f5b44.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832689 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-82408ec054eef4c0be3519663bc7e5cbaff7a4fea1d8b7490fa3cc5a74bac60.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832697 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-82408ec054eef4c0bce3519663c7e5cbaff7a4fea1d8b7490fa3cc5a74bac60.mount", but ignoring.
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832706 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-82408ec054eef4c0bce3519663bc7e5cbaff7a4fea1d8b490fa3cc5a74bac60.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832752 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-kubelet-pods-b77b0858\\x2dc9a8\\x2d11e7\\2d89f4\\x2dc6b053eac242-volumes-kubernetes.io\\x7esecret-weave\\x2dnet\\x2dtoken\\x2drn6j7.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832762 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-kubelet-pods-b77b0858\\x2dc9a8\\x2d11e7\\x2d89f4\\2dc6b053eac242-volumes-kubernetes.io\\x7esecret-weave\\x2dnet\\x2dtoken\\x2drn6j7.mount", but ignoring.
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832774 27751 manager.go:901] ignoring container "/system.slice/var-lib-kubelet-pods-b77b0858\\x2dc9a8\\x2d11e7\\x2d89f4\\x2dc6b053eac242-volues-kubernetes.io\\x7esecret-weave\\x2dnet\\x2dtoken\\x2drn6j7.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832784 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/run-docker-netns-default.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832791 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/run-docker-netns-default.mount", but ignoring.
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832800 27751 manager.go:901] ignoring container "/system.slice/run-docker-netns-default.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832806 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-be14b0552100f929052c8e9e1b7cf587c4218e321d811314e58f2cf312f1c0f-shm.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832814 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-be14b0552100f929052c8e94e1b7cf58c4218e321d811314e58f2cf312f1c0f-shm.mount", but ignoring.
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832823 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-containers-be14b0552100f929052c8e94e1b7cf587c4218e321d811314e582cf312f1c0f-shm.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832831 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-d5076ad0e9fb270d1b8c4ffcdadbf32d6b1e30dc42ae24dfbc5cb01bb5aa934-shm.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832839 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-d5076ad0e9fb270d1b8c4ff7cdadbf326b1e30dc42ae24dfbc5cb01bb5aa934-shm.mount", but ignoring.
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832848 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-containers-d5076ad0e9fb270d1b8c4ff7cdadbf32d6b1e30dc42ae24dfbc5b01bb5aa934-shm.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832856 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-9ba72245089e9b36e01a816a7446b004869c57dbf880db7daeb3edac0f81e51.mount"
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832864 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-9ba72245089e9b36e901a816a746b004869c57dbf880db7daeb3edac0f81e51.mount", but ignoring.
Nov 15 02:16:50 af867b kubelet[27751]: I1115 02:16:50.832874 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-9ba72245089e9b36e901a816a7446b004869c57dbf880d7daeb3edac0f81e51.mount"
Nov 15 02:16:51 af867b kubelet[27751]: I1115 02:16:51.227958 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=true reason: mesage:
Nov 15 02:16:52 af867b kubelet[27751]: I1115 02:16:52.523852 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:16:52 af867b kubelet[27751]: I1115 02:16:52.960894 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 2379, Path: /health
Nov 15 02:16:52 af867b kubelet[27751]: I1115 02:16:52.960939 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:16:53 af867b kubelet[27751]: I1115 02:16:53.197839 27751 prober.go:160] HTTP-Probe Host: http://10.32.0.2, Port: 10054, Path: /healthcheck/kubedns
Nov 15 02:16:53 af867b kubelet[27751]: I1115 02:16:53.197867 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:16:53 af867b kubelet[27751]: I1115 02:16:53.199321 27751 http.go:96] Probe succeeded for http://10.32.0.2:10054/healthcheck/kubedns, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Type:application/json] Date:[Wed, 15 Nov 2017 02:16:53 GMT] Content-Length:[51]] 0xc421144720 51 [] true false map[] 0xc4200dd600 <nil>}
Nov 15 02:16:53 af867b kubelet[27751]: I1115 02:16:53.199378 27751 prober.go:113] Liveness probe for "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242):kubedns" succeede
Nov 15 02:16:53 af867b kubelet[27751]: I1115 02:16:53.467882 27751 http.go:96] Probe succeeded for http://127.0.0.1:2379/health, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:1:53 GMT] Content-Length:[18] Content-Type:[text/plain; charset=utf-8]] 0xc42121c520 18 [] true false map[] 0xc4200dd000 <nil>}
Nov 15 02:16:53 af867b kubelet[27751]: I1115 02:16:53.467949 27751 prober.go:113] Liveness probe for "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250):etcd" succeeded
Nov 15 02:16:53 af867b kubelet[27751]: I1115 02:16:53.618829 27751 prober.go:160] HTTP-Probe Host: https://127.0.0.1, Port: 6443, Path: /healthz
Nov 15 02:16:53 af867b kubelet[27751]: I1115 02:16:53.618869 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:16:53 af867b kubelet[27751]: I1115 02:16:53.626801 27751 http.go:96] Probe succeeded for https://127.0.0.1:6443/healthz, Response: {200 OK 200 HTTP/2.0 2 0 map[Date:[Wed, 15 Nov 2017 0216:53 GMT] Content-Type:[text/plain; charset=utf-8] Content-Length:[2]] 0xc421263880 2 [] false false map[] 0xc4200ddc00 0xc421f50580}
Nov 15 02:16:53 af867b kubelet[27751]: I1115 02:16:53.626841 27751 prober.go:113] Liveness probe for "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d):kube-apiserver" succeeded
Nov 15 02:16:53 af867b kubelet[27751]: I1115 02:16:53.836175 27751 prober.go:160] HTTP-Probe Host: http://10.32.0.2, Port: 10054, Path: /healthcheck/dnsmasq
Nov 15 02:16:53 af867b kubelet[27751]: I1115 02:16:53.836238 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:16:53 af867b kubelet[27751]: I1115 02:16:53.837693 27751 http.go:96] Probe succeeded for http://10.32.0.2:10054/healthcheck/dnsmasq, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Type:application/json] Date:[Wed, 15 Nov 2017 02:16:53 GMT] Content-Length:[51]] 0xc4212e11c0 51 [] true false map[] 0xc420430c00 <nil>}
Nov 15 02:16:53 af867b kubelet[27751]: I1115 02:16:53.837759 27751 prober.go:113] Liveness probe for "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242):dnsmasq" succeede
Nov 15 02:16:54 af867b kubelet[27751]: I1115 02:16:54.504578 27751 prober.go:160] HTTP-Probe Host: http://10.32.0.2, Port: 10054, Path: /metrics
Nov 15 02:16:54 af867b kubelet[27751]: I1115 02:16:54.504617 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:16:54 af867b kubelet[27751]: I1115 02:16:54.511205 27751 http.go:96] Probe succeeded for http://10.32.0.2:10054/metrics, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Type:[text/plain;version=0.0.4] Date:[Wed, 15 Nov 2017 02:16:54 GMT]] 0xc420f3e180 -1 [] true true map[] 0xc420431a00 <nil>}
Nov 15 02:16:54 af867b kubelet[27751]: I1115 02:16:54.511247 27751 prober.go:113] Liveness probe for "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242):sidecar" succeede
Nov 15 02:16:54 af867b kubelet[27751]: I1115 02:16:54.526795 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:16:55 af867b kubelet[27751]: I1115 02:16:55.363904 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 6784, Path: /status
Nov 15 02:16:55 af867b kubelet[27751]: I1115 02:16:55.363945 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:16:55 af867b kubelet[27751]: I1115 02:16:55.368050 27751 http.go:96] Probe succeeded for http://127.0.0.1:6784/status, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:1:55 GMT] Content-Length:[445] Content-Type:[text/plain; charset=utf-8]] 0xc4218b97e0 445 [] true false map[] 0xc421178700 <nil>}
Nov 15 02:16:55 af867b kubelet[27751]: I1115 02:16:55.368111 27751 prober.go:113] Liveness probe for "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242):weave" succeeded
Nov 15 02:16:55 af867b kubelet[27751]: I1115 02:16:55.879324 27751 config.go:282] Setting pods for source api
Nov 15 02:16:55 af867b kubelet[27751]: I1115 02:16:55.880480 27751 config.go:404] Receiving a new pod "snginx-10_default(0d5631da-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:16:55 af867b kubelet[27751]: I1115 02:16:55.881055 27751 kubelet.go:1837] SyncLoop (ADD, "api"): "snginx-10_default(0d5631da-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:16:55 af867b kubelet[27751]: I1115 02:16:55.881195 27751 kubelet_pods.go:1284] Generating status for "snginx-10_default(0d5631da-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:16:55 af867b kubelet[27751]: I1115 02:16:55.882469 27751 qos_container_manager_linux.go:320] [ContainerManager]: Updated QoS cgroup configuration
Nov 15 02:16:55 af867b kubelet[27751]: I1115 02:16:55.907077 27751 factory.go:116] Factory "docker" was unable to handle container "/kubepods/besteffort/pod0d5631da-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:16:55 af867b kubelet[27751]: I1115 02:16:55.907103 27751 factory.go:105] Error trying to work out if we can handle /kubepods/besteffort/pod0d5631da-c9ab-11e7-89f4-c6b053eac242: /kubepod/besteffort/pod0d5631da-c9ab-11e7-89f4-c6b053eac242 not handled by systemd handler
Nov 15 02:16:55 af867b kubelet[27751]: I1115 02:16:55.907141 27751 factory.go:116] Factory "systemd" was unable to handle container "/kubepods/besteffort/pod0d5631da-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:16:55 af867b kubelet[27751]: I1115 02:16:55.907149 27751 factory.go:112] Using factory "raw" for container "/kubepods/besteffort/pod0d5631da-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:16:55 af867b kubelet[27751]: I1115 02:16:55.908143 27751 manager.go:932] Added container: "/kubepods/besteffort/pod0d5631da-c9ab-11e7-89f4-c6b053eac242" (aliases: [], namespace: "")
Nov 15 02:16:55 af867b kubelet[27751]: I1115 02:16:55.908331 27751 handler.go:325] Added event &{/kubepods/besteffort/pod0d5631da-c9ab-11e7-89f4-c6b053eac242 2017-11-15 02:16:55.903194711 +0000 UC containerCreation {<nil>}}
Nov 15 02:16:55 af867b kubelet[27751]: I1115 02:16:55.908368 27751 container.go:409] Start housekeeping for container "/kubepods/besteffort/pod0d5631da-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:16:55 af867b kubelet[27751]: I1115 02:16:55.920067 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "snginx-10_default(0d5631da-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:16:55 af867b kubelet[27751]: I1115 02:16:55.968016 27751 config.go:282] Setting pods for source api
Nov 15 02:16:55 af867b kubelet[27751]: I1115 02:16:55.969524 27751 kubelet.go:1850] SyncLoop (RECONCILE, "api"): "snginx-10_default(0d5631da-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:16:55 af867b kubelet[27751]: I1115 02:16:55.973949 27751 status_manager.go:451] Status for pod "snginx-10_default(0d5631da-c9ab-11e7-89f4-c6b053eac242)" updated successfully: (1, {PhasePending Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:16:55 +0000 UTC Reason: Message:} {Type:Ready Status:False LastProbeTie:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:16:55 +0000 UTC Reason:ContainersNotReady Message:containers with unready status: [nginx]} {Type:PodScheduled Status:True LastProbeTme:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:16:55 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP: StartTime:2017-11-15 02:16:55 +0000 UTC InitContainrStatuses:[] ContainerStatuses:[{Name:nginx State:{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,} Running:nil Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Termiated:nil} Ready:false RestartCount:0 Image:nginx ImageID: ContainerID:}] QOSClass:BestEffort})
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.076998 27751 reconciler.go:212] operationExecutor.VerifyControllerAttachedVolume started for volume "default-token-qjbsf" (UniqueName: "kuberetes.io/secret/0d5631da-c9ab-11e7-89f4-c6b053eac242-default-token-qjbsf") pod "snginx-10" (UID: "0d5631da-c9ab-11e7-89f4-c6b053eac242")
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.178310 27751 reconciler.go:257] operationExecutor.MountVolume started for volume "default-token-qjbsf" (UniqueName: "kubernetes.io/secret/0d531da-c9ab-11e7-89f4-c6b053eac242-default-token-qjbsf") pod "snginx-10" (UID: "0d5631da-c9ab-11e7-89f4-c6b053eac242")
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.178387 27751 secret.go:186] Setting up volume default-token-qjbsf for pod 0d5631da-c9ab-11e7-89f4-c6b053eac242 at /var/lib/kubelet/pods/0d563da-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.178601 27751 empty_dir.go:264] pod 0d5631da-c9ab-11e7-89f4-c6b053eac242: mounting tmpfs for volume wrapped_default-token-qjbsf
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.178623 27751 mount_linux.go:135] Mounting cmd (systemd-run) with arguments ([--description=Kubernetes transient mount for /var/lib/kubelet/pos/0d5631da-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf --scope -- mount -t tmpfs tmpfs /var/lib/kubelet/pods/0d5631da-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetesio~secret/default-token-qjbsf])
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.191695 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/run-1306.scope"
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.191764 27751 factory.go:105] Error trying to work out if we can handle /system.slice/run-1306.scope: /system.slice/run-1306.scope not handledby systemd handler
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.191773 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/run-1306.scope"
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.191781 27751 factory.go:112] Using factory "raw" for container "/system.slice/run-1306.scope"
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.191934 27751 manager.go:932] Added container: "/system.slice/run-1306.scope" (aliases: [], namespace: "")
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.192037 27751 handler.go:325] Added event &{/system.slice/run-1306.scope 2017-11-15 02:16:56.185194711 +0000 UTC containerCreation {<nil>}}
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.192070 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-8dc12d8756ef2ac44557ac1346841f65a298d1c068a94bca68af19724fac1b0.mount"
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.192083 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-8dc12d8756ef2ac445557ac134841f65a298d1c068a94bca68af19724fac1b0.mount", but ignoring.
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.192094 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-8dc12d8756ef2ac445557ac1346841f65a298d1c068a94ca68af19724fac1b0.mount"
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.192104 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-4c898f767718d5bed6f7aa6de5dd63e3b65aa41665cb76b96b0e4e8bd126526.mount"
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.192112 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-4c898f767718d5bed36f7aa6dedd63e3b65aa41665cb76b96b0e4e8bd126526.mount", but ignoring.
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.192122 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-4c898f767718d5bed36f7aa6de5dd63e3b65aa41665cb7b96b0e4e8bd126526.mount"
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.192131 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-bad218fed21ce57f7150c24da0486c9ae20ab13e29dc9dfbd021de166c11a8b.mount"
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.192139 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-bad218fed21ce57f77150c24da486c9ae20ab13e29dc9dfbd021de166c11a8b.mount", but ignoring.
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.192149 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-bad218fed21ce57f77150c24da0486c9ae20ab13e29dc9fbd021de166c11a8b.mount"
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.192157 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-kubelet-pods-97270c63\\x2dc9a8\\x2d11e7\\2d89f4\\x2dc6b053eac242-volumes-kubernetes.io\\x7esecret-kube\\x2ddns\\x2dtoken\\x2d987zv.mount"
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.192167 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-kubelet-pods-97270c63\\x2dc9a8\\x2d11e7\\x2d89f4\\2dc6b053eac242-volumes-kubernetes.io\\x7esecret-kube\\x2ddns\\x2dtoken\\x2d987zv.mount", but ignoring.
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.192179 27751 manager.go:901] ignoring container "/system.slice/var-lib-kubelet-pods-97270c63\\x2dc9a8\\x2d11e7\\x2d89f4\\x2dc6b053eac242-volues-kubernetes.io\\x7esecret-kube\\x2ddns\\x2dtoken\\x2d987zv.mount"
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.192193 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-57f84a53f69dda718423e2a2b069d129afa4d226103816e6fa21ee5fab8b82b-shm.mount"
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.192202 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-57f84a53f69dda718423e2a72b069d12afa4d226103816e6fa21ee5fab8b82b-shm.mount", but ignoring.
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.192211 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-containers-57f84a53f69dda718423e2a72b069d129afa4d226103816e6fa2ee5fab8b82b-shm.mount"
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.192220 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-43c233c30408e57d9b41952d031e6274ca80b99953da44b0ab20a811a1643cd.mount"
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.192228 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-43c233c30408e57d98b41952d01e6274ca80b99953da44b0ab20a811a1643cd.mount", but ignoring.
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.192241 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-43c233c30408e57d98b41952d031e6274ca80b99953da4b0ab20a811a1643cd.mount"
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.192250 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-e4d771176f4a77939775443f08578068442b107ec0f99073f00532cfd4c7e31.mount"
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.192258 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-e4d771176f4a779391775443f0578068442b107ec0f99073f00532cfd4c7e31.mount", but ignoring.
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.192268 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-e4d771176f4a779391775443f08578068442b107ec0f9973f00532cfd4c7e31.mount"
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.192277 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/run-docker-netns-26165e268da2.mount"
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.192284 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/run-docker-netns-26165e268da2.mount", but ignoring.
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.192292 27751 manager.go:901] ignoring container "/system.slice/run-docker-netns-26165e268da2.mount"
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.192299 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-a05e8c51434768693d26caf8b1a97742f40b6197e992c2a2edcd1a9cb3b597f-shm.mount"
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.192307 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-a05e8c51434768693d26caf18b1a9774f40b6197e992c2a2edcd1a9cb3b597f-shm.mount", but ignoring.
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.192316 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-containers-a05e8c51434768693d26caf18b1a97742f40b6197e992c2a2edc1a9cb3b597f-shm.mount"
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.192325 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-de36c5b16bb7b3560a83dd6d52bbd656881ea6228a27d7578e51bcb2fa16f3f.mount"
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.192334 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-de36c5b16bb7b35606a83dd6d5bbd656881ea6228a27d7578e51bcb2fa16f3f.mount", but ignoring.
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.192343 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-de36c5b16bb7b35606a83dd6d52bbd656881ea6228a27d578e51bcb2fa16f3f.mount"
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.192366 27751 container.go:409] Start housekeeping for container "/system.slice/run-1306.scope"
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.196728 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-3b26b88e613ebc9ff1ea92e255fe70cbbd3e8e67980d4be55241a1198e378cc.mount"
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.196748 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-3b26b88e613ebc9ff51ea92e25fe70cbbd3e8e67980d4be55241a1198e378cc.mount", but ignoring.
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.196777 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-3b26b88e613ebc9ff51ea92e255fe70cbbd3e8e67980d4e55241a1198e378cc.mount"
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.206221 27751 manager.go:989] Destroyed container: "/system.slice/run-1306.scope" (aliases: [], namespace: "")
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.206241 27751 handler.go:325] Added event &{/system.slice/run-1306.scope 2017-11-15 02:16:56.206236648 +0000 UTC containerDeletion {<nil>}}
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.227835 27751 secret.go:217] Received secret default/default-token-qjbsf containing (3) pieces of data, 1878 total bytes
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.227918 27751 atomic_writer.go:145] pod default/snginx-10 volume default-token-qjbsf: write required for target directory /var/lib/kubelet/pod/0d5631da-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.228023 27751 atomic_writer.go:160] pod default/snginx-10 volume default-token-qjbsf: performed write of new data to ts data directory: /var/lb/kubelet/pods/0d5631da-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf/..119811_15_11_02_16_56.160831717
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.228115 27751 operation_generator.go:484] MountVolume.SetUp succeeded for volume "default-token-qjbsf" (UniqueName: "kubernetes.io/secret/0d561da-c9ab-11e7-89f4-c6b053eac242-default-token-qjbsf") pod "snginx-10" (UID: "0d5631da-c9ab-11e7-89f4-c6b053eac242")
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.228350 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"default", Name:"snginx-10", UID:"0d5631da-c9ab-11e7-89f4-c6b053ea242", APIVersion:"v1", ResourceVersion:"1308", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "default-token-qjbsf"
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.231847 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=true reason: mesage:
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.520428 27751 volume_manager.go:366] All volumes are attached and mounted for pod "snginx-10_default(0d5631da-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.520475 27751 kuberuntime_manager.go:370] No sandbox for pod "snginx-10_default(0d5631da-c9ab-11e7-89f4-c6b053eac242)" can be found. Need to sart a new one
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.520488 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:true CreateSandbox:true SandboxID: Attempt:0 NextInitContainerToStartnil ContainersToStart:[0] ContainersToKill:map[]} for pod "snginx-10_default(0d5631da-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.520550 27751 kuberuntime_manager.go:565] SyncPod received new pod "snginx-10_default(0d5631da-c9ab-11e7-89f4-c6b053eac242)", will create a sadbox for it
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.520559 27751 kuberuntime_manager.go:574] Stopping PodSandbox for "snginx-10_default(0d5631da-c9ab-11e7-89f4-c6b053eac242)", will start new on
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.520579 27751 kuberuntime_manager.go:626] Creating sandbox for pod "snginx-10_default(0d5631da-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.523431 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.526707 27751 expiration_cache.go:98] Entry version: {key:version obj:0xc4223dcb40} has expired
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.527153 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/besteffort/pod0d5631da-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.527169 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/besteffort/pod0d5631da-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.551794 27751 eviction_manager.go:221] eviction manager: synchronize housekeeping
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.603355 27751 helpers.go:871] eviction manager: observations: signal=nodefs.inodesFree, available: 10384476, capacity: 10208Ki, time: 2017-11-5 02:16:41.962872214 +0000 UTC
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.603390 27751 helpers.go:871] eviction manager: observations: signal=imagefs.available, available: 39632Mi, capacity: 45Gi, time: 2017-11-15 0:16:41.962872214 +0000 UTC
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.603399 27751 helpers.go:873] eviction manager: observations: signal=allocatableMemory.available, available: 6884476Ki, capacity: 7393360Ki
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.603406 27751 helpers.go:873] eviction manager: observations: signal=allocatableNodeFs.available, available: 9624040228, capacity: 10198Mi
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.603413 27751 helpers.go:871] eviction manager: observations: signal=memory.available, available: 5609296Ki, capacity: 7393360Ki, time: 2017-1-15 02:16:41.962872214 +0000 UTC
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.603421 27751 helpers.go:871] eviction manager: observations: signal=nodefs.available, available: 7399696Ki, capacity: 10198Mi, time: 2017-11-5 02:16:41.962872214 +0000 UTC
Nov 15 02:16:56 af867b kubelet[27751]: I1115 02:16:56.603440 27751 eviction_manager.go:325] eviction manager: no resources are starved
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.191300 27751 config.go:282] Setting pods for source api
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.192475 27751 config.go:404] Receiving a new pod "snginx-25_default(0e141d4f-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.193059 27751 kubelet.go:1837] SyncLoop (ADD, "api"): "snginx-25_default(0e141d4f-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.193220 27751 kubelet_pods.go:1284] Generating status for "snginx-25_default(0e141d4f-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.193519 27751 qos_container_manager_linux.go:320] [ContainerManager]: Updated QoS cgroup configuration
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.194751 27751 reconciler.go:212] operationExecutor.VerifyControllerAttachedVolume started for volume "default-token-qjbsf" (UniqueName: "kuberetes.io/secret/0e141d4f-c9ab-11e7-89f4-c6b053eac242-default-token-qjbsf") pod "snginx-25" (UID: "0e141d4f-c9ab-11e7-89f4-c6b053eac242")
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.213955 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "snginx-25_default(0e141d4f-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.296945 27751 reconciler.go:257] operationExecutor.MountVolume started for volume "default-token-qjbsf" (UniqueName: "kubernetes.io/secret/0e11d4f-c9ab-11e7-89f4-c6b053eac242-default-token-qjbsf") pod "snginx-25" (UID: "0e141d4f-c9ab-11e7-89f4-c6b053eac242")
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.297016 27751 secret.go:186] Setting up volume default-token-qjbsf for pod 0e141d4f-c9ab-11e7-89f4-c6b053eac242 at /var/lib/kubelet/pods/0e1414f-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.297197 27751 empty_dir.go:264] pod 0e141d4f-c9ab-11e7-89f4-c6b053eac242: mounting tmpfs for volume wrapped_default-token-qjbsf
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.297222 27751 mount_linux.go:135] Mounting cmd (systemd-run) with arguments ([--description=Kubernetes transient mount for /var/lib/kubelet/pos/0e141d4f-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf --scope -- mount -t tmpfs tmpfs /var/lib/kubelet/pods/0e141d4f-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetesio~secret/default-token-qjbsf])
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.465063 27751 secret.go:217] Received secret default/default-token-qjbsf containing (3) pieces of data, 1878 total bytes
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.465176 27751 atomic_writer.go:145] pod default/snginx-25 volume default-token-qjbsf: write required for target directory /var/lib/kubelet/pod/0e141d4f-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.465287 27751 atomic_writer.go:160] pod default/snginx-25 volume default-token-qjbsf: performed write of new data to ts data directory: /var/lb/kubelet/pods/0e141d4f-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf/..119811_15_11_02_16_57.864151808
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.465396 27751 operation_generator.go:484] MountVolume.SetUp succeeded for volume "default-token-qjbsf" (UniqueName: "kubernetes.io/secret/0e14d4f-c9ab-11e7-89f4-c6b053eac242-default-token-qjbsf") pod "snginx-25" (UID: "0e141d4f-c9ab-11e7-89f4-c6b053eac242")
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.465436 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"default", Name:"snginx-25", UID:"0e141d4f-c9ab-11e7-89f4-c6b053ea242", APIVersion:"v1", ResourceVersion:"1314", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "default-token-qjbsf"
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.514195 27751 volume_manager.go:366] All volumes are attached and mounted for pod "snginx-25_default(0e141d4f-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.514238 27751 kuberuntime_manager.go:370] No sandbox for pod "snginx-25_default(0e141d4f-c9ab-11e7-89f4-c6b053eac242)" can be found. Need to sart a new one
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.514252 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:true CreateSandbox:true SandboxID: Attempt:0 NextInitContainerToStartnil ContainersToStart:[0] ContainersToKill:map[]} for pod "snginx-25_default(0e141d4f-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.514327 27751 kuberuntime_manager.go:565] SyncPod received new pod "snginx-25_default(0e141d4f-c9ab-11e7-89f4-c6b053eac242)", will create a sadbox for it
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.514339 27751 kuberuntime_manager.go:574] Stopping PodSandbox for "snginx-25_default(0e141d4f-c9ab-11e7-89f4-c6b053eac242)", will start new on
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.514362 27751 kuberuntime_manager.go:626] Creating sandbox for pod "snginx-25_default(0e141d4f-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.516376 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/besteffort/pod0e141d4f-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.516392 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/besteffort/pod0e141d4f-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.600967 27751 prober.go:160] HTTP-Probe Host: http://10.32.0.2, Port: 8081, Path: /readiness
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.601007 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.606469 27751 http.go:96] Probe succeeded for http://10.32.0.2:8081/readiness, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 0:16:57 GMT] Content-Length:[3] Content-Type:[text/plain; charset=utf-8]] 0xc420cc3280 3 [] true false map[] 0xc420afd300 <nil>}
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.606518 27751 prober.go:113] Readiness probe for "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242):kubedns" succeedd
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.681846 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10251, Path: /healthz
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.681906 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.682872 27751 http.go:96] Probe succeeded for http://127.0.0.1:10251/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 0216:57 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc420d82000 2 [] true false map[] 0xc420afdf00 <nil>}
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.682926 27751 prober.go:113] Liveness probe for "kube-scheduler-af867b_kube-system(bc22704d9f4dc5d62a8217cfd5c14373):kube-scheduler" succeeded
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.710661 27751 docker_sandbox.go:691] Will attempt to re-write config file /var/lib/docker/containers/5fb7fcf7c72f470128f788803221316e45e7426df6260371e72e5a65e412678/resolv.conf with:
Nov 15 02:16:57 af867b kubelet[27751]: [nameserver 10.96.0.10 search default.svc.cluster.local svc.cluster.local cluster.local opcwlaas.oraclecloud.internal. options ndots:5]
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.710877 27751 plugins.go:392] Calling network plugin cni to set up pod "snginx-10_default"
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.716034 27751 cni.go:326] Got netns path /proc/1421/ns/net
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.716048 27751 cni.go:327] Using netns path default
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.716149 27751 cni.go:298] About to add CNI network cni-loopback (type=loopback)
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.728525 27751 factory.go:112] Using factory "docker" for container "/kubepods/besteffort/pod0d5631da-c9ab-11e7-89f4-c6b053eac242/5fb7fcf7c72f40128f788803221316e45e7426df36260371e72e5a65e412678"
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.732585 27751 manager.go:932] Added container: "/kubepods/besteffort/pod0d5631da-c9ab-11e7-89f4-c6b053eac242/5fb7fcf7c72f470128f788803221316e4e7426df36260371e72e5a65e412678" (aliases: [k8s_POD_snginx-10_default_0d5631da-c9ab-11e7-89f4-c6b053eac242_0 5fb7fcf7c72f470128f788803221316e45e7426df36260371e72e5a65e412678], namespace: "docker")
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.732751 27751 handler.go:325] Added event &{/kubepods/besteffort/pod0d5631da-c9ab-11e7-89f4-c6b053eac242/5fb7fcf7c72f470128f788803221316e45e746df36260371e72e5a65e412678 2017-11-15 02:16:57.189194711 +0000 UTC containerCreation {<nil>}}
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.732798 27751 factory.go:116] Factory "docker" was unable to handle container "/kubepods/besteffort/pod0e141d4f-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.732810 27751 factory.go:105] Error trying to work out if we can handle /kubepods/besteffort/pod0e141d4f-c9ab-11e7-89f4-c6b053eac242: /kubepod/besteffort/pod0e141d4f-c9ab-11e7-89f4-c6b053eac242 not handled by systemd handler
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.732816 27751 factory.go:116] Factory "systemd" was unable to handle container "/kubepods/besteffort/pod0e141d4f-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.732824 27751 factory.go:112] Using factory "raw" for container "/kubepods/besteffort/pod0e141d4f-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.732994 27751 manager.go:932] Added container: "/kubepods/besteffort/pod0e141d4f-c9ab-11e7-89f4-c6b053eac242" (aliases: [], namespace: "")
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.733110 27751 handler.go:325] Added event &{/kubepods/besteffort/pod0e141d4f-c9ab-11e7-89f4-c6b053eac242 2017-11-15 02:16:57.203194711 +0000 UC containerCreation {<nil>}}
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.733131 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/run-1482.scope"
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.733140 27751 factory.go:105] Error trying to work out if we can handle /system.slice/run-1482.scope: /system.slice/run-1482.scope not handledby systemd handler
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.733145 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/run-1482.scope"
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.733151 27751 factory.go:112] Using factory "raw" for container "/system.slice/run-1482.scope"
Nov 15 02:16:57 af867b kubelet[27751]: W1115 02:16:57.733219 27751 container.go:354] Failed to create summary reader for "/system.slice/run-1482.scope": none of the resources are being tracked.
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.733230 27751 manager.go:932] Added container: "/system.slice/run-1482.scope" (aliases: [], namespace: "")
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.733260 27751 handler.go:325] Added event &{/system.slice/run-1482.scope 0001-01-01 00:00:00 +0000 UTC containerCreation {<nil>}}
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.733276 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-92b5e8c6024d48911cd69220d7bd1096230880041c5fccc7046f3e126499b74.mount"
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.733286 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-92b5e8c6024d489118cd69220dbd1096230880041c5fccc7046f3e126499b74.mount", but ignoring.
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.733297 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-92b5e8c6024d489118cd69220d7bd1096230880041c5fcc7046f3e126499b74.mount"
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.733306 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-kubelet-pods-0d5631da\\x2dc9ab\\x2d11e7\\2d89f4\\x2dc6b053eac242-volumes-kubernetes.io\\x7esecret-default\\x2dtoken\\x2dqjbsf.mount"
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.733316 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-kubelet-pods-0d5631da\\x2dc9ab\\x2d11e7\\x2d89f4\\2dc6b053eac242-volumes-kubernetes.io\\x7esecret-default\\x2dtoken\\x2dqjbsf.mount", but ignoring.
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.733327 27751 manager.go:901] ignoring container "/system.slice/var-lib-kubelet-pods-0d5631da\\x2dc9ab\\x2d11e7\\x2d89f4\\x2dc6b053eac242-volues-kubernetes.io\\x7esecret-default\\x2dtoken\\x2dqjbsf.mount"
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.733342 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-5fb7fcf7c72f470128f78883221316e45e7426df36260371e72e5a65e412678-shm.mount"
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.733351 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-5fb7fcf7c72f470128f788803221316e5e7426df36260371e72e5a65e412678-shm.mount", but ignoring.
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.733361 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-containers-5fb7fcf7c72f470128f788803221316e45e7426df36260371e725a65e412678-shm.mount"
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.733373 27751 manager.go:989] Destroyed container: "/system.slice/run-1482.scope" (aliases: [], namespace: "")
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.733381 27751 handler.go:325] Added event &{/system.slice/run-1482.scope 2017-11-15 02:16:57.733377639 +0000 UTC containerDeletion {<nil>}}
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.733402 27751 container.go:409] Start housekeeping for container "/system.slice/run-1482.scope"
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.733416 27751 container.go:409] Start housekeeping for container "/kubepods/besteffort/pod0d5631da-c9ab-11e7-89f4-c6b053eac242/5fb7fcf7c72f47028f788803221316e45e7426df36260371e72e5a65e412678"
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.752447 27751 container.go:409] Start housekeeping for container "/kubepods/besteffort/pod0e141d4f-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.764449 27751 generic.go:146] GenericPLEG: 0d5631da-c9ab-11e7-89f4-c6b053eac242/5fb7fcf7c72f470128f788803221316e45e7426df36260371e72e5a65e41268: non-existent -> running
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.765929 27751 cni.go:326] Got netns path /proc/1421/ns/net
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.765940 27751 cni.go:327] Using netns path default
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.766042 27751 cni.go:298] About to add CNI network weave (type=weave-net)
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.797861 27751 config.go:282] Setting pods for source api
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.798081 27751 kuberuntime_manager.go:833] getSandboxIDByPodUID got sandbox IDs ["5fb7fcf7c72f470128f788803221316e45e7426df36260371e72e5a65e41278"] for pod "snginx-10_default(0d5631da-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.798198 27751 status_manager.go:451] Status for pod "snginx-25_default(0e141d4f-c9ab-11e7-89f4-c6b053eac242)" updated successfully: (1, {PhasePending Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:16:57 +0000 UTC Reason: Message:} {Type:Ready Status:False LastProbeTie:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:16:57 +0000 UTC Reason:ContainersNotReady Message:containers with unready status: [nginx]} {Type:PodScheduled Status:True LastProbeTme:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:16:57 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP: StartTime:2017-11-15 02:16:57 +0000 UTC InitContainrStatuses:[] ContainerStatuses:[{Name:nginx State:{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,} Running:nil Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Termiated:nil} Ready:false RestartCount:0 Image:nginx ImageID: ContainerID:}] QOSClass:BestEffort})
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.804291 27751 generic.go:345] PLEG: Write status for snginx-10/default: &container.PodStatus{ID:"0d5631da-c9ab-11e7-89f4-c6b053eac242", Name:"nginx-10", Namespace:"default", IP:"", ContainerStatuses:[]*container.ContainerStatus{}, SandboxStatuses:[]*runtime.PodSandboxStatus{(*runtime.PodSandboxStatus)(0xc421e424b0)}} (err: <nil>)
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.804346 27751 kubelet.go:1871] SyncLoop (PLEG): "snginx-10_default(0d5631da-c9ab-11e7-89f4-c6b053eac242)", event: &pleg.PodLifecycleEvent{ID:"d5631da-c9ab-11e7-89f4-c6b053eac242", Type:"ContainerStarted", Data:"5fb7fcf7c72f470128f788803221316e45e7426df36260371e72e5a65e412678"}
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.807456 27751 kubelet.go:1850] SyncLoop (RECONCILE, "api"): "snginx-25_default(0e141d4f-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.914004 27751 config.go:282] Setting pods for source api
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.915783 27751 config.go:404] Receiving a new pod "snginx-16_default(0e4740f5-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.916308 27751 kubelet.go:1837] SyncLoop (ADD, "api"): "snginx-16_default(0e4740f5-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.916443 27751 kubelet_pods.go:1284] Generating status for "snginx-16_default(0e4740f5-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.917493 27751 qos_container_manager_linux.go:320] [ContainerManager]: Updated QoS cgroup configuration
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.923054 27751 factory.go:116] Factory "docker" was unable to handle container "/kubepods/besteffort/pod0e4740f5-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.926198 27751 factory.go:105] Error trying to work out if we can handle /kubepods/besteffort/pod0e4740f5-c9ab-11e7-89f4-c6b053eac242: /kubepod/besteffort/pod0e4740f5-c9ab-11e7-89f4-c6b053eac242 not handled by systemd handler
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.926212 27751 factory.go:116] Factory "systemd" was unable to handle container "/kubepods/besteffort/pod0e4740f5-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.926220 27751 factory.go:112] Using factory "raw" for container "/kubepods/besteffort/pod0e4740f5-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.927546 27751 manager.go:932] Added container: "/kubepods/besteffort/pod0e4740f5-c9ab-11e7-89f4-c6b053eac242" (aliases: [], namespace: "")
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.933104 27751 handler.go:325] Added event &{/kubepods/besteffort/pod0e4740f5-c9ab-11e7-89f4-c6b053eac242 2017-11-15 02:16:57.920194711 +0000 UC containerCreation {<nil>}}
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.933172 27751 container.go:409] Start housekeeping for container "/kubepods/besteffort/pod0e4740f5-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.943182 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "snginx-16_default(0e4740f5-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.946662 27751 status_manager.go:451] Status for pod "snginx-16_default(0e4740f5-c9ab-11e7-89f4-c6b053eac242)" updated successfully: (1, {PhasePending Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:16:57 +0000 UTC Reason: Message:} {Type:Ready Status:False LastProbeTie:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:16:57 +0000 UTC Reason:ContainersNotReady Message:containers with unready status: [nginx]} {Type:PodScheduled Status:True LastProbeTme:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:16:57 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP: StartTime:2017-11-15 02:16:57 +0000 UTC InitContainrStatuses:[] ContainerStatuses:[{Name:nginx State:{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,} Running:nil Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Termiated:nil} Ready:false RestartCount:0 Image:nginx ImageID: ContainerID:}] QOSClass:BestEffort})
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.946829 27751 config.go:282] Setting pods for source api
Nov 15 02:16:57 af867b kubelet[27751]: I1115 02:16:57.948518 27751 kubelet.go:1850] SyncLoop (RECONCILE, "api"): "snginx-16_default(0e4740f5-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.096722 27751 kuberuntime_manager.go:640] Created PodSandbox "5fb7fcf7c72f470128f788803221316e45e7426df36260371e72e5a65e412678" for pod "sngin-10_default(0d5631da-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.116791 27751 reconciler.go:212] operationExecutor.VerifyControllerAttachedVolume started for volume "default-token-qjbsf" (UniqueName: "kuberetes.io/secret/0e4740f5-c9ab-11e7-89f4-c6b053eac242-default-token-qjbsf") pod "snginx-16" (UID: "0e4740f5-c9ab-11e7-89f4-c6b053eac242")
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.176584 27751 config.go:282] Setting pods for source api
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.177575 27751 config.go:404] Receiving a new pod "snginx-28_default(0eb57d14-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.179040 27751 kubelet.go:1837] SyncLoop (ADD, "api"): "snginx-28_default(0eb57d14-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.179193 27751 kubelet_pods.go:1284] Generating status for "snginx-28_default(0eb57d14-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.179473 27751 qos_container_manager_linux.go:320] [ContainerManager]: Updated QoS cgroup configuration
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.185043 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "snginx-28_default(0eb57d14-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.185106 27751 factory.go:116] Factory "docker" was unable to handle container "/kubepods/besteffort/pod0eb57d14-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.185121 27751 factory.go:105] Error trying to work out if we can handle /kubepods/besteffort/pod0eb57d14-c9ab-11e7-89f4-c6b053eac242: /kubepod/besteffort/pod0eb57d14-c9ab-11e7-89f4-c6b053eac242 not handled by systemd handler
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.185128 27751 factory.go:116] Factory "systemd" was unable to handle container "/kubepods/besteffort/pod0eb57d14-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.185137 27751 factory.go:112] Using factory "raw" for container "/kubepods/besteffort/pod0eb57d14-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.185286 27751 manager.go:932] Added container: "/kubepods/besteffort/pod0eb57d14-c9ab-11e7-89f4-c6b053eac242" (aliases: [], namespace: "")
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.185390 27751 handler.go:325] Added event &{/kubepods/besteffort/pod0eb57d14-c9ab-11e7-89f4-c6b053eac242 2017-11-15 02:16:58.183194711 +0000 UC containerCreation {<nil>}}
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.185427 27751 container.go:409] Start housekeeping for container "/kubepods/besteffort/pod0eb57d14-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.210909 27751 status_manager.go:451] Status for pod "snginx-28_default(0eb57d14-c9ab-11e7-89f4-c6b053eac242)" updated successfully: (1, {PhasePending Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:16:58 +0000 UTC Reason: Message:} {Type:Ready Status:False LastProbeTie:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:16:58 +0000 UTC Reason:ContainersNotReady Message:containers with unready status: [nginx]} {Type:PodScheduled Status:True LastProbeTme:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:16:58 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP: StartTime:2017-11-15 02:16:58 +0000 UTC InitContainrStatuses:[] ContainerStatuses:[{Name:nginx State:{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,} Running:nil Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Termiated:nil} Ready:false RestartCount:0 Image:nginx ImageID: ContainerID:}] QOSClass:BestEffort})
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.211040 27751 config.go:282] Setting pods for source api
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.213793 27751 kubelet.go:1850] SyncLoop (RECONCILE, "api"): "snginx-28_default(0eb57d14-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.217167 27751 reconciler.go:257] operationExecutor.MountVolume started for volume "default-token-qjbsf" (UniqueName: "kubernetes.io/secret/0e440f5-c9ab-11e7-89f4-c6b053eac242-default-token-qjbsf") pod "snginx-16" (UID: "0e4740f5-c9ab-11e7-89f4-c6b053eac242")
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.217206 27751 reconciler.go:212] operationExecutor.VerifyControllerAttachedVolume started for volume "default-token-qjbsf" (UniqueName: "kuberetes.io/secret/0eb57d14-c9ab-11e7-89f4-c6b053eac242-default-token-qjbsf") pod "snginx-28" (UID: "0eb57d14-c9ab-11e7-89f4-c6b053eac242")
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.217281 27751 secret.go:186] Setting up volume default-token-qjbsf for pod 0e4740f5-c9ab-11e7-89f4-c6b053eac242 at /var/lib/kubelet/pods/0e474f5-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.217455 27751 empty_dir.go:264] pod 0e4740f5-c9ab-11e7-89f4-c6b053eac242: mounting tmpfs for volume wrapped_default-token-qjbsf
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.217473 27751 mount_linux.go:135] Mounting cmd (systemd-run) with arguments ([--description=Kubernetes transient mount for /var/lib/kubelet/pos/0e4740f5-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf --scope -- mount -t tmpfs tmpfs /var/lib/kubelet/pods/0e4740f5-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetesio~secret/default-token-qjbsf])
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.229494 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/run-1596.scope"
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.229524 27751 factory.go:105] Error trying to work out if we can handle /system.slice/run-1596.scope: /system.slice/run-1596.scope not handledby systemd handler
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.229534 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/run-1596.scope"
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.229542 27751 factory.go:112] Using factory "raw" for container "/system.slice/run-1596.scope"
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.229715 27751 manager.go:932] Added container: "/system.slice/run-1596.scope" (aliases: [], namespace: "")
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.229834 27751 handler.go:325] Added event &{/system.slice/run-1596.scope 2017-11-15 02:16:58.227194711 +0000 UTC containerCreation {<nil>}}
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.229860 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/run-docker-netns-8930558db511.mount"
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.229867 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/run-docker-netns-8930558db511.mount", but ignoring.
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.229877 27751 manager.go:901] ignoring container "/system.slice/run-docker-netns-8930558db511.mount"
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.229884 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-kubelet-pods-0e141d4f\\x2dc9ab\\x2d11e7\\2d89f4\\x2dc6b053eac242-volumes-kubernetes.io\\x7esecret-default\\x2dtoken\\x2dqjbsf.mount"
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.229897 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-kubelet-pods-0e141d4f\\x2dc9ab\\x2d11e7\\x2d89f4\\2dc6b053eac242-volumes-kubernetes.io\\x7esecret-default\\x2dtoken\\x2dqjbsf.mount", but ignoring.
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.229908 27751 manager.go:901] ignoring container "/system.slice/var-lib-kubelet-pods-0e141d4f\\x2dc9ab\\x2d11e7\\x2d89f4\\x2dc6b053eac242-volues-kubernetes.io\\x7esecret-default\\x2dtoken\\x2dqjbsf.mount"
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.229932 27751 container.go:409] Start housekeeping for container "/system.slice/run-1596.scope"
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.238462 27751 manager.go:989] Destroyed container: "/system.slice/run-1596.scope" (aliases: [], namespace: "")
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.238486 27751 handler.go:325] Added event &{/system.slice/run-1596.scope 2017-11-15 02:16:58.238479785 +0000 UTC containerDeletion {<nil>}}
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.318103 27751 reconciler.go:257] operationExecutor.MountVolume started for volume "default-token-qjbsf" (UniqueName: "kubernetes.io/secret/0eb7d14-c9ab-11e7-89f4-c6b053eac242-default-token-qjbsf") pod "snginx-28" (UID: "0eb57d14-c9ab-11e7-89f4-c6b053eac242")
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.318178 27751 secret.go:186] Setting up volume default-token-qjbsf for pod 0eb57d14-c9ab-11e7-89f4-c6b053eac242 at /var/lib/kubelet/pods/0eb5714-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.318355 27751 empty_dir.go:264] pod 0eb57d14-c9ab-11e7-89f4-c6b053eac242: mounting tmpfs for volume wrapped_default-token-qjbsf
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.318380 27751 mount_linux.go:135] Mounting cmd (systemd-run) with arguments ([--description=Kubernetes transient mount for /var/lib/kubelet/pos/0eb57d14-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf --scope -- mount -t tmpfs tmpfs /var/lib/kubelet/pods/0eb57d14-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetesio~secret/default-token-qjbsf])
Nov 15 02:16:58 af867b kubelet[27751]: W1115 02:16:58.357080 27751 raw.go:87] Error while processing event ("/sys/fs/cgroup/cpu,cpuacct/system.slice/run-1603.scope": 0x40000100 == IN_CREATE|IN_ISIR): inotify_add_watch /sys/fs/cgroup/cpu,cpuacct/system.slice/run-1603.scope: no such file or directory
Nov 15 02:16:58 af867b kubelet[27751]: W1115 02:16:58.357833 27751 raw.go:87] Error while processing event ("/sys/fs/cgroup/blkio/system.slice/run-1603.scope": 0x40000100 == IN_CREATE|IN_ISDIR): notify_add_watch /sys/fs/cgroup/blkio/system.slice/run-1603.scope: no such file or directory
Nov 15 02:16:58 af867b kubelet[27751]: W1115 02:16:58.358627 27751 raw.go:87] Error while processing event ("/sys/fs/cgroup/memory/system.slice/run-1603.scope": 0x40000100 == IN_CREATE|IN_ISDIR):inotify_add_watch /sys/fs/cgroup/memory/system.slice/run-1603.scope: no such file or directory
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.359187 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-kubelet-pods-0e4740f5\\x2dc9ab\\x2d11e7\\2d89f4\\x2dc6b053eac242-volumes-kubernetes.io\\x7esecret-default\\x2dtoken\\x2dqjbsf.mount"
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.359937 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-kubelet-pods-0e4740f5\\x2dc9ab\\x2d11e7\\x2d89f4\\2dc6b053eac242-volumes-kubernetes.io\\x7esecret-default\\x2dtoken\\x2dqjbsf.mount", but ignoring.
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.359951 27751 manager.go:901] ignoring container "/system.slice/var-lib-kubelet-pods-0e4740f5\\x2dc9ab\\x2d11e7\\x2d89f4\\x2dc6b053eac242-volues-kubernetes.io\\x7esecret-default\\x2dtoken\\x2dqjbsf.mount"
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.366666 27751 secret.go:217] Received secret default/default-token-qjbsf containing (3) pieces of data, 1878 total bytes
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.366768 27751 atomic_writer.go:145] pod default/snginx-28 volume default-token-qjbsf: write required for target directory /var/lib/kubelet/pod/0eb57d14-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.366863 27751 atomic_writer.go:160] pod default/snginx-28 volume default-token-qjbsf: performed write of new data to ts data directory: /var/lb/kubelet/pods/0eb57d14-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf/..119811_15_11_02_16_58.948849759
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.366955 27751 operation_generator.go:484] MountVolume.SetUp succeeded for volume "default-token-qjbsf" (UniqueName: "kubernetes.io/secret/0eb5d14-c9ab-11e7-89f4-c6b053eac242-default-token-qjbsf") pod "snginx-28" (UID: "0eb57d14-c9ab-11e7-89f4-c6b053eac242")
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.367216 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"default", Name:"snginx-28", UID:"0eb57d14-c9ab-11e7-89f4-c6b053ea242", APIVersion:"v1", ResourceVersion:"1325", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "default-token-qjbsf"
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.378118 27751 kuberuntime_manager.go:654] Determined the ip "10.32.0.3" for pod "snginx-10_default(0d5631da-c9ab-11e7-89f4-c6b053eac242)" afte sandbox changed
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.378227 27751 kuberuntime_manager.go:705] Creating container &Container{Name:nginx,Image:nginx,Command:[],Args:[],WorkingDir:,Ports:[{ 0 80 TC }],Env:[],Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[{default-token-qjbsf true /var/run/secrets/kubernetes.io/serviceaccount <nil>}],LivenessProb:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[],TerminationMessagePolic:File,} in pod snginx-10_default(0d5631da-c9ab-11e7-89f4-c6b053eac242)
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.380710 27751 provider.go:119] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.380760 27751 config.go:131] looking for config.json at /var/lib/kubelet/config.json
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.380802 27751 config.go:131] looking for config.json at /config.json
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.380817 27751 config.go:131] looking for config.json at /.docker/config.json
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.380825 27751 config.go:131] looking for config.json at /.docker/config.json
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.380838 27751 config.go:101] looking for .dockercfg at /var/lib/kubelet/.dockercfg
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.380849 27751 config.go:101] looking for .dockercfg at /.dockercfg
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.380860 27751 config.go:101] looking for .dockercfg at /.dockercfg
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.380867 27751 config.go:101] looking for .dockercfg at /.dockercfg
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.380876 27751 provider.go:89] Unable to parse Docker config file: couldn't find valid .dockercfg after checking in [/var/lib/kubelet /]
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.380888 27751 kuberuntime_image.go:46] Pulling image "docker.io/library/nginx:latest" without credentials
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.380942 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"default", Name:"snginx-10", UID:"0d5631da-c9ab-11e7-89f4-c6b053ea242", APIVersion:"v1", ResourceVersion:"1308", FieldPath:"spec.containers{nginx}"}): type: 'Normal' reason: 'Pulling' pulling image "nginx"
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.485875 27751 volume_manager.go:366] All volumes are attached and mounted for pod "snginx-28_default(0eb57d14-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.485927 27751 kuberuntime_manager.go:370] No sandbox for pod "snginx-28_default(0eb57d14-c9ab-11e7-89f4-c6b053eac242)" can be found. Need to sart a new one
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.485942 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:true CreateSandbox:true SandboxID: Attempt:0 NextInitContainerToStartnil ContainersToStart:[0] ContainersToKill:map[]} for pod "snginx-28_default(0eb57d14-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.486006 27751 kuberuntime_manager.go:565] SyncPod received new pod "snginx-28_default(0eb57d14-c9ab-11e7-89f4-c6b053eac242)", will create a sadbox for it
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.486015 27751 kuberuntime_manager.go:574] Stopping PodSandbox for "snginx-28_default(0eb57d14-c9ab-11e7-89f4-c6b053eac242)", will start new on
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.486035 27751 kuberuntime_manager.go:626] Creating sandbox for pod "snginx-28_default(0eb57d14-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.490113 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/besteffort/pod0eb57d14-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.490131 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/besteffort/pod0eb57d14-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.524080 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.639816 27751 secret.go:217] Received secret default/default-token-qjbsf containing (3) pieces of data, 1878 total bytes
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.639901 27751 atomic_writer.go:145] pod default/snginx-16 volume default-token-qjbsf: write required for target directory /var/lib/kubelet/pod/0e4740f5-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.639982 27751 atomic_writer.go:160] pod default/snginx-16 volume default-token-qjbsf: performed write of new data to ts data directory: /var/lb/kubelet/pods/0e4740f5-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf/..119811_15_11_02_16_58.130212914
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.640068 27751 operation_generator.go:484] MountVolume.SetUp succeeded for volume "default-token-qjbsf" (UniqueName: "kubernetes.io/secret/0e470f5-c9ab-11e7-89f4-c6b053eac242-default-token-qjbsf") pod "snginx-16" (UID: "0e4740f5-c9ab-11e7-89f4-c6b053eac242")
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.640308 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"default", Name:"snginx-16", UID:"0e4740f5-c9ab-11e7-89f4-c6b053ea242", APIVersion:"v1", ResourceVersion:"1320", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "default-token-qjbsf"
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.847645 27751 volume_manager.go:366] All volumes are attached and mounted for pod "snginx-16_default(0e4740f5-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.847699 27751 kuberuntime_manager.go:370] No sandbox for pod "snginx-16_default(0e4740f5-c9ab-11e7-89f4-c6b053eac242)" can be found. Need to sart a new one
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.847713 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:true CreateSandbox:true SandboxID: Attempt:0 NextInitContainerToStartnil ContainersToStart:[0] ContainersToKill:map[]} for pod "snginx-16_default(0e4740f5-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.847772 27751 kuberuntime_manager.go:565] SyncPod received new pod "snginx-16_default(0e4740f5-c9ab-11e7-89f4-c6b053eac242)", will create a sadbox for it
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.847780 27751 kuberuntime_manager.go:574] Stopping PodSandbox for "snginx-16_default(0e4740f5-c9ab-11e7-89f4-c6b053eac242)", will start new on
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.847801 27751 kuberuntime_manager.go:626] Creating sandbox for pod "snginx-16_default(0e4740f5-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.852428 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/besteffort/pod0e4740f5-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:16:58 af867b kubelet[27751]: I1115 02:16:58.852447 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/besteffort/pod0e4740f5-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:16:59 af867b kubelet[27751]: I1115 02:16:59.214620 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10252, Path: /healthz
Nov 15 02:16:59 af867b kubelet[27751]: I1115 02:16:59.214670 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:16:59 af867b kubelet[27751]: I1115 02:16:59.215970 27751 http.go:96] Probe succeeded for http://127.0.0.1:10252/healthz, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 0216:59 GMT] Content-Length:[2] Content-Type:[text/plain; charset=utf-8]] 0xc421bf2da0 2 [] true false map[] 0xc421e41500 <nil>}
Nov 15 02:16:59 af867b kubelet[27751]: I1115 02:16:59.216017 27751 prober.go:113] Liveness probe for "kube-controller-manager-af867b_kube-system(f49ee4da5c66af63a0b4bcea4f69baf9):kube-controller-anager" succeeded
Nov 15 02:16:59 af867b kubelet[27751]: I1115 02:16:59.234459 27751 docker_sandbox.go:691] Will attempt to re-write config file /var/lib/docker/containers/b5b36e0aa65ce37565a6b968f3df8c3fbe6f5bf70af61470f9c8b3e1f835576/resolv.conf with:
Nov 15 02:16:59 af867b kubelet[27751]: [nameserver 10.96.0.10 search default.svc.cluster.local svc.cluster.local cluster.local opcwlaas.oraclecloud.internal. options ndots:5]
Nov 15 02:16:59 af867b kubelet[27751]: I1115 02:16:59.234625 27751 plugins.go:392] Calling network plugin cni to set up pod "snginx-25_default"
Nov 15 02:16:59 af867b kubelet[27751]: I1115 02:16:59.238777 27751 factory.go:112] Using factory "docker" for container "/kubepods/besteffort/pod0e141d4f-c9ab-11e7-89f4-c6b053eac242/b5b36e0aa65ce7565a6b968f3df8c3fbe6f5bf701af61470f9c8b3e1f835576"
Nov 15 02:16:59 af867b kubelet[27751]: I1115 02:16:59.419575 27751 cni.go:326] Got netns path /proc/1642/ns/net
Nov 15 02:16:59 af867b kubelet[27751]: I1115 02:16:59.419591 27751 cni.go:327] Using netns path default
Nov 15 02:16:59 af867b kubelet[27751]: I1115 02:16:59.419731 27751 cni.go:298] About to add CNI network cni-loopback (type=loopback)
Nov 15 02:16:59 af867b kubelet[27751]: I1115 02:16:59.436579 27751 generic.go:146] GenericPLEG: 0e141d4f-c9ab-11e7-89f4-c6b053eac242/b5b36e0aa65ce37565a6b968f3df8c3fbe6f5bf701af61470f9c8b3e1f83556: non-existent -> running
Nov 15 02:16:59 af867b kubelet[27751]: I1115 02:16:59.442961 27751 cni.go:326] Got netns path /proc/1642/ns/net
Nov 15 02:16:59 af867b kubelet[27751]: I1115 02:16:59.442975 27751 cni.go:327] Using netns path default
Nov 15 02:16:59 af867b kubelet[27751]: I1115 02:16:59.443092 27751 cni.go:298] About to add CNI network weave (type=weave-net)
Nov 15 02:16:59 af867b kubelet[27751]: I1115 02:16:59.555930 27751 manager.go:932] Added container: "/kubepods/besteffort/pod0e141d4f-c9ab-11e7-89f4-c6b053eac242/b5b36e0aa65ce37565a6b968f3df8c3fb6f5bf701af61470f9c8b3e1f835576" (aliases: [k8s_POD_snginx-25_default_0e141d4f-c9ab-11e7-89f4-c6b053eac242_0 b5b36e0aa65ce37565a6b968f3df8c3fbe6f5bf701af61470f9c8b3e1f835576], namespace: "docker")
Nov 15 02:16:59 af867b kubelet[27751]: I1115 02:16:59.556083 27751 handler.go:325] Added event &{/kubepods/besteffort/pod0e141d4f-c9ab-11e7-89f4-c6b053eac242/b5b36e0aa65ce37565a6b968f3df8c3fbe6f5f701af61470f9c8b3e1f835576 2017-11-15 02:16:58.497194711 +0000 UTC containerCreation {<nil>}}
Nov 15 02:16:59 af867b kubelet[27751]: I1115 02:16:59.556131 27751 container.go:409] Start housekeeping for container "/kubepods/besteffort/pod0e141d4f-c9ab-11e7-89f4-c6b053eac242/b5b36e0aa65ce3765a6b968f3df8c3fbe6f5bf701af61470f9c8b3e1f835576"
Nov 15 02:16:59 af867b kubelet[27751]: I1115 02:16:59.676375 27751 kuberuntime_manager.go:640] Created PodSandbox "b5b36e0aa65ce37565a6b968f3df8c3fbe6f5bf701af61470f9c8b3e1f835576" for pod "sngin-25_default(0e141d4f-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:16:59 af867b kubelet[27751]: I1115 02:16:59.917614 27751 kuberuntime_manager.go:654] Determined the ip "10.32.0.4" for pod "snginx-25_default(0e141d4f-c9ab-11e7-89f4-c6b053eac242)" afte sandbox changed
Nov 15 02:16:59 af867b kubelet[27751]: I1115 02:16:59.917769 27751 kuberuntime_manager.go:705] Creating container &Container{Name:nginx,Image:nginx,Command:[],Args:[],WorkingDir:,Ports:[{ 0 80 TC }],Env:[],Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[{default-token-qjbsf true /var/run/secrets/kubernetes.io/serviceaccount <nil>}],LivenessProb:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[],TerminationMessagePolic:File,} in pod snginx-25_default(0e141d4f-c9ab-11e7-89f4-c6b053eac242)
Nov 15 02:16:59 af867b kubelet[27751]: I1115 02:16:59.919195 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"default", Name:"snginx-25", UID:"0e141d4f-c9ab-11e7-89f4-c6b053ea242", APIVersion:"v1", ResourceVersion:"1314", FieldPath:"spec.containers{nginx}"}): type: 'Normal' reason: 'Pulling' pulling image "nginx"
Nov 15 02:17:00 af867b kubelet[27751]: I1115 02:17:00.449784 27751 factory.go:112] Using factory "docker" for container "/kubepods/besteffort/pod0eb57d14-c9ab-11e7-89f4-c6b053eac242/0fab8639e26123f063ff03cc306506a34283f24ab70834abda8a699113657cc"
Nov 15 02:17:00 af867b kubelet[27751]: I1115 02:17:00.451830 27751 kuberuntime_manager.go:833] getSandboxIDByPodUID got sandbox IDs ["b5b36e0aa65ce37565a6b968f3df8c3fbe6f5bf701af61470f9c8b3e1f83576"] for pod "snginx-25_default(0e141d4f-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:00 af867b kubelet[27751]: I1115 02:17:00.452469 27751 docker_sandbox.go:691] Will attempt to re-write config file /var/lib/docker/containers/0fab8639e261263f063ff03cc306506a34283f24a70834abda8a699113657cc/resolv.conf with:
Nov 15 02:17:00 af867b kubelet[27751]: [nameserver 10.96.0.10 search default.svc.cluster.local svc.cluster.local cluster.local opcwlaas.oraclecloud.internal. options ndots:5]
Nov 15 02:17:00 af867b kubelet[27751]: I1115 02:17:00.452618 27751 plugins.go:392] Calling network plugin cni to set up pod "snginx-28_default"
Nov 15 02:17:00 af867b kubelet[27751]: I1115 02:17:00.454516 27751 manager.go:932] Added container: "/kubepods/besteffort/pod0eb57d14-c9ab-11e7-89f4-c6b053eac242/0fab8639e261263f063ff03cc306506a3283f24ab70834abda8a699113657cc" (aliases: [k8s_POD_snginx-28_default_0eb57d14-c9ab-11e7-89f4-c6b053eac242_0 0fab8639e261263f063ff03cc306506a34283f24ab70834abda8a699113657cc], namespace: "docker")
Nov 15 02:17:00 af867b kubelet[27751]: I1115 02:17:00.454679 27751 handler.go:325] Added event &{/kubepods/besteffort/pod0eb57d14-c9ab-11e7-89f4-c6b053eac242/0fab8639e261263f063ff03cc306506a3428324ab70834abda8a699113657cc 2017-11-15 02:16:59.698194711 +0000 UTC containerCreation {<nil>}}
Nov 15 02:17:00 af867b kubelet[27751]: I1115 02:17:00.455024 27751 container.go:409] Start housekeeping for container "/kubepods/besteffort/pod0eb57d14-c9ab-11e7-89f4-c6b053eac242/0fab8639e261263063ff03cc306506a34283f24ab70834abda8a699113657cc"
Nov 15 02:17:00 af867b kubelet[27751]: I1115 02:17:00.465914 27751 cni.go:326] Got netns path /proc/1844/ns/net
Nov 15 02:17:00 af867b kubelet[27751]: I1115 02:17:00.465946 27751 cni.go:327] Using netns path default
Nov 15 02:17:00 af867b kubelet[27751]: I1115 02:17:00.466289 27751 cni.go:298] About to add CNI network cni-loopback (type=loopback)
Nov 15 02:17:00 af867b kubelet[27751]: I1115 02:17:00.530175 27751 cni.go:326] Got netns path /proc/1844/ns/net
Nov 15 02:17:00 af867b kubelet[27751]: I1115 02:17:00.530196 27751 cni.go:327] Using netns path default
Nov 15 02:17:00 af867b kubelet[27751]: I1115 02:17:00.532189 27751 cni.go:298] About to add CNI network weave (type=weave-net)
Nov 15 02:17:00 af867b kubelet[27751]: I1115 02:17:00.823489 27751 factory.go:112] Using factory "docker" for container "/kubepods/besteffort/pod0e4740f5-c9ab-11e7-89f4-c6b053eac242/66b9c434d36eb27e5ef517d07f9a34fc5e788400450810f5065e3e2ce1a0d07"
Nov 15 02:17:00 af867b kubelet[27751]: I1115 02:17:00.825773 27751 manager.go:932] Added container: "/kubepods/besteffort/pod0e4740f5-c9ab-11e7-89f4-c6b053eac242/66b9c434d36eb927e5ef517d07f9a34fce788400450810f5065e3e2ce1a0d07" (aliases: [k8s_POD_snginx-16_default_0e4740f5-c9ab-11e7-89f4-c6b053eac242_0 66b9c434d36eb927e5ef517d07f9a34fc5e788400450810f5065e3e2ce1a0d07], namespace: "docker")
Nov 15 02:17:00 af867b kubelet[27751]: I1115 02:17:00.825912 27751 handler.go:325] Added event &{/kubepods/besteffort/pod0e4740f5-c9ab-11e7-89f4-c6b053eac242/66b9c434d36eb927e5ef517d07f9a34fc5e78400450810f5065e3e2ce1a0d07 2017-11-15 02:17:00.215194711 +0000 UTC containerCreation {<nil>}}
Nov 15 02:17:00 af867b kubelet[27751]: I1115 02:17:00.825953 27751 container.go:409] Start housekeeping for container "/kubepods/besteffort/pod0e4740f5-c9ab-11e7-89f4-c6b053eac242/66b9c434d36eb92e5ef517d07f9a34fc5e788400450810f5065e3e2ce1a0d07"
Nov 15 02:17:00 af867b kubelet[27751]: I1115 02:17:00.832924 27751 kuberuntime_manager.go:640] Created PodSandbox "0fab8639e261263f063ff03cc306506a34283f24ab70834abda8a699113657cc" for pod "sngin-28_default(0eb57d14-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:00 af867b kubelet[27751]: I1115 02:17:00.835053 27751 generic.go:345] PLEG: Write status for snginx-25/default: &container.PodStatus{ID:"0e141d4f-c9ab-11e7-89f4-c6b053eac242", Name:"nginx-25", Namespace:"default", IP:"10.32.0.4", ContainerStatuses:[]*container.ContainerStatus{}, SandboxStatuses:[]*runtime.PodSandboxStatus{(*runtime.PodSandboxStatus)(0xc4209887d0)}} (err: <nil>
Nov 15 02:17:00 af867b kubelet[27751]: I1115 02:17:00.835492 27751 docker_sandbox.go:691] Will attempt to re-write config file /var/lib/docker/containers/66b9c434d36eb927e5ef517d07f9a34fc5e78840050810f5065e3e2ce1a0d07/resolv.conf with:
Nov 15 02:17:00 af867b kubelet[27751]: [nameserver 10.96.0.10 search default.svc.cluster.local svc.cluster.local cluster.local opcwlaas.oraclecloud.internal. options ndots:5]
Nov 15 02:17:00 af867b kubelet[27751]: I1115 02:17:00.835659 27751 plugins.go:392] Calling network plugin cni to set up pod "snginx-16_default"
Nov 15 02:17:00 af867b kubelet[27751]: I1115 02:17:00.836174 27751 kubelet.go:1871] SyncLoop (PLEG): "snginx-25_default(0e141d4f-c9ab-11e7-89f4-c6b053eac242)", event: &pleg.PodLifecycleEvent{ID:"e141d4f-c9ab-11e7-89f4-c6b053eac242", Type:"ContainerStarted", Data:"b5b36e0aa65ce37565a6b968f3df8c3fbe6f5bf701af61470f9c8b3e1f835576"}
Nov 15 02:17:00 af867b kubelet[27751]: I1115 02:17:00.836212 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:17:00 af867b kubelet[27751]: I1115 02:17:00.848570 27751 cni.go:326] Got netns path /proc/1928/ns/net
Nov 15 02:17:00 af867b kubelet[27751]: I1115 02:17:00.848583 27751 cni.go:327] Using netns path default
Nov 15 02:17:00 af867b kubelet[27751]: I1115 02:17:00.848737 27751 cni.go:298] About to add CNI network cni-loopback (type=loopback)
Nov 15 02:17:00 af867b kubelet[27751]: I1115 02:17:00.881339 27751 cni.go:326] Got netns path /proc/1928/ns/net
Nov 15 02:17:00 af867b kubelet[27751]: I1115 02:17:00.881585 27751 cni.go:327] Using netns path default
Nov 15 02:17:00 af867b kubelet[27751]: I1115 02:17:00.881734 27751 cni.go:298] About to add CNI network weave (type=weave-net)
Nov 15 02:17:00 af867b kubelet[27751]: I1115 02:17:00.900226 27751 kuberuntime_manager.go:654] Determined the ip "10.32.0.5" for pod "snginx-28_default(0eb57d14-c9ab-11e7-89f4-c6b053eac242)" afte sandbox changed
Nov 15 02:17:00 af867b kubelet[27751]: I1115 02:17:00.901402 27751 kuberuntime_manager.go:705] Creating container &Container{Name:nginx,Image:nginx,Command:[],Args:[],WorkingDir:,Ports:[{ 0 80 TC }],Env:[],Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[{default-token-qjbsf true /var/run/secrets/kubernetes.io/serviceaccount <nil>}],LivenessProb:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[],TerminationMessagePolic:File,} in pod snginx-28_default(0eb57d14-c9ab-11e7-89f4-c6b053eac242)
Nov 15 02:17:00 af867b kubelet[27751]: I1115 02:17:00.908045 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"default", Name:"snginx-28", UID:"0eb57d14-c9ab-11e7-89f4-c6b053ea242", APIVersion:"v1", ResourceVersion:"1325", FieldPath:"spec.containers{nginx}"}): type: 'Normal' reason: 'Pulling' pulling image "nginx"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.015010 27751 config.go:282] Setting pods for source api
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.016617 27751 config.go:404] Receiving a new pod "snginx-1_default(106570f0-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.018534 27751 kubelet.go:1837] SyncLoop (ADD, "api"): "snginx-1_default(106570f0-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.018686 27751 kubelet_pods.go:1284] Generating status for "snginx-1_default(106570f0-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.025996 27751 qos_container_manager_linux.go:320] [ContainerManager]: Updated QoS cgroup configuration
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.028655 27751 factory.go:116] Factory "docker" was unable to handle container "/kubepods/besteffort/pod106570f0-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.028697 27751 factory.go:105] Error trying to work out if we can handle /kubepods/besteffort/pod106570f0-c9ab-11e7-89f4-c6b053eac242: /kubepod/besteffort/pod106570f0-c9ab-11e7-89f4-c6b053eac242 not handled by systemd handler
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.028705 27751 factory.go:116] Factory "systemd" was unable to handle container "/kubepods/besteffort/pod106570f0-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.028713 27751 factory.go:112] Using factory "raw" for container "/kubepods/besteffort/pod106570f0-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.028963 27751 manager.go:932] Added container: "/kubepods/besteffort/pod106570f0-c9ab-11e7-89f4-c6b053eac242" (aliases: [], namespace: "")
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.029177 27751 handler.go:325] Added event &{/kubepods/besteffort/pod106570f0-c9ab-11e7-89f4-c6b053eac242 2017-11-15 02:17:01.027194711 +0000 UC containerCreation {<nil>}}
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.029238 27751 container.go:409] Start housekeeping for container "/kubepods/besteffort/pod106570f0-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.045700 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "snginx-1_default(106570f0-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.094545 27751 config.go:282] Setting pods for source api
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.096928 27751 status_manager.go:451] Status for pod "snginx-1_default(106570f0-c9ab-11e7-89f4-c6b053eac242)" updated successfully: (1, {Phase:ending Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:17:01 +0000 UTC Reason: Message:} {Type:Ready Status:False LastProbeTim:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:17:01 +0000 UTC Reason:ContainersNotReady Message:containers with unready status: [nginx]} {Type:PodScheduled Status:True LastProbeTie:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:17:00 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP: StartTime:2017-11-15 02:17:01 +0000 UTC InitContaineStatuses:[] ContainerStatuses:[{Name:nginx State:{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,} Running:nil Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminted:nil} Ready:false RestartCount:0 Image:nginx ImageID: ContainerID:}] QOSClass:BestEffort})
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.101040 27751 kubelet.go:1850] SyncLoop (RECONCILE, "api"): "snginx-1_default(106570f0-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.225190 27751 reconciler.go:212] operationExecutor.VerifyControllerAttachedVolume started for volume "default-token-qjbsf" (UniqueName: "kuberetes.io/secret/106570f0-c9ab-11e7-89f4-c6b053eac242-default-token-qjbsf") pod "snginx-1" (UID: "106570f0-c9ab-11e7-89f4-c6b053eac242")
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.289495 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=true reason: mesage:
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.332296 27751 reconciler.go:257] operationExecutor.MountVolume started for volume "default-token-qjbsf" (UniqueName: "kubernetes.io/secret/10670f0-c9ab-11e7-89f4-c6b053eac242-default-token-qjbsf") pod "snginx-1" (UID: "106570f0-c9ab-11e7-89f4-c6b053eac242")
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.332370 27751 secret.go:186] Setting up volume default-token-qjbsf for pod 106570f0-c9ab-11e7-89f4-c6b053eac242 at /var/lib/kubelet/pods/10657f0-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.332580 27751 empty_dir.go:264] pod 106570f0-c9ab-11e7-89f4-c6b053eac242: mounting tmpfs for volume wrapped_default-token-qjbsf
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.332602 27751 mount_linux.go:135] Mounting cmd (systemd-run) with arguments ([--description=Kubernetes transient mount for /var/lib/kubelet/pos/106570f0-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf --scope -- mount -t tmpfs tmpfs /var/lib/kubelet/pods/106570f0-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetesio~secret/default-token-qjbsf])
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.365002 27751 kuberuntime_manager.go:640] Created PodSandbox "66b9c434d36eb927e5ef517d07f9a34fc5e788400450810f5065e3e2ce1a0d07" for pod "sngin-16_default(0e4740f5-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:01 af867b kubelet[27751]: W1115 02:17:01.382681 27751 raw.go:87] Error while processing event ("/sys/fs/cgroup/cpu,cpuacct/system.slice/run-2090.scope": 0x40000100 == IN_CREATE|IN_ISIR): inotify_add_watch /sys/fs/cgroup/cpu,cpuacct/system.slice/run-2090.scope: no such file or directory
Nov 15 02:17:01 af867b kubelet[27751]: W1115 02:17:01.382733 27751 raw.go:87] Error while processing event ("/sys/fs/cgroup/blkio/system.slice/run-2090.scope": 0x40000100 == IN_CREATE|IN_ISDIR): notify_add_watch /sys/fs/cgroup/blkio/system.slice/run-2090.scope: no such file or directory
Nov 15 02:17:01 af867b kubelet[27751]: W1115 02:17:01.383509 27751 raw.go:87] Error while processing event ("/sys/fs/cgroup/memory/system.slice/run-2090.scope": 0x40000100 == IN_CREATE|IN_ISDIR):inotify_add_watch /sys/fs/cgroup/memory/system.slice/run-2090.scope: no such file or directory
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.383691 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-kubelet-pods-0eb57d14\\x2dc9ab\\x2d11e7\\2d89f4\\x2dc6b053eac242-volumes-kubernetes.io\\x7esecret-default\\x2dtoken\\x2dqjbsf.mount"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.383712 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-kubelet-pods-0eb57d14\\x2dc9ab\\x2d11e7\\x2d89f4\\2dc6b053eac242-volumes-kubernetes.io\\x7esecret-default\\x2dtoken\\x2dqjbsf.mount", but ignoring.
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.383729 27751 manager.go:901] ignoring container "/system.slice/var-lib-kubelet-pods-0eb57d14\\x2dc9ab\\x2d11e7\\x2d89f4\\x2dc6b053eac242-volues-kubernetes.io\\x7esecret-default\\x2dtoken\\x2dqjbsf.mount"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.388446 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-66b9c434d36eb927e5ef51707f9a34fc5e788400450810f5065e3e2ce1a0d07-shm.mount"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.388476 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-66b9c434d36eb927e5ef517d07f9a34f5e788400450810f5065e3e2ce1a0d07-shm.mount", but ignoring.
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.388489 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-containers-66b9c434d36eb927e5ef517d07f9a34fc5e788400450810f50653e2ce1a0d07-shm.mount"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.388500 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-11d04dfdbcc14d0839b62173cbec9d6618e35afc75e3122e0b5e582cec7bd24.mount"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.388509 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-11d04dfdbcc14d083d9b62173cec9d6618e35afc75e3122e0b5e582cec7bd24.mount", but ignoring.
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.388519 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-11d04dfdbcc14d083d9b62173cbec9d6618e35afc75e312e0b5e582cec7bd24.mount"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.388528 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-0fab8639e261263f063ff03c306506a34283f24ab70834abda8a699113657cc-shm.mount"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.388536 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-0fab8639e261263f063ff03cc306506a4283f24ab70834abda8a699113657cc-shm.mount", but ignoring.
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.388545 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-containers-0fab8639e261263f063ff03cc306506a34283f24ab70834abda8699113657cc-shm.mount"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.388554 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/run-docker-netns-cb26277dba54.mount"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.388575 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/run-docker-netns-cb26277dba54.mount", but ignoring.
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.388588 27751 manager.go:901] ignoring container "/system.slice/run-docker-netns-cb26277dba54.mount"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.388596 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/run-docker-netns-b3cd55c6dd34.mount"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.388603 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/run-docker-netns-b3cd55c6dd34.mount", but ignoring.
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.388611 27751 manager.go:901] ignoring container "/system.slice/run-docker-netns-b3cd55c6dd34.mount"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.393069 27751 secret.go:217] Received secret default/default-token-qjbsf containing (3) pieces of data, 1878 total bytes
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.393161 27751 atomic_writer.go:145] pod default/snginx-1 volume default-token-qjbsf: write required for target directory /var/lib/kubelet/pods106570f0-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.393274 27751 atomic_writer.go:160] pod default/snginx-1 volume default-token-qjbsf: performed write of new data to ts data directory: /var/li/kubelet/pods/106570f0-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf/..119811_15_11_02_17_01.269455337
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.393396 27751 operation_generator.go:484] MountVolume.SetUp succeeded for volume "default-token-qjbsf" (UniqueName: "kubernetes.io/secret/10650f0-c9ab-11e7-89f4-c6b053eac242-default-token-qjbsf") pod "snginx-1" (UID: "106570f0-c9ab-11e7-89f4-c6b053eac242")
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.393637 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"default", Name:"snginx-1", UID:"106570f0-c9ab-11e7-89f4-c6b053eac42", APIVersion:"v1", ResourceVersion:"1337", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "default-token-qjbsf"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.398152 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-c1fb4b3943b5e595ebc1e33980b6b2bca6fac3ac5b74374d709559e02d55896.mount"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.398177 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-c1fb4b3943b5e595ecbc1e3398b6b2bca6fac3ac5b74374d709559e02d55896.mount", but ignoring.
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.398188 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-c1fb4b3943b5e595ecbc1e33980b6b2bca6fac3ac5b7434d709559e02d55896.mount"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.401591 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-containers-b5b36e0aa65ce37565a6b96f3df8c3fbe6f5bf701af61470f9c8b3e1f835576-shm.mount"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.401608 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-containers-b5b36e0aa65ce37565a6b968f3df8c3fe6f5bf701af61470f9c8b3e1f835576-shm.mount", but ignoring.
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.401622 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-containers-b5b36e0aa65ce37565a6b968f3df8c3fbe6f5bf701af61470f9cb3e1f835576-shm.mount"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.410344 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/run-docker-netns-9a5667b14714.mount"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.410372 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/run-docker-netns-9a5667b14714.mount", but ignoring.
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.410382 27751 manager.go:901] ignoring container "/system.slice/run-docker-netns-9a5667b14714.mount"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.410390 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-docker-devicemapper-mnt-7849cd5bef6ac97e51cb3dff128a214cd542fe2d505cf379e8cacea700428f5.mount"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.410399 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-docker-devicemapper-mnt-7849cd5bef6ac97e5e1cb3dff18a214cd542fe2d505cf379e8cacea700428f5.mount", but ignoring.
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.410408 27751 manager.go:901] ignoring container "/system.slice/var-lib-docker-devicemapper-mnt-7849cd5bef6ac97e5e1cb3dff128a214cd542fe2d505cf79e8cacea700428f5.mount"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.412660 27751 kuberuntime_manager.go:654] Determined the ip "10.32.0.6" for pod "snginx-16_default(0e4740f5-c9ab-11e7-89f4-c6b053eac242)" afte sandbox changed
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.412782 27751 kuberuntime_manager.go:705] Creating container &Container{Name:nginx,Image:nginx,Command:[],Args:[],WorkingDir:,Ports:[{ 0 80 TC }],Env:[],Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[{default-token-qjbsf true /var/run/secrets/kubernetes.io/serviceaccount <nil>}],LivenessProb:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[],TerminationMessagePolic:File,} in pod snginx-16_default(0e4740f5-c9ab-11e7-89f4-c6b053eac242)
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.414115 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"default", Name:"snginx-16", UID:"0e4740f5-c9ab-11e7-89f4-c6b053ea242", APIVersion:"v1", ResourceVersion:"1320", FieldPath:"spec.containers{nginx}"}): type: 'Normal' reason: 'Pulling' pulling image "nginx"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.546967 27751 config.go:282] Setting pods for source api
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.548807 27751 config.go:404] Receiving a new pod "snginx-6_default(10b34ac6-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.549285 27751 kubelet.go:1837] SyncLoop (ADD, "api"): "snginx-6_default(10b34ac6-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.549443 27751 kubelet_pods.go:1284] Generating status for "snginx-6_default(10b34ac6-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.549790 27751 qos_container_manager_linux.go:320] [ContainerManager]: Updated QoS cgroup configuration
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.552868 27751 factory.go:116] Factory "docker" was unable to handle container "/kubepods/besteffort/pod10b34ac6-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.552899 27751 factory.go:105] Error trying to work out if we can handle /kubepods/besteffort/pod10b34ac6-c9ab-11e7-89f4-c6b053eac242: /kubepod/besteffort/pod10b34ac6-c9ab-11e7-89f4-c6b053eac242 not handled by systemd handler
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.552924 27751 factory.go:116] Factory "systemd" was unable to handle container "/kubepods/besteffort/pod10b34ac6-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.552933 27751 factory.go:112] Using factory "raw" for container "/kubepods/besteffort/pod10b34ac6-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.553123 27751 manager.go:932] Added container: "/kubepods/besteffort/pod10b34ac6-c9ab-11e7-89f4-c6b053eac242" (aliases: [], namespace: "")
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.553207 27751 handler.go:325] Added event &{/kubepods/besteffort/pod10b34ac6-c9ab-11e7-89f4-c6b053eac242 2017-11-15 02:17:01.552194711 +0000 UC containerCreation {<nil>}}
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.553251 27751 container.go:409] Start housekeeping for container "/kubepods/besteffort/pod10b34ac6-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:01 af867b kubelet[27751]: E1115 02:17:01.553551 27751 helpers.go:468] PercpuUsage had 0 cpus, but the actual number is 2; ignoring extra CPUs
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.557754 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "snginx-6_default(10b34ac6-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.567869 27751 config.go:282] Setting pods for source api
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.569746 27751 config.go:404] Receiving a new pod "snginx-4_default(10af02e8-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.569986 27751 kubelet.go:1837] SyncLoop (ADD, "api"): "snginx-4_default(10af02e8-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.570125 27751 kubelet_pods.go:1284] Generating status for "snginx-4_default(10af02e8-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.570363 27751 qos_container_manager_linux.go:320] [ContainerManager]: Updated QoS cgroup configuration
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.575222 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "snginx-4_default(10af02e8-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.576775 27751 factory.go:116] Factory "docker" was unable to handle container "/kubepods/besteffort/pod10af02e8-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.576815 27751 factory.go:105] Error trying to work out if we can handle /kubepods/besteffort/pod10af02e8-c9ab-11e7-89f4-c6b053eac242: /kubepod/besteffort/pod10af02e8-c9ab-11e7-89f4-c6b053eac242 not handled by systemd handler
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.576823 27751 factory.go:116] Factory "systemd" was unable to handle container "/kubepods/besteffort/pod10af02e8-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.576830 27751 factory.go:112] Using factory "raw" for container "/kubepods/besteffort/pod10af02e8-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.576977 27751 manager.go:932] Added container: "/kubepods/besteffort/pod10af02e8-c9ab-11e7-89f4-c6b053eac242" (aliases: [], namespace: "")
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.577082 27751 handler.go:325] Added event &{/kubepods/besteffort/pod10af02e8-c9ab-11e7-89f4-c6b053eac242 2017-11-15 02:17:01.574194711 +0000 UC containerCreation {<nil>}}
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.577113 27751 container.go:409] Start housekeeping for container "/kubepods/besteffort/pod10af02e8-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.638655 27751 reconciler.go:212] operationExecutor.VerifyControllerAttachedVolume started for volume "default-token-qjbsf" (UniqueName: "kuberetes.io/secret/10b34ac6-c9ab-11e7-89f4-c6b053eac242-default-token-qjbsf") pod "snginx-6" (UID: "10b34ac6-c9ab-11e7-89f4-c6b053eac242")
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.638740 27751 reconciler.go:212] operationExecutor.VerifyControllerAttachedVolume started for volume "default-token-qjbsf" (UniqueName: "kuberetes.io/secret/10af02e8-c9ab-11e7-89f4-c6b053eac242-default-token-qjbsf") pod "snginx-4" (UID: "10af02e8-c9ab-11e7-89f4-c6b053eac242")
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.639989 27751 config.go:282] Setting pods for source api
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.642003 27751 kubelet.go:1850] SyncLoop (RECONCILE, "api"): "snginx-6_default(10b34ac6-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.646324 27751 volume_manager.go:366] All volumes are attached and mounted for pod "snginx-1_default(106570f0-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.646352 27751 kuberuntime_manager.go:370] No sandbox for pod "snginx-1_default(106570f0-c9ab-11e7-89f4-c6b053eac242)" can be found. Need to strt a new one
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.646370 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:true CreateSandbox:true SandboxID: Attempt:0 NextInitContainerToStartnil ContainersToStart:[0] ContainersToKill:map[]} for pod "snginx-1_default(106570f0-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.646435 27751 kuberuntime_manager.go:565] SyncPod received new pod "snginx-1_default(106570f0-c9ab-11e7-89f4-c6b053eac242)", will create a sanbox for it
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.646446 27751 kuberuntime_manager.go:574] Stopping PodSandbox for "snginx-1_default(106570f0-c9ab-11e7-89f4-c6b053eac242)", will start new one
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.646468 27751 kuberuntime_manager.go:626] Creating sandbox for pod "snginx-1_default(106570f0-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.649147 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/besteffort/pod106570f0-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.649173 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/besteffort/pod106570f0-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.653936 27751 status_manager.go:451] Status for pod "snginx-6_default(10b34ac6-c9ab-11e7-89f4-c6b053eac242)" updated successfully: (1, {Phase:ending Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:17:01 +0000 UTC Reason: Message:} {Type:Ready Status:False LastProbeTim:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:17:01 +0000 UTC Reason:ContainersNotReady Message:containers with unready status: [nginx]} {Type:PodScheduled Status:True LastProbeTie:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:17:01 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP: StartTime:2017-11-15 02:17:01 +0000 UTC InitContaineStatuses:[] ContainerStatuses:[{Name:nginx State:{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,} Running:nil Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminted:nil} Ready:false RestartCount:0 Image:nginx ImageID: ContainerID:}] QOSClass:BestEffort})
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.733038 27751 config.go:282] Setting pods for source api
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.735321 27751 config.go:404] Receiving a new pod "snginx-11_default(10c857c2-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.736980 27751 kubelet.go:1837] SyncLoop (ADD, "api"): "snginx-11_default(10c857c2-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.737138 27751 kubelet_pods.go:1284] Generating status for "snginx-11_default(10c857c2-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.737429 27751 qos_container_manager_linux.go:320] [ContainerManager]: Updated QoS cgroup configuration
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.740486 27751 reconciler.go:257] operationExecutor.MountVolume started for volume "default-token-qjbsf" (UniqueName: "kubernetes.io/secret/10b4ac6-c9ab-11e7-89f4-c6b053eac242-default-token-qjbsf") pod "snginx-6" (UID: "10b34ac6-c9ab-11e7-89f4-c6b053eac242")
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.740534 27751 reconciler.go:257] operationExecutor.MountVolume started for volume "default-token-qjbsf" (UniqueName: "kubernetes.io/secret/10a02e8-c9ab-11e7-89f4-c6b053eac242-default-token-qjbsf") pod "snginx-4" (UID: "10af02e8-c9ab-11e7-89f4-c6b053eac242")
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.740577 27751 secret.go:186] Setting up volume default-token-qjbsf for pod 10af02e8-c9ab-11e7-89f4-c6b053eac242 at /var/lib/kubelet/pods/10af0e8-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.740761 27751 empty_dir.go:264] pod 10af02e8-c9ab-11e7-89f4-c6b053eac242: mounting tmpfs for volume wrapped_default-token-qjbsf
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.740779 27751 mount_linux.go:135] Mounting cmd (systemd-run) with arguments ([--description=Kubernetes transient mount for /var/lib/kubelet/pos/10af02e8-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf --scope -- mount -t tmpfs tmpfs /var/lib/kubelet/pods/10af02e8-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetesio~secret/default-token-qjbsf])
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.746237 27751 status_manager.go:451] Status for pod "snginx-4_default(10af02e8-c9ab-11e7-89f4-c6b053eac242)" updated successfully: (1, {Phase:ending Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:17:01 +0000 UTC Reason: Message:} {Type:Ready Status:False LastProbeTim:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:17:01 +0000 UTC Reason:ContainersNotReady Message:containers with unready status: [nginx]} {Type:PodScheduled Status:True LastProbeTie:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:17:01 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP: StartTime:2017-11-15 02:17:01 +0000 UTC InitContaineStatuses:[] ContainerStatuses:[{Name:nginx State:{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,} Running:nil Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Terminted:nil} Ready:false RestartCount:0 Image:nginx ImageID: ContainerID:}] QOSClass:BestEffort})
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.752091 27751 secret.go:186] Setting up volume default-token-qjbsf for pod 10b34ac6-c9ab-11e7-89f4-c6b053eac242 at /var/lib/kubelet/pods/10b34c6-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.752247 27751 empty_dir.go:264] pod 10b34ac6-c9ab-11e7-89f4-c6b053eac242: mounting tmpfs for volume wrapped_default-token-qjbsf
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.752266 27751 mount_linux.go:135] Mounting cmd (systemd-run) with arguments ([--description=Kubernetes transient mount for /var/lib/kubelet/pos/10b34ac6-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf --scope -- mount -t tmpfs tmpfs /var/lib/kubelet/pods/10b34ac6-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetesio~secret/default-token-qjbsf])
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.755385 27751 factory.go:116] Factory "docker" was unable to handle container "/kubepods/besteffort/pod10c857c2-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.755421 27751 factory.go:105] Error trying to work out if we can handle /kubepods/besteffort/pod10c857c2-c9ab-11e7-89f4-c6b053eac242: /kubepod/besteffort/pod10c857c2-c9ab-11e7-89f4-c6b053eac242 not handled by systemd handler
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.755433 27751 factory.go:116] Factory "systemd" was unable to handle container "/kubepods/besteffort/pod10c857c2-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.755441 27751 factory.go:112] Using factory "raw" for container "/kubepods/besteffort/pod10c857c2-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.755729 27751 manager.go:932] Added container: "/kubepods/besteffort/pod10c857c2-c9ab-11e7-89f4-c6b053eac242" (aliases: [], namespace: "")
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.755869 27751 handler.go:325] Added event &{/kubepods/besteffort/pod10c857c2-c9ab-11e7-89f4-c6b053eac242 2017-11-15 02:17:01.749194711 +0000 UC containerCreation {<nil>}}
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.755899 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/run-2103.scope"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.755908 27751 factory.go:105] Error trying to work out if we can handle /system.slice/run-2103.scope: /system.slice/run-2103.scope not handledby systemd handler
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.755914 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/run-2103.scope"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.755921 27751 factory.go:112] Using factory "raw" for container "/system.slice/run-2103.scope"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.756063 27751 manager.go:932] Added container: "/system.slice/run-2103.scope" (aliases: [], namespace: "")
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.766470 27751 container.go:409] Start housekeeping for container "/kubepods/besteffort/pod10c857c2-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.767242 27751 handler.go:325] Added event &{/system.slice/run-2103.scope 2017-11-15 02:17:01.749194711 +0000 UTC containerCreation {<nil>}}
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.767266 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-kubelet-pods-106570f0\\x2dc9ab\\x2d11e7\\2d89f4\\x2dc6b053eac242-volumes-kubernetes.io\\x7esecret-default\\x2dtoken\\x2dqjbsf.mount"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.767280 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-kubelet-pods-106570f0\\x2dc9ab\\x2d11e7\\x2d89f4\\2dc6b053eac242-volumes-kubernetes.io\\x7esecret-default\\x2dtoken\\x2dqjbsf.mount", but ignoring.
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.767292 27751 manager.go:901] ignoring container "/system.slice/var-lib-kubelet-pods-106570f0\\x2dc9ab\\x2d11e7\\x2d89f4\\x2dc6b053eac242-volues-kubernetes.io\\x7esecret-default\\x2dtoken\\x2dqjbsf.mount"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.767317 27751 container.go:409] Start housekeeping for container "/system.slice/run-2103.scope"
Nov 15 02:17:01 af867b kubelet[27751]: E1115 02:17:01.767370 27751 helpers.go:468] PercpuUsage had 0 cpus, but the actual number is 2; ignoring extra CPUs
Nov 15 02:17:01 af867b kubelet[27751]: W1115 02:17:01.779138 27751 raw.go:87] Error while processing event ("/sys/fs/cgroup/blkio/system.slice/run-2106.scope": 0x40000100 == IN_CREATE|IN_ISDIR): notify_add_watch /sys/fs/cgroup/blkio/system.slice/run-2106.scope: no such file or directory
Nov 15 02:17:01 af867b kubelet[27751]: W1115 02:17:01.779172 27751 raw.go:87] Error while processing event ("/sys/fs/cgroup/memory/system.slice/run-2106.scope": 0x40000100 == IN_CREATE|IN_ISDIR):inotify_add_watch /sys/fs/cgroup/memory/system.slice/run-2106.scope: no such file or directory
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.780600 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "snginx-11_default(10c857c2-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.826200 27751 manager.go:989] Destroyed container: "/system.slice/run-2103.scope" (aliases: [], namespace: "")
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.826235 27751 handler.go:325] Added event &{/system.slice/run-2103.scope 2017-11-15 02:17:01.826224627 +0000 UTC containerDeletion {<nil>}}
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.826277 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/run-2106.scope"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.826289 27751 factory.go:105] Error trying to work out if we can handle /system.slice/run-2106.scope: /system.slice/run-2106.scope not handledby systemd handler
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.826295 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/run-2106.scope"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.826304 27751 factory.go:112] Using factory "raw" for container "/system.slice/run-2106.scope"
Nov 15 02:17:01 af867b kubelet[27751]: W1115 02:17:01.826410 27751 container.go:354] Failed to create summary reader for "/system.slice/run-2106.scope": none of the resources are being tracked.
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.826424 27751 manager.go:932] Added container: "/system.slice/run-2106.scope" (aliases: [], namespace: "")
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.826456 27751 handler.go:325] Added event &{/system.slice/run-2106.scope 0001-01-01 00:00:00 +0000 UTC containerCreation {<nil>}}
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.826475 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-kubelet-pods-10af02e8\\x2dc9ab\\x2d11e7\\2d89f4\\x2dc6b053eac242-volumes-kubernetes.io\\x7esecret-default\\x2dtoken\\x2dqjbsf.mount"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.826487 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-kubelet-pods-10af02e8\\x2dc9ab\\x2d11e7\\x2d89f4\\2dc6b053eac242-volumes-kubernetes.io\\x7esecret-default\\x2dtoken\\x2dqjbsf.mount", but ignoring.
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.826500 27751 manager.go:901] ignoring container "/system.slice/var-lib-kubelet-pods-10af02e8\\x2dc9ab\\x2d11e7\\x2d89f4\\x2dc6b053eac242-volues-kubernetes.io\\x7esecret-default\\x2dtoken\\x2dqjbsf.mount"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.826517 27751 manager.go:989] Destroyed container: "/system.slice/run-2106.scope" (aliases: [], namespace: "")
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.826525 27751 handler.go:325] Added event &{/system.slice/run-2106.scope 2017-11-15 02:17:01.826522316 +0000 UTC containerDeletion {<nil>}}
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.826554 27751 container.go:409] Start housekeeping for container "/system.slice/run-2106.scope"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.826709 27751 config.go:282] Setting pods for source api
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.845433 27751 reconciler.go:212] operationExecutor.VerifyControllerAttachedVolume started for volume "default-token-qjbsf" (UniqueName: "kuberetes.io/secret/10c857c2-c9ab-11e7-89f4-c6b053eac242-default-token-qjbsf") pod "snginx-11" (UID: "10c857c2-c9ab-11e7-89f4-c6b053eac242")
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.845771 27751 kubelet.go:1850] SyncLoop (RECONCILE, "api"): "snginx-4_default(10af02e8-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.853670 27751 generic.go:146] GenericPLEG: 0eb57d14-c9ab-11e7-89f4-c6b053eac242/0fab8639e261263f063ff03cc306506a34283f24ab70834abda8a699113657c: non-existent -> running
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.853689 27751 generic.go:146] GenericPLEG: 0e4740f5-c9ab-11e7-89f4-c6b053eac242/66b9c434d36eb927e5ef517d07f9a34fc5e788400450810f5065e3e2ce1a0d7: non-existent -> running
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.855863 27751 kuberuntime_manager.go:833] getSandboxIDByPodUID got sandbox IDs ["0fab8639e261263f063ff03cc306506a34283f24ab70834abda8a69911365cc"] for pod "snginx-28_default(0eb57d14-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.931909 27751 generic.go:345] PLEG: Write status for snginx-28/default: &container.PodStatus{ID:"0eb57d14-c9ab-11e7-89f4-c6b053eac242", Name:"nginx-28", Namespace:"default", IP:"10.32.0.5", ContainerStatuses:[]*container.ContainerStatus{}, SandboxStatuses:[]*runtime.PodSandboxStatus{(*runtime.PodSandboxStatus)(0xc421be4000)}} (err: <nil>
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.932893 27751 kubelet.go:1871] SyncLoop (PLEG): "snginx-28_default(0eb57d14-c9ab-11e7-89f4-c6b053eac242)", event: &pleg.PodLifecycleEvent{ID:"eb57d14-c9ab-11e7-89f4-c6b053eac242", Type:"ContainerStarted", Data:"0fab8639e261263f063ff03cc306506a34283f24ab70834abda8a699113657cc"}
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.934864 27751 kuberuntime_manager.go:833] getSandboxIDByPodUID got sandbox IDs ["66b9c434d36eb927e5ef517d07f9a34fc5e788400450810f5065e3e2ce1a007"] for pod "snginx-16_default(0e4740f5-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.946525 27751 reconciler.go:257] operationExecutor.MountVolume started for volume "default-token-qjbsf" (UniqueName: "kubernetes.io/secret/10c57c2-c9ab-11e7-89f4-c6b053eac242-default-token-qjbsf") pod "snginx-11" (UID: "10c857c2-c9ab-11e7-89f4-c6b053eac242")
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.969078 27751 secret.go:186] Setting up volume default-token-qjbsf for pod 10c857c2-c9ab-11e7-89f4-c6b053eac242 at /var/lib/kubelet/pods/10c85c2-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.971971 27751 empty_dir.go:264] pod 10c857c2-c9ab-11e7-89f4-c6b053eac242: mounting tmpfs for volume wrapped_default-token-qjbsf
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.972009 27751 mount_linux.go:135] Mounting cmd (systemd-run) with arguments ([--description=Kubernetes transient mount for /var/lib/kubelet/pos/10c857c2-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf --scope -- mount -t tmpfs tmpfs /var/lib/kubelet/pods/10c857c2-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetesio~secret/default-token-qjbsf])
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.983179 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/run-2123.scope"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.983270 27751 factory.go:105] Error trying to work out if we can handle /system.slice/run-2123.scope: /system.slice/run-2123.scope not handledby systemd handler
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.983278 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/run-2123.scope"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.983286 27751 factory.go:112] Using factory "raw" for container "/system.slice/run-2123.scope"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.983540 27751 manager.go:932] Added container: "/system.slice/run-2123.scope" (aliases: [], namespace: "")
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.990264 27751 handler.go:325] Added event &{/system.slice/run-2123.scope 2017-11-15 02:17:01.980194711 +0000 UTC containerCreation {<nil>}}
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.990314 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-kubelet-pods-10b34ac6\\x2dc9ab\\x2d11e7\\2d89f4\\x2dc6b053eac242-volumes-kubernetes.io\\x7esecret-default\\x2dtoken\\x2dqjbsf.mount"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.990333 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-kubelet-pods-10b34ac6\\x2dc9ab\\x2d11e7\\x2d89f4\\2dc6b053eac242-volumes-kubernetes.io\\x7esecret-default\\x2dtoken\\x2dqjbsf.mount", but ignoring.
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.990345 27751 manager.go:901] ignoring container "/system.slice/var-lib-kubelet-pods-10b34ac6\\x2dc9ab\\x2d11e7\\x2d89f4\\x2dc6b053eac242-volues-kubernetes.io\\x7esecret-default\\x2dtoken\\x2dqjbsf.mount"
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.990371 27751 container.go:409] Start housekeeping for container "/system.slice/run-2123.scope"
Nov 15 02:17:01 af867b kubelet[27751]: W1115 02:17:01.990621 27751 container.go:367] Failed to get RecentStats("/system.slice/run-2123.scope") while determining the next housekeeping: unable to fnd data for container /system.slice/run-2123.scope
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.991062 27751 manager.go:989] Destroyed container: "/system.slice/run-2123.scope" (aliases: [], namespace: "")
Nov 15 02:17:01 af867b kubelet[27751]: I1115 02:17:01.991092 27751 handler.go:325] Added event &{/system.slice/run-2123.scope 2017-11-15 02:17:01.991084523 +0000 UTC containerDeletion {<nil>}}
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.004208 27751 generic.go:345] PLEG: Write status for snginx-16/default: &container.PodStatus{ID:"0e4740f5-c9ab-11e7-89f4-c6b053eac242", Name:"nginx-16", Namespace:"default", IP:"10.32.0.6", ContainerStatuses:[]*container.ContainerStatus{}, SandboxStatuses:[]*runtime.PodSandboxStatus{(*runtime.PodSandboxStatus)(0xc420d76000)}} (err: <nil>
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.004264 27751 kubelet.go:1871] SyncLoop (PLEG): "snginx-16_default(0e4740f5-c9ab-11e7-89f4-c6b053eac242)", event: &pleg.PodLifecycleEvent{ID:"e4740f5-c9ab-11e7-89f4-c6b053eac242", Type:"ContainerStarted", Data:"66b9c434d36eb927e5ef517d07f9a34fc5e788400450810f5065e3e2ce1a0d07"}
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.017809 27751 config.go:282] Setting pods for source api
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.018836 27751 secret.go:217] Received secret default/default-token-qjbsf containing (3) pieces of data, 1878 total bytes
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.018931 27751 atomic_writer.go:145] pod default/snginx-4 volume default-token-qjbsf: write required for target directory /var/lib/kubelet/pods10af02e8-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.019035 27751 atomic_writer.go:160] pod default/snginx-4 volume default-token-qjbsf: performed write of new data to ts data directory: /var/li/kubelet/pods/10af02e8-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf/..119811_15_11_02_17_02.062005300
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.019140 27751 operation_generator.go:484] MountVolume.SetUp succeeded for volume "default-token-qjbsf" (UniqueName: "kubernetes.io/secret/10af2e8-c9ab-11e7-89f4-c6b053eac242-default-token-qjbsf") pod "snginx-4" (UID: "10af02e8-c9ab-11e7-89f4-c6b053eac242")
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.019422 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"default", Name:"snginx-4", UID:"10af02e8-c9ab-11e7-89f4-c6b053eac42", APIVersion:"v1", ResourceVersion:"1345", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "default-token-qjbsf"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.066879 27751 config.go:404] Receiving a new pod "snginx-33_default(10d84c3d-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.070139 27751 kubelet.go:1837] SyncLoop (ADD, "api"): "snginx-33_default(10d84c3d-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.070283 27751 kubelet_pods.go:1284] Generating status for "snginx-33_default(10d84c3d-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.070556 27751 qos_container_manager_linux.go:320] [ContainerManager]: Updated QoS cgroup configuration
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.082417 27751 reconciler.go:212] operationExecutor.VerifyControllerAttachedVolume started for volume "default-token-qjbsf" (UniqueName: "kuberetes.io/secret/10d84c3d-c9ab-11e7-89f4-c6b053eac242-default-token-qjbsf") pod "snginx-33" (UID: "10d84c3d-c9ab-11e7-89f4-c6b053eac242")
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.082526 27751 factory.go:116] Factory "docker" was unable to handle container "/kubepods/besteffort/pod10d84c3d-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.082541 27751 factory.go:105] Error trying to work out if we can handle /kubepods/besteffort/pod10d84c3d-c9ab-11e7-89f4-c6b053eac242: /kubepod/besteffort/pod10d84c3d-c9ab-11e7-89f4-c6b053eac242 not handled by systemd handler
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.082547 27751 factory.go:116] Factory "systemd" was unable to handle container "/kubepods/besteffort/pod10d84c3d-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.082554 27751 factory.go:112] Using factory "raw" for container "/kubepods/besteffort/pod10d84c3d-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.082698 27751 manager.go:932] Added container: "/kubepods/besteffort/pod10d84c3d-c9ab-11e7-89f4-c6b053eac242" (aliases: [], namespace: "")
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.082836 27751 handler.go:325] Added event &{/kubepods/besteffort/pod10d84c3d-c9ab-11e7-89f4-c6b053eac242 2017-11-15 02:17:02.080194711 +0000 UC containerCreation {<nil>}}
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.082866 27751 container.go:409] Start housekeeping for container "/kubepods/besteffort/pod10d84c3d-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.087609 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "snginx-33_default(10d84c3d-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.133079 27751 status_manager.go:451] Status for pod "snginx-11_default(10c857c2-c9ab-11e7-89f4-c6b053eac242)" updated successfully: (1, {PhasePending Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:17:01 +0000 UTC Reason: Message:} {Type:Ready Status:False LastProbeTie:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:17:01 +0000 UTC Reason:ContainersNotReady Message:containers with unready status: [nginx]} {Type:PodScheduled Status:True LastProbeTme:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:17:01 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP: StartTime:2017-11-15 02:17:01 +0000 UTC InitContainrStatuses:[] ContainerStatuses:[{Name:nginx State:{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,} Running:nil Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Termiated:nil} Ready:false RestartCount:0 Image:nginx ImageID: ContainerID:}] QOSClass:BestEffort})
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.133393 27751 config.go:282] Setting pods for source api
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.140131 27751 secret.go:217] Received secret default/default-token-qjbsf containing (3) pieces of data, 1878 total bytes
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.140218 27751 atomic_writer.go:145] pod default/snginx-6 volume default-token-qjbsf: write required for target directory /var/lib/kubelet/pods10b34ac6-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.140305 27751 atomic_writer.go:160] pod default/snginx-6 volume default-token-qjbsf: performed write of new data to ts data directory: /var/li/kubelet/pods/10b34ac6-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf/..119811_15_11_02_17_02.321763843
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.140396 27751 operation_generator.go:484] MountVolume.SetUp succeeded for volume "default-token-qjbsf" (UniqueName: "kubernetes.io/secret/10b3ac6-c9ab-11e7-89f4-c6b053eac242-default-token-qjbsf") pod "snginx-6" (UID: "10b34ac6-c9ab-11e7-89f4-c6b053eac242")
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.140429 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"default", Name:"snginx-6", UID:"10b34ac6-c9ab-11e7-89f4-c6b053eac42", APIVersion:"v1", ResourceVersion:"1344", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "default-token-qjbsf"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.145178 27751 kubelet.go:1850] SyncLoop (RECONCILE, "api"): "snginx-11_default(10c857c2-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.148820 27751 secret.go:217] Received secret default/default-token-qjbsf containing (3) pieces of data, 1878 total bytes
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.148879 27751 atomic_writer.go:145] pod default/snginx-11 volume default-token-qjbsf: write required for target directory /var/lib/kubelet/pod/10c857c2-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.148955 27751 atomic_writer.go:160] pod default/snginx-11 volume default-token-qjbsf: performed write of new data to ts data directory: /var/lb/kubelet/pods/10c857c2-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf/..119811_15_11_02_17_02.926896518
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.149031 27751 operation_generator.go:484] MountVolume.SetUp succeeded for volume "default-token-qjbsf" (UniqueName: "kubernetes.io/secret/10c87c2-c9ab-11e7-89f4-c6b053eac242-default-token-qjbsf") pod "snginx-11" (UID: "10c857c2-c9ab-11e7-89f4-c6b053eac242")
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.149055 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"default", Name:"snginx-11", UID:"10c857c2-c9ab-11e7-89f4-c6b053ea242", APIVersion:"v1", ResourceVersion:"1349", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "default-token-qjbsf"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.158772 27751 volume_manager.go:366] All volumes are attached and mounted for pod "snginx-6_default(10b34ac6-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.158797 27751 kuberuntime_manager.go:370] No sandbox for pod "snginx-6_default(10b34ac6-c9ab-11e7-89f4-c6b053eac242)" can be found. Need to strt a new one
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.158809 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:true CreateSandbox:true SandboxID: Attempt:0 NextInitContainerToStartnil ContainersToStart:[0] ContainersToKill:map[]} for pod "snginx-6_default(10b34ac6-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.158844 27751 kuberuntime_manager.go:565] SyncPod received new pod "snginx-6_default(10b34ac6-c9ab-11e7-89f4-c6b053eac242)", will create a sanbox for it
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.158852 27751 kuberuntime_manager.go:574] Stopping PodSandbox for "snginx-6_default(10b34ac6-c9ab-11e7-89f4-c6b053eac242)", will start new one
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.158887 27751 kuberuntime_manager.go:626] Creating sandbox for pod "snginx-6_default(10b34ac6-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.167769 27751 config.go:282] Setting pods for source api
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.169058 27751 config.go:404] Receiving a new pod "snginx-24_default(10e8f77b-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.172804 27751 kubelet.go:1837] SyncLoop (ADD, "api"): "snginx-24_default(10e8f77b-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.172949 27751 kubelet_pods.go:1284] Generating status for "snginx-24_default(10e8f77b-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.173227 27751 qos_container_manager_linux.go:320] [ContainerManager]: Updated QoS cgroup configuration
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.180726 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "snginx-24_default(10e8f77b-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.181494 27751 volume_manager.go:366] All volumes are attached and mounted for pod "snginx-4_default(10af02e8-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.181512 27751 kuberuntime_manager.go:370] No sandbox for pod "snginx-4_default(10af02e8-c9ab-11e7-89f4-c6b053eac242)" can be found. Need to strt a new one
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.181522 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:true CreateSandbox:true SandboxID: Attempt:0 NextInitContainerToStartnil ContainersToStart:[0] ContainersToKill:map[]} for pod "snginx-4_default(10af02e8-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.181555 27751 kuberuntime_manager.go:565] SyncPod received new pod "snginx-4_default(10af02e8-c9ab-11e7-89f4-c6b053eac242)", will create a sanbox for it
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.181563 27751 kuberuntime_manager.go:574] Stopping PodSandbox for "snginx-4_default(10af02e8-c9ab-11e7-89f4-c6b053eac242)", will start new one
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.181579 27751 kuberuntime_manager.go:626] Creating sandbox for pod "snginx-4_default(10af02e8-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.186806 27751 factory.go:116] Factory "docker" was unable to handle container "/kubepods/besteffort/pod10e8f77b-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.186826 27751 factory.go:105] Error trying to work out if we can handle /kubepods/besteffort/pod10e8f77b-c9ab-11e7-89f4-c6b053eac242: /kubepod/besteffort/pod10e8f77b-c9ab-11e7-89f4-c6b053eac242 not handled by systemd handler
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.186834 27751 factory.go:116] Factory "systemd" was unable to handle container "/kubepods/besteffort/pod10e8f77b-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.186854 27751 factory.go:112] Using factory "raw" for container "/kubepods/besteffort/pod10e8f77b-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.187031 27751 manager.go:932] Added container: "/kubepods/besteffort/pod10e8f77b-c9ab-11e7-89f4-c6b053eac242" (aliases: [], namespace: "")
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.187148 27751 handler.go:325] Added event &{/kubepods/besteffort/pod10e8f77b-c9ab-11e7-89f4-c6b053eac242 2017-11-15 02:17:02.180194711 +0000 UC containerCreation {<nil>}}
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.187197 27751 container.go:409] Start housekeeping for container "/kubepods/besteffort/pod10e8f77b-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.188509 27751 reconciler.go:212] operationExecutor.VerifyControllerAttachedVolume started for volume "default-token-qjbsf" (UniqueName: "kuberetes.io/secret/10e8f77b-c9ab-11e7-89f4-c6b053eac242-default-token-qjbsf") pod "snginx-24" (UID: "10e8f77b-c9ab-11e7-89f4-c6b053eac242")
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.194180 27751 reconciler.go:257] operationExecutor.MountVolume started for volume "default-token-qjbsf" (UniqueName: "kubernetes.io/secret/10d4c3d-c9ab-11e7-89f4-c6b053eac242-default-token-qjbsf") pod "snginx-33" (UID: "10d84c3d-c9ab-11e7-89f4-c6b053eac242")
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.194233 27751 secret.go:186] Setting up volume default-token-qjbsf for pod 10d84c3d-c9ab-11e7-89f4-c6b053eac242 at /var/lib/kubelet/pods/10d843d-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.194379 27751 empty_dir.go:264] pod 10d84c3d-c9ab-11e7-89f4-c6b053eac242: mounting tmpfs for volume wrapped_default-token-qjbsf
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.194397 27751 mount_linux.go:135] Mounting cmd (systemd-run) with arguments ([--description=Kubernetes transient mount for /var/lib/kubelet/pos/10d84c3d-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf --scope -- mount -t tmpfs tmpfs /var/lib/kubelet/pods/10d84c3d-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetesio~secret/default-token-qjbsf])
Nov 15 02:17:02 af867b kubelet[27751]: W1115 02:17:02.253782 27751 raw.go:87] Error while processing event ("/sys/fs/cgroup/blkio/system.slice/run-2136.scope": 0x40000100 == IN_CREATE|IN_ISDIR): notify_add_watch /sys/fs/cgroup/blkio/system.slice/run-2136.scope: no such file or directory
Nov 15 02:17:02 af867b kubelet[27751]: W1115 02:17:02.253845 27751 raw.go:87] Error while processing event ("/sys/fs/cgroup/memory/system.slice/run-2136.scope": 0x40000100 == IN_CREATE|IN_ISDIR):inotify_add_watch /sys/fs/cgroup/memory/system.slice/run-2136.scope: no such file or directory
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.257012 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/run-2136.scope"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.258110 27751 factory.go:105] Error trying to work out if we can handle /system.slice/run-2136.scope: /system.slice/run-2136.scope not handledby systemd handler
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.258727 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/run-2136.scope"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.258745 27751 factory.go:112] Using factory "raw" for container "/system.slice/run-2136.scope"
Nov 15 02:17:02 af867b kubelet[27751]: W1115 02:17:02.260229 27751 container.go:354] Failed to create summary reader for "/system.slice/run-2136.scope": none of the resources are being tracked.
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.260249 27751 manager.go:932] Added container: "/system.slice/run-2136.scope" (aliases: [], namespace: "")
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.260293 27751 handler.go:325] Added event &{/system.slice/run-2136.scope 0001-01-01 00:00:00 +0000 UTC containerCreation {<nil>}}
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.261699 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-kubelet-pods-10c857c2\\x2dc9ab\\x2d11e7\\2d89f4\\x2dc6b053eac242-volumes-kubernetes.io\\x7esecret-default\\x2dtoken\\x2dqjbsf.mount"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.261713 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-kubelet-pods-10c857c2\\x2dc9ab\\x2d11e7\\x2d89f4\\2dc6b053eac242-volumes-kubernetes.io\\x7esecret-default\\x2dtoken\\x2dqjbsf.mount", but ignoring.
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.261724 27751 manager.go:901] ignoring container "/system.slice/var-lib-kubelet-pods-10c857c2\\x2dc9ab\\x2d11e7\\x2d89f4\\x2dc6b053eac242-volues-kubernetes.io\\x7esecret-default\\x2dtoken\\x2dqjbsf.mount"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.263414 27751 manager.go:989] Destroyed container: "/system.slice/run-2136.scope" (aliases: [], namespace: "")
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.264767 27751 handler.go:325] Added event &{/system.slice/run-2136.scope 2017-11-15 02:17:02.263422817 +0000 UTC containerDeletion {<nil>}}
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.264844 27751 container.go:409] Start housekeeping for container "/system.slice/run-2136.scope"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.268922 27751 config.go:282] Setting pods for source api
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.274985 27751 config.go:404] Receiving a new pod "snginx-30_default(110c8f01-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.275353 27751 kubelet.go:1837] SyncLoop (ADD, "api"): "snginx-30_default(110c8f01-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.275498 27751 kubelet_pods.go:1284] Generating status for "snginx-30_default(110c8f01-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.277228 27751 qos_container_manager_linux.go:320] [ContainerManager]: Updated QoS cgroup configuration
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.279496 27751 factory.go:116] Factory "docker" was unable to handle container "/kubepods/besteffort/pod110c8f01-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.279515 27751 factory.go:105] Error trying to work out if we can handle /kubepods/besteffort/pod110c8f01-c9ab-11e7-89f4-c6b053eac242: /kubepod/besteffort/pod110c8f01-c9ab-11e7-89f4-c6b053eac242 not handled by systemd handler
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.279521 27751 factory.go:116] Factory "systemd" was unable to handle container "/kubepods/besteffort/pod110c8f01-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.279528 27751 factory.go:112] Using factory "raw" for container "/kubepods/besteffort/pod110c8f01-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.282086 27751 manager.go:932] Added container: "/kubepods/besteffort/pod110c8f01-c9ab-11e7-89f4-c6b053eac242" (aliases: [], namespace: "")
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.282155 27751 handler.go:325] Added event &{/kubepods/besteffort/pod110c8f01-c9ab-11e7-89f4-c6b053eac242 2017-11-15 02:17:02.278194711 +0000 UC containerCreation {<nil>}}
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.282842 27751 container.go:409] Start housekeeping for container "/kubepods/besteffort/pod110c8f01-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:02 af867b kubelet[27751]: E1115 02:17:02.282905 27751 helpers.go:468] PercpuUsage had 0 cpus, but the actual number is 2; ignoring extra CPUs
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.293849 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "snginx-30_default(110c8f01-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.294075 27751 status_manager.go:451] Status for pod "snginx-33_default(10d84c3d-c9ab-11e7-89f4-c6b053eac242)" updated successfully: (1, {PhasePending Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:17:02 +0000 UTC Reason: Message:} {Type:Ready Status:False LastProbeTie:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:17:02 +0000 UTC Reason:ContainersNotReady Message:containers with unready status: [nginx]} {Type:PodScheduled Status:True LastProbeTme:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:17:01 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP: StartTime:2017-11-15 02:17:02 +0000 UTC InitContainrStatuses:[] ContainerStatuses:[{Name:nginx State:{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,} Running:nil Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Termiated:nil} Ready:false RestartCount:0 Image:nginx ImageID: ContainerID:}] QOSClass:BestEffort})
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.297965 27751 reconciler.go:212] operationExecutor.VerifyControllerAttachedVolume started for volume "default-token-qjbsf" (UniqueName: "kuberetes.io/secret/110c8f01-c9ab-11e7-89f4-c6b053eac242-default-token-qjbsf") pod "snginx-30" (UID: "110c8f01-c9ab-11e7-89f4-c6b053eac242")
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.298058 27751 reconciler.go:257] operationExecutor.MountVolume started for volume "default-token-qjbsf" (UniqueName: "kubernetes.io/secret/10ef77b-c9ab-11e7-89f4-c6b053eac242-default-token-qjbsf") pod "snginx-24" (UID: "10e8f77b-c9ab-11e7-89f4-c6b053eac242")
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.298108 27751 secret.go:186] Setting up volume default-token-qjbsf for pod 10e8f77b-c9ab-11e7-89f4-c6b053eac242 at /var/lib/kubelet/pods/10e8f7b-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.298245 27751 empty_dir.go:264] pod 10e8f77b-c9ab-11e7-89f4-c6b053eac242: mounting tmpfs for volume wrapped_default-token-qjbsf
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.298263 27751 mount_linux.go:135] Mounting cmd (systemd-run) with arguments ([--description=Kubernetes transient mount for /var/lib/kubelet/pos/10e8f77b-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf --scope -- mount -t tmpfs tmpfs /var/lib/kubelet/pods/10e8f77b-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetesio~secret/default-token-qjbsf])
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.324295 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/run-2146.scope"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.324331 27751 factory.go:105] Error trying to work out if we can handle /system.slice/run-2146.scope: /system.slice/run-2146.scope not handledby systemd handler
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.324337 27751 factory.go:116] Factory "systemd" was unable to handle container "/system.slice/run-2146.scope"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.324345 27751 factory.go:112] Using factory "raw" for container "/system.slice/run-2146.scope"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.327697 27751 manager.go:932] Added container: "/system.slice/run-2146.scope" (aliases: [], namespace: "")
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.328651 27751 handler.go:325] Added event &{/system.slice/run-2146.scope 2017-11-15 02:17:02.322194711 +0000 UTC containerCreation {<nil>}}
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.329445 27751 container.go:409] Start housekeeping for container "/system.slice/run-2146.scope"
Nov 15 02:17:02 af867b kubelet[27751]: E1115 02:17:02.336789 27751 helpers.go:468] PercpuUsage had 0 cpus, but the actual number is 2; ignoring extra CPUs
Nov 15 02:17:02 af867b kubelet[27751]: W1115 02:17:02.339945 27751 raw.go:87] Error while processing event ("/sys/fs/cgroup/blkio/system.slice/run-2146.scope": 0x40000100 == IN_CREATE|IN_ISDIR): eaddirent: no such file or directory
Nov 15 02:17:02 af867b kubelet[27751]: W1115 02:17:02.340006 27751 raw.go:87] Error while processing event ("/sys/fs/cgroup/memory/system.slice/run-2146.scope": 0x40000100 == IN_CREATE|IN_ISDIR):inotify_add_watch /sys/fs/cgroup/memory/system.slice/run-2146.scope: no such file or directory
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.361405 27751 config.go:282] Setting pods for source api
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.364513 27751 kubelet.go:1850] SyncLoop (RECONCILE, "api"): "snginx-33_default(10d84c3d-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.364560 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-kubelet-pods-10d84c3d\\x2dc9ab\\x2d11e7\\2d89f4\\x2dc6b053eac242-volumes-kubernetes.io\\x7esecret-default\\x2dtoken\\x2dqjbsf.mount"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.364574 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-kubelet-pods-10d84c3d\\x2dc9ab\\x2d11e7\\x2d89f4\\2dc6b053eac242-volumes-kubernetes.io\\x7esecret-default\\x2dtoken\\x2dqjbsf.mount", but ignoring.
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.364586 27751 manager.go:901] ignoring container "/system.slice/var-lib-kubelet-pods-10d84c3d\\x2dc9ab\\x2d11e7\\x2d89f4\\x2dc6b053eac242-volues-kubernetes.io\\x7esecret-default\\x2dtoken\\x2dqjbsf.mount"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.366337 27751 manager.go:989] Destroyed container: "/system.slice/run-2146.scope" (aliases: [], namespace: "")
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.366356 27751 handler.go:325] Added event &{/system.slice/run-2146.scope 2017-11-15 02:17:02.366350324 +0000 UTC containerDeletion {<nil>}}
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.381096 27751 volume_manager.go:366] All volumes are attached and mounted for pod "snginx-11_default(10c857c2-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.381117 27751 kuberuntime_manager.go:370] No sandbox for pod "snginx-11_default(10c857c2-c9ab-11e7-89f4-c6b053eac242)" can be found. Need to sart a new one
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.381131 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:true CreateSandbox:true SandboxID: Attempt:0 NextInitContainerToStartnil ContainersToStart:[0] ContainersToKill:map[]} for pod "snginx-11_default(10c857c2-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.381174 27751 kuberuntime_manager.go:565] SyncPod received new pod "snginx-11_default(10c857c2-c9ab-11e7-89f4-c6b053eac242)", will create a sadbox for it
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.381186 27751 kuberuntime_manager.go:574] Stopping PodSandbox for "snginx-11_default(10c857c2-c9ab-11e7-89f4-c6b053eac242)", will start new on
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.381244 27751 kuberuntime_manager.go:626] Creating sandbox for pod "snginx-11_default(10c857c2-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.411090 27751 reconciler.go:257] operationExecutor.MountVolume started for volume "default-token-qjbsf" (UniqueName: "kubernetes.io/secret/1108f01-c9ab-11e7-89f4-c6b053eac242-default-token-qjbsf") pod "snginx-30" (UID: "110c8f01-c9ab-11e7-89f4-c6b053eac242")
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.411211 27751 secret.go:186] Setting up volume default-token-qjbsf for pod 110c8f01-c9ab-11e7-89f4-c6b053eac242 at /var/lib/kubelet/pods/110c801-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.411357 27751 empty_dir.go:264] pod 110c8f01-c9ab-11e7-89f4-c6b053eac242: mounting tmpfs for volume wrapped_default-token-qjbsf
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.411375 27751 mount_linux.go:135] Mounting cmd (systemd-run) with arguments ([--description=Kubernetes transient mount for /var/lib/kubelet/pos/110c8f01-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf --scope -- mount -t tmpfs tmpfs /var/lib/kubelet/pods/110c8f01-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetesio~secret/default-token-qjbsf])
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.432348 27751 secret.go:217] Received secret default/default-token-qjbsf containing (3) pieces of data, 1878 total bytes
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.432441 27751 atomic_writer.go:145] pod default/snginx-33 volume default-token-qjbsf: write required for target directory /var/lib/kubelet/pod/10d84c3d-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.432532 27751 atomic_writer.go:160] pod default/snginx-33 volume default-token-qjbsf: performed write of new data to ts data directory: /var/lb/kubelet/pods/10d84c3d-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf/..119811_15_11_02_17_02.827992365
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.432618 27751 operation_generator.go:484] MountVolume.SetUp succeeded for volume "default-token-qjbsf" (UniqueName: "kubernetes.io/secret/10d8c3d-c9ab-11e7-89f4-c6b053eac242-default-token-qjbsf") pod "snginx-33" (UID: "10d84c3d-c9ab-11e7-89f4-c6b053eac242")
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.432652 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"default", Name:"snginx-33", UID:"10d84c3d-c9ab-11e7-89f4-c6b053ea242", APIVersion:"v1", ResourceVersion:"1353", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "default-token-qjbsf"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.432737 27751 request.go:462] Throttling request took 108.438482ms, request: GET:https://10.241.226.117:6443/api/v1/namespaces/default/secretsdefault-token-qjbsf
Nov 15 02:17:02 af867b kubelet[27751]: W1115 02:17:02.434147 27751 raw.go:87] Error while processing event ("/sys/fs/cgroup/cpu,cpuacct/system.slice/run-2151.scope": 0x40000100 == IN_CREATE|IN_ISIR): inotify_add_watch /sys/fs/cgroup/cpu,cpuacct/system.slice/run-2151.scope: no such file or directory
Nov 15 02:17:02 af867b kubelet[27751]: W1115 02:17:02.434181 27751 raw.go:87] Error while processing event ("/sys/fs/cgroup/blkio/system.slice/run-2151.scope": 0x40000100 == IN_CREATE|IN_ISDIR): notify_add_watch /sys/fs/cgroup/blkio/system.slice/run-2151.scope: no such file or directory
Nov 15 02:17:02 af867b kubelet[27751]: W1115 02:17:02.434204 27751 raw.go:87] Error while processing event ("/sys/fs/cgroup/memory/system.slice/run-2151.scope": 0x40000100 == IN_CREATE|IN_ISDIR):inotify_add_watch /sys/fs/cgroup/memory/system.slice/run-2151.scope: no such file or directory
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.434336 27751 factory.go:116] Factory "docker" was unable to handle container "/system.slice/var-lib-kubelet-pods-10e8f77b\\x2dc9ab\\x2d11e7\\2d89f4\\x2dc6b053eac242-volumes-kubernetes.io\\x7esecret-default\\x2dtoken\\x2dqjbsf.mount"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.434353 27751 factory.go:109] Factory "systemd" can handle container "/system.slice/var-lib-kubelet-pods-10e8f77b\\x2dc9ab\\x2d11e7\\x2d89f4\\2dc6b053eac242-volumes-kubernetes.io\\x7esecret-default\\x2dtoken\\x2dqjbsf.mount", but ignoring.
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.434366 27751 manager.go:901] ignoring container "/system.slice/var-lib-kubelet-pods-10e8f77b\\x2dc9ab\\x2d11e7\\x2d89f4\\x2dc6b053eac242-volues-kubernetes.io\\x7esecret-default\\x2dtoken\\x2dqjbsf.mount"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.470230 27751 secret.go:217] Received secret default/default-token-qjbsf containing (3) pieces of data, 1878 total bytes
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.470312 27751 atomic_writer.go:145] pod default/snginx-24 volume default-token-qjbsf: write required for target directory /var/lib/kubelet/pod/10e8f77b-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.470388 27751 atomic_writer.go:160] pod default/snginx-24 volume default-token-qjbsf: performed write of new data to ts data directory: /var/lb/kubelet/pods/10e8f77b-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf/..119811_15_11_02_17_02.323676328
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.470471 27751 operation_generator.go:484] MountVolume.SetUp succeeded for volume "default-token-qjbsf" (UniqueName: "kubernetes.io/secret/10e877b-c9ab-11e7-89f4-c6b053eac242-default-token-qjbsf") pod "snginx-24" (UID: "10e8f77b-c9ab-11e7-89f4-c6b053eac242")
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.470497 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"default", Name:"snginx-24", UID:"10e8f77b-c9ab-11e7-89f4-c6b053ea242", APIVersion:"v1", ResourceVersion:"1357", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "default-token-qjbsf"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.480916 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/besteffort/pod10c857c2-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.480937 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/besteffort/pod10c857c2-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.481276 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/besteffort/pod10b34ac6-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.481288 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/besteffort/pod10b34ac6-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.481577 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/besteffort/pod10af02e8-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.481592 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/besteffort/pod10af02e8-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.482007 27751 volume_manager.go:366] All volumes are attached and mounted for pod "snginx-24_default(10e8f77b-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.482027 27751 kuberuntime_manager.go:370] No sandbox for pod "snginx-24_default(10e8f77b-c9ab-11e7-89f4-c6b053eac242)" can be found. Need to sart a new one
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.482037 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:true CreateSandbox:true SandboxID: Attempt:0 NextInitContainerToStartnil ContainersToStart:[0] ContainersToKill:map[]} for pod "snginx-24_default(10e8f77b-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.482095 27751 kuberuntime_manager.go:565] SyncPod received new pod "snginx-24_default(10e8f77b-c9ab-11e7-89f4-c6b053eac242)", will create a sadbox for it
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.482107 27751 kuberuntime_manager.go:574] Stopping PodSandbox for "snginx-24_default(10e8f77b-c9ab-11e7-89f4-c6b053eac242)", will start new on
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.482126 27751 kuberuntime_manager.go:626] Creating sandbox for pod "snginx-24_default(10e8f77b-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.488052 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/besteffort/pod10e8f77b-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.488068 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/besteffort/pod10e8f77b-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.523810 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.589583 27751 config.go:282] Setting pods for source api
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.591369 27751 config.go:404] Receiving a new pod "snginx-47_default(114df326-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.615016 27751 request.go:462] Throttling request took 180.918686ms, request: GET:https://10.241.226.117:6443/api/v1/namespaces/default/secretsdefault-token-qjbsf
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.723653 27751 volume_manager.go:366] All volumes are attached and mounted for pod "snginx-33_default(10d84c3d-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.723796 27751 kuberuntime_manager.go:370] No sandbox for pod "snginx-33_default(10d84c3d-c9ab-11e7-89f4-c6b053eac242)" can be found. Need to sart a new one
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.723818 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:true CreateSandbox:true SandboxID: Attempt:0 NextInitContainerToStartnil ContainersToStart:[0] ContainersToKill:map[]} for pod "snginx-33_default(10d84c3d-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.727736 27751 kuberuntime_manager.go:565] SyncPod received new pod "snginx-33_default(10d84c3d-c9ab-11e7-89f4-c6b053eac242)", will create a sadbox for it
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.727822 27751 kuberuntime_manager.go:574] Stopping PodSandbox for "snginx-33_default(10d84c3d-c9ab-11e7-89f4-c6b053eac242)", will start new on
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.727859 27751 kuberuntime_manager.go:626] Creating sandbox for pod "snginx-33_default(10d84c3d-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.807844 27751 secret.go:217] Received secret default/default-token-qjbsf containing (3) pieces of data, 1878 total bytes
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.811009 27751 atomic_writer.go:145] pod default/snginx-30 volume default-token-qjbsf: write required for target directory /var/lib/kubelet/pod/110c8f01-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.811125 27751 atomic_writer.go:160] pod default/snginx-30 volume default-token-qjbsf: performed write of new data to ts data directory: /var/lb/kubelet/pods/110c8f01-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf/..119811_15_11_02_17_02.735224807
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.811238 27751 operation_generator.go:484] MountVolume.SetUp succeeded for volume "default-token-qjbsf" (UniqueName: "kubernetes.io/secret/110cf01-c9ab-11e7-89f4-c6b053eac242-default-token-qjbsf") pod "snginx-30" (UID: "110c8f01-c9ab-11e7-89f4-c6b053eac242")
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.811383 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"default", Name:"snginx-30", UID:"110c8f01-c9ab-11e7-89f4-c6b053ea242", APIVersion:"v1", ResourceVersion:"1359", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "default-token-qjbsf"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.811737 27751 request.go:462] Throttling request took 338.947459ms, request: PUT:https://10.241.226.117:6443/api/v1/namespaces/default/pods/sninx-24/status
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.867181 27751 config.go:282] Setting pods for source api
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.869949 27751 config.go:404] Receiving a new pod "snginx-32_default(1164e65d-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.870197 27751 config.go:282] Setting pods for source api
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.875453 27751 status_manager.go:451] Status for pod "snginx-24_default(10e8f77b-c9ab-11e7-89f4-c6b053eac242)" updated successfully: (1, {PhasePending Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:17:02 +0000 UTC Reason: Message:} {Type:Ready Status:False LastProbeTie:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:17:02 +0000 UTC Reason:ContainersNotReady Message:containers with unready status: [nginx]} {Type:PodScheduled Status:True LastProbeTme:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:17:02 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP: StartTime:2017-11-15 02:17:02 +0000 UTC InitContainrStatuses:[] ContainerStatuses:[{Name:nginx State:{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,} Running:nil Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Termiated:nil} Ready:false RestartCount:0 Image:nginx ImageID: ContainerID:}] QOSClass:BestEffort})
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.877578 27751 config.go:404] Receiving a new pod "snginx-52_default(1167792f-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.883526 27751 config.go:282] Setting pods for source api
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.903207 27751 volume_manager.go:366] All volumes are attached and mounted for pod "snginx-30_default(110c8f01-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.903231 27751 kuberuntime_manager.go:370] No sandbox for pod "snginx-30_default(110c8f01-c9ab-11e7-89f4-c6b053eac242)" can be found. Need to sart a new one
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.903244 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:true CreateSandbox:true SandboxID: Attempt:0 NextInitContainerToStartnil ContainersToStart:[0] ContainersToKill:map[]} for pod "snginx-30_default(110c8f01-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.903290 27751 kuberuntime_manager.go:565] SyncPod received new pod "snginx-30_default(110c8f01-c9ab-11e7-89f4-c6b053eac242)", will create a sadbox for it
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.903301 27751 kuberuntime_manager.go:574] Stopping PodSandbox for "snginx-30_default(110c8f01-c9ab-11e7-89f4-c6b053eac242)", will start new on
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.903323 27751 kuberuntime_manager.go:626] Creating sandbox for pod "snginx-30_default(110c8f01-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.960136 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 2379, Path: /health
Nov 15 02:17:02 af867b kubelet[27751]: I1115 02:17:02.960226 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.014792 27751 request.go:462] Throttling request took 137.359425ms, request: GET:https://10.241.226.117:6443/api/v1/namespaces/default/pods/sninx-30
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.114103 27751 config.go:282] Setting pods for source api
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.118014 27751 config.go:404] Receiving a new pod "snginx-12_default(1188bbdb-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.183941 27751 config.go:282] Setting pods for source api
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.186254 27751 config.go:404] Receiving a new pod "snginx-7_default(119ec189-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.197799 27751 prober.go:160] HTTP-Probe Host: http://10.32.0.2, Port: 10054, Path: /healthcheck/kubedns
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.197824 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.210998 27751 config.go:282] Setting pods for source api
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.213504 27751 config.go:404] Receiving a new pod "snginx-26_default(1195f2d8-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.214211 27751 request.go:462] Throttling request took 91.34839ms, request: PUT:https://10.241.226.117:6443/api/v1/namespaces/default/pods/sngix-30/status
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.214980 27751 config.go:282] Setting pods for source api
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.221679 27751 config.go:404] Receiving a new pod "snginx-27_default(11a56a59-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.222209 27751 config.go:282] Setting pods for source api
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.226056 27751 config.go:404] Receiving a new pod "snginx-8_default(11946462-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.246288 27751 http.go:96] Probe succeeded for http://127.0.0.1:2379/health, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:1:03 GMT] Content-Length:[18] Content-Type:[text/plain; charset=utf-8]] 0xc420f64ea0 18 [] true false map[] 0xc420a30700 <nil>}
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.246336 27751 prober.go:113] Liveness probe for "etcd-af867b_kube-system(d76e26fba3bf2bfd215eb29011d55250):etcd" succeeded
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.260857 27751 http.go:96] Probe succeeded for http://10.32.0.2:10054/healthcheck/kubedns, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15Nov 2017 02:17:03 GMT] Content-Length:[51] Content-Type:[application/json]] 0xc420f64fe0 51 [] true false map[] 0xc42271d200 <nil>}
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.260914 27751 prober.go:113] Liveness probe for "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242):kubedns" succeede
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.418917 27751 status_manager.go:451] Status for pod "snginx-30_default(110c8f01-c9ab-11e7-89f4-c6b053eac242)" updated successfully: (1, {PhasePending Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:17:02 +0000 UTC Reason: Message:} {Type:Ready Status:False LastProbeTie:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:17:02 +0000 UTC Reason:ContainersNotReady Message:containers with unready status: [nginx]} {Type:PodScheduled Status:True LastProbeTme:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:17:02 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP: StartTime:2017-11-15 02:17:02 +0000 UTC InitContainrStatuses:[] ContainerStatuses:[{Name:nginx State:{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,} Running:nil Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Termiated:nil} Ready:false RestartCount:0 Image:nginx ImageID: ContainerID:}] QOSClass:BestEffort})
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.460024 27751 config.go:282] Setting pods for source api
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.464102 27751 config.go:282] Setting pods for source api
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.466632 27751 config.go:404] Receiving a new pod "snginx-51_default(11c10f74-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.507058 27751 config.go:282] Setting pods for source api
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.510292 27751 config.go:404] Receiving a new pod "snginx-80_default(11be9038-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.582043 27751 config.go:282] Setting pods for source api
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.584747 27751 config.go:404] Receiving a new pod "snginx-31_default(11dedb1e-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.589345 27751 config.go:282] Setting pods for source api
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.594357 27751 config.go:404] Receiving a new pod "snginx-34_default(11e01340-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.620825 27751 prober.go:160] HTTP-Probe Host: https://127.0.0.1, Port: 6443, Path: /healthz
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.620858 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.645075 27751 http.go:96] Probe succeeded for https://127.0.0.1:6443/healthz, Response: {200 OK 200 HTTP/2.0 2 0 map[Content-Type:[text/plain;charset=utf-8] Content-Length:[2] Date:[Wed, 15 Nov 2017 02:17:03 GMT]] 0xc42121d620 2 [] false false map[] 0xc420a31c00 0xc421b98a50}
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.645119 27751 prober.go:113] Liveness probe for "kube-apiserver-af867b_kube-system(4e0fac5dee63099d647b4d031a37ad7d):kube-apiserver" succeeded
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.657902 27751 config.go:282] Setting pods for source api
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.659829 27751 config.go:404] Receiving a new pod "snginx-57_default(11e0958a-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.708913 27751 config.go:282] Setting pods for source api
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.710643 27751 config.go:404] Receiving a new pod "snginx-50_default(11eaf831-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.754419 27751 config.go:282] Setting pods for source api
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.756893 27751 config.go:404] Receiving a new pod "snginx-19_default(11fb7e52-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.757623 27751 config.go:282] Setting pods for source api
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.760664 27751 config.go:404] Receiving a new pod "snginx-36_default(11fbf7a4-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.761607 27751 config.go:282] Setting pods for source api
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.763977 27751 config.go:404] Receiving a new pod "snginx-45_default(11fbce72-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.835400 27751 prober.go:160] HTTP-Probe Host: http://10.32.0.2, Port: 10054, Path: /healthcheck/dnsmasq
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.835437 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.837171 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/besteffort/pod110c8f01-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.837196 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/besteffort/pod110c8f01-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.839250 27751 http.go:96] Probe succeeded for http://10.32.0.2:10054/healthcheck/dnsmasq, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Type:application/json] Date:[Wed, 15 Nov 2017 02:17:03 GMT] Content-Length:[51]] 0xc421101ee0 51 [] true false map[] 0xc421178800 <nil>}
Nov 15 02:17:03 af867b kubelet[27751]: I1115 02:17:03.839301 27751 prober.go:113] Liveness probe for "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242):dnsmasq" succeede
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:03.998970 27751 config.go:282] Setting pods for source api
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.001614 27751 config.go:404] Receiving a new pod "snginx-62_default(121c97f0-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.005031 27751 factory.go:112] Using factory "docker" for container "/kubepods/besteffort/pod106570f0-c9ab-11e7-89f4-c6b053eac242/aef8399af06a53e2faec702ec96747200084cadab8b41510604cf1759a7983f"
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.009868 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/besteffort/pod10d84c3d-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.009886 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/besteffort/pod10d84c3d-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.010308 27751 docker_sandbox.go:691] Will attempt to re-write config file /var/lib/docker/containers/aef8399af06a593e2faec702ec96747200084cada8b41510604cf1759a7983f/resolv.conf with:
Nov 15 02:17:04 af867b kubelet[27751]: [nameserver 10.96.0.10 search default.svc.cluster.local svc.cluster.local cluster.local opcwlaas.oraclecloud.internal. options ndots:5]
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.010455 27751 plugins.go:392] Calling network plugin cni to set up pod "snginx-1_default"
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.068181 27751 config.go:282] Setting pods for source api
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.071465 27751 config.go:404] Receiving a new pod "snginx-22_default(1222faad-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.084970 27751 cni.go:326] Got netns path /proc/2206/ns/net
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.084983 27751 cni.go:327] Using netns path default
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.085081 27751 cni.go:298] About to add CNI network cni-loopback (type=loopback)
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.095442 27751 manager.go:932] Added container: "/kubepods/besteffort/pod106570f0-c9ab-11e7-89f4-c6b053eac242/aef8399af06a593e2faec702ec9674720084cadab8b41510604cf1759a7983f" (aliases: [k8s_POD_snginx-1_default_106570f0-c9ab-11e7-89f4-c6b053eac242_0 aef8399af06a593e2faec702ec96747200084cadab8b41510604cf1759a7983f], namespace: "docker")
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.095586 27751 handler.go:325] Added event &{/kubepods/besteffort/pod106570f0-c9ab-11e7-89f4-c6b053eac242/aef8399af06a593e2faec702ec96747200084adab8b41510604cf1759a7983f 2017-11-15 02:17:02.951194711 +0000 UTC containerCreation {<nil>}}
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.095627 27751 container.go:409] Start housekeeping for container "/kubepods/besteffort/pod106570f0-c9ab-11e7-89f4-c6b053eac242/aef8399af06a5932faec702ec96747200084cadab8b41510604cf1759a7983f"
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.116967 27751 cni.go:326] Got netns path /proc/2206/ns/net
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.116998 27751 cni.go:327] Using netns path default
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.117129 27751 cni.go:298] About to add CNI network weave (type=weave-net)
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.162447 27751 config.go:282] Setting pods for source api
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.164516 27751 config.go:404] Receiving a new pod "snginx-3_default(123d32e8-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.431330 27751 config.go:282] Setting pods for source api
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.472654 27751 config.go:404] Receiving a new pod "snginx-41_default(124c80b8-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.504600 27751 prober.go:160] HTTP-Probe Host: http://10.32.0.2, Port: 10054, Path: /metrics
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.504626 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.518452 27751 http.go:96] Probe succeeded for http://10.32.0.2:10054/metrics, Response: {200 OK 200 HTTP/1.1 1 1 map[Content-Type:[text/plain;version=0.0.4] Date:[Wed, 15 Nov 2017 02:17:04 GMT]] 0xc421638ea0 -1 [] true true map[] 0xc420c90400 <nil>}
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.518530 27751 prober.go:113] Liveness probe for "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242):sidecar" succeede
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.591026 27751 config.go:282] Setting pods for source api
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.594301 27751 config.go:404] Receiving a new pod "snginx-58_default(12526146-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.596566 27751 kuberuntime_manager.go:640] Created PodSandbox "aef8399af06a593e2faec702ec96747200084cadab8b41510604cf1759a7983f" for pod "sngin-1_default(106570f0-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.601998 27751 config.go:282] Setting pods for source api
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.613639 27751 config.go:404] Receiving a new pod "snginx-20_default(125d0f6c-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.841184 27751 config.go:282] Setting pods for source api
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.847705 27751 config.go:404] Receiving a new pod "snginx-21_default(12634156-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.848663 27751 config.go:282] Setting pods for source api
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.851649 27751 config.go:404] Receiving a new pod "snginx-42_default(1275c2ca-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.853417 27751 config.go:282] Setting pods for source api
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.882551 27751 config.go:404] Receiving a new pod "snginx-56_default(1275fae4-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.883365 27751 config.go:282] Setting pods for source api
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.894587 27751 config.go:404] Receiving a new pod "snginx-9_default(1275c6f5-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.900636 27751 config.go:282] Setting pods for source api
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.905502 27751 config.go:404] Receiving a new pod "snginx-44_default(127ec404-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.923953 27751 config.go:282] Setting pods for source api
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.958257 27751 config.go:404] Receiving a new pod "snginx-53_default(12867215-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:04 af867b kubelet[27751]: I1115 02:17:04.994539 27751 config.go:282] Setting pods for source api
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.008229 27751 config.go:404] Receiving a new pod "snginx-2_default(127ed64d-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.231553 27751 config.go:282] Setting pods for source api
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.235551 27751 config.go:404] Receiving a new pod "snginx-48_default(12b81849-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.235809 27751 config.go:282] Setting pods for source api
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.239541 27751 config.go:404] Receiving a new pod "snginx-61_default(12b2d816-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.240148 27751 config.go:282] Setting pods for source api
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.243432 27751 config.go:404] Receiving a new pod "snginx-29_default(12b0c1b4-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.245678 27751 config.go:282] Setting pods for source api
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.254601 27751 config.go:404] Receiving a new pod "snginx-40_default(12b094bf-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.257408 27751 config.go:282] Setting pods for source api
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.260393 27751 config.go:404] Receiving a new pod "snginx-5_default(12b8bcf8-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.364895 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 6784, Path: /status
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.364939 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.377892 27751 http.go:96] Probe succeeded for http://127.0.0.1:6784/status, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 02:1:05 GMT] Content-Length:[445] Content-Type:[text/plain; charset=utf-8]] 0xc421398a40 445 [] true false map[] 0xc4200ddb00 <nil>}
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.377949 27751 prober.go:113] Liveness probe for "weave-net-rg7fn_kube-system(b77b0858-c9a8-11e7-89f4-c6b053eac242):weave" succeeded
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.382026 27751 config.go:282] Setting pods for source api
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.387229 27751 config.go:404] Receiving a new pod "snginx-88_default(12e591cf-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.388303 27751 config.go:282] Setting pods for source api
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.390688 27751 config.go:404] Receiving a new pod "snginx-66_default(12d0b0e9-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.392464 27751 config.go:282] Setting pods for source api
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.399026 27751 config.go:404] Receiving a new pod "snginx-23_default(12ce91d1-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.399128 27751 config.go:282] Setting pods for source api
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.402667 27751 config.go:404] Receiving a new pod "snginx-38_default(12e1b910-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.413546 27751 config.go:282] Setting pods for source api
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.417415 27751 config.go:404] Receiving a new pod "snginx-70_default(12df4528-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.536955 27751 config.go:282] Setting pods for source api
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.548055 27751 config.go:404] Receiving a new pod "snginx-77_default(12fcc791-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.549460 27751 config.go:282] Setting pods for source api
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.572245 27751 config.go:404] Receiving a new pod "snginx-95_default(13008e85-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.581520 27751 config.go:282] Setting pods for source api
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.591515 27751 docker_sandbox.go:691] Will attempt to re-write config file /var/lib/docker/containers/e96d5281f1655945493aeaeeae76261f1190dfd412f7d754f19ce7b72309658/resolv.conf with:
Nov 15 02:17:05 af867b kubelet[27751]: [nameserver 10.96.0.10 search default.svc.cluster.local svc.cluster.local cluster.local opcwlaas.oraclecloud.internal. options ndots:5]
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.591653 27751 plugins.go:392] Calling network plugin cni to set up pod "snginx-11_default"
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.599403 27751 factory.go:112] Using factory "docker" for container "/kubepods/besteffort/pod10c857c2-c9ab-11e7-89f4-c6b053eac242/e96d5281f165545493aeaeeae76261f1190dfd41d2f7d754f19ce7b72309658"
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.600275 27751 config.go:404] Receiving a new pod "snginx-63_default(130cfe94-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.600445 27751 config.go:282] Setting pods for source api
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.611839 27751 config.go:404] Receiving a new pod "snginx-69_default(1315f5b8-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.612467 27751 config.go:282] Setting pods for source api
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.631771 27751 config.go:404] Receiving a new pod "snginx-54_default(131477f7-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.796973 27751 config.go:282] Setting pods for source api
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.802027 27751 config.go:404] Receiving a new pod "snginx-13_default(133bae95-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.929075 27751 kubelet.go:1837] SyncLoop (ADD, "api"): "snginx-47_default(114df326-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.929194 27751 kubelet.go:1913] SyncLoop (housekeeping)
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.929616 27751 generic.go:146] GenericPLEG: 106570f0-c9ab-11e7-89f4-c6b053eac242/aef8399af06a593e2faec702ec96747200084cadab8b41510604cf1759a798f: non-existent -> running
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.930476 27751 kubelet_pods.go:1284] Generating status for "snginx-47_default(114df326-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.930837 27751 qos_container_manager_linux.go:320] [ContainerManager]: Updated QoS cgroup configuration
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.949769 27751 volume_manager.go:337] Waiting for volumes to attach and mount for pod "snginx-47_default(114df326-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.974942 27751 config.go:282] Setting pods for source api
Nov 15 02:17:05 af867b kubelet[27751]: I1115 02:17:05.982327 27751 config.go:404] Receiving a new pod "snginx-71_default(13482d80-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.023985 27751 reconciler.go:212] operationExecutor.VerifyControllerAttachedVolume started for volume "default-token-qjbsf" (UniqueName: "kuberetes.io/secret/114df326-c9ab-11e7-89f4-c6b053eac242-default-token-qjbsf") pod "snginx-47" (UID: "114df326-c9ab-11e7-89f4-c6b053eac242")
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.034822 27751 config.go:282] Setting pods for source api
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.039675 27751 config.go:404] Receiving a new pod "snginx-14_default(134dfefb-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.061925 27751 config.go:282] Setting pods for source api
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.062935 27751 status_manager.go:451] Status for pod "snginx-47_default(114df326-c9ab-11e7-89f4-c6b053eac242)" updated successfully: (1, {PhasePending Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:17:05 +0000 UTC Reason: Message:} {Type:Ready Status:False LastProbeTie:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:17:05 +0000 UTC Reason:ContainersNotReady Message:containers with unready status: [nginx]} {Type:PodScheduled Status:True LastProbeTme:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2017-11-15 02:17:02 +0000 UTC Reason: Message:}] Message: Reason: HostIP:10.196.65.210 PodIP: StartTime:2017-11-15 02:17:05 +0000 UTC InitContainrStatuses:[] ContainerStatuses:[{Name:nginx State:{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,} Running:nil Terminated:nil} LastTerminationState:{Waiting:nil Running:nil Termiated:nil} Ready:false RestartCount:0 Image:nginx ImageID: ContainerID:}] QOSClass:BestEffort})
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.068545 27751 config.go:404] Receiving a new pod "snginx-97_default(135fa4d9-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.071311 27751 config.go:282] Setting pods for source api
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.124549 27751 reconciler.go:257] operationExecutor.MountVolume started for volume "default-token-qjbsf" (UniqueName: "kubernetes.io/secret/114f326-c9ab-11e7-89f4-c6b053eac242-default-token-qjbsf") pod "snginx-47" (UID: "114df326-c9ab-11e7-89f4-c6b053eac242")
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.124631 27751 secret.go:186] Setting up volume default-token-qjbsf for pod 114df326-c9ab-11e7-89f4-c6b053eac242 at /var/lib/kubelet/pods/114df26-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.124828 27751 empty_dir.go:264] pod 114df326-c9ab-11e7-89f4-c6b053eac242: mounting tmpfs for volume wrapped_default-token-qjbsf
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.124850 27751 mount_linux.go:135] Mounting cmd (systemd-run) with arguments ([--description=Kubernetes transient mount for /var/lib/kubelet/pos/114df326-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf --scope -- mount -t tmpfs tmpfs /var/lib/kubelet/pods/114df326-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetesio~secret/default-token-qjbsf])
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.228200 27751 secret.go:217] Received secret default/default-token-qjbsf containing (3) pieces of data, 1878 total bytes
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.228307 27751 atomic_writer.go:145] pod default/snginx-47 volume default-token-qjbsf: write required for target directory /var/lib/kubelet/pod/114df326-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.228397 27751 atomic_writer.go:160] pod default/snginx-47 volume default-token-qjbsf: performed write of new data to ts data directory: /var/lb/kubelet/pods/114df326-c9ab-11e7-89f4-c6b053eac242/volumes/kubernetes.io~secret/default-token-qjbsf/..119811_15_11_02_17_06.068422170
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.228483 27751 operation_generator.go:484] MountVolume.SetUp succeeded for volume "default-token-qjbsf" (UniqueName: "kubernetes.io/secret/114d326-c9ab-11e7-89f4-c6b053eac242-default-token-qjbsf") pod "snginx-47" (UID: "114df326-c9ab-11e7-89f4-c6b053eac242")
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.228799 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"default", Name:"snginx-47", UID:"114df326-c9ab-11e7-89f4-c6b053ea242", APIVersion:"v1", ResourceVersion:"1367", FieldPath:""}): type: 'Normal' reason: 'SuccessfulMountVolume' MountVolume.SetUp succeeded for volume "default-token-qjbsf"
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.258550 27751 volume_manager.go:366] All volumes are attached and mounted for pod "snginx-47_default(114df326-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.258594 27751 kuberuntime_manager.go:370] No sandbox for pod "snginx-47_default(114df326-c9ab-11e7-89f4-c6b053eac242)" can be found. Need to sart a new one
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.258608 27751 kuberuntime_manager.go:556] computePodActions got {KillPod:true CreateSandbox:true SandboxID: Attempt:0 NextInitContainerToStartnil ContainersToStart:[0] ContainersToKill:map[]} for pod "snginx-47_default(114df326-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.258664 27751 kuberuntime_manager.go:565] SyncPod received new pod "snginx-47_default(114df326-c9ab-11e7-89f4-c6b053eac242)", will create a sadbox for it
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.258673 27751 kuberuntime_manager.go:574] Stopping PodSandbox for "snginx-47_default(114df326-c9ab-11e7-89f4-c6b053eac242)", will start new on
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.258693 27751 kuberuntime_manager.go:626] Creating sandbox for pod "snginx-47_default(114df326-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.292018 27751 kubelet.go:2092] Container runtime status: Runtime Conditions: RuntimeReady=true reason: message:, NetworkReady=true reason: mesage:
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.420750 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/besteffort/pod114df326-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.420782 27751 docker_service.go:407] Setting cgroup parent to: "/kubepods/besteffort/pod114df326-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.421369 27751 cni.go:326] Got netns path /proc/2335/ns/net
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.421378 27751 cni.go:327] Using netns path default
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.421490 27751 cni.go:298] About to add CNI network cni-loopback (type=loopback)
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.470939 27751 manager.go:932] Added container: "/kubepods/besteffort/pod10c857c2-c9ab-11e7-89f4-c6b053eac242/e96d5281f1655945493aeaeeae76261f190dfd41d2f7d754f19ce7b72309658" (aliases: [k8s_POD_snginx-11_default_10c857c2-c9ab-11e7-89f4-c6b053eac242_0 e96d5281f1655945493aeaeeae76261f1190dfd41d2f7d754f19ce7b72309658], namespace: "docker")
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.475580 27751 handler.go:325] Added event &{/kubepods/besteffort/pod10c857c2-c9ab-11e7-89f4-c6b053eac242/e96d5281f1655945493aeaeeae76261f1190dd41d2f7d754f19ce7b72309658 2017-11-15 02:17:04.153194711 +0000 UTC containerCreation {<nil>}}
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.475640 27751 factory.go:116] Factory "docker" was unable to handle container "/kubepods/besteffort/pod114df326-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.475659 27751 factory.go:105] Error trying to work out if we can handle /kubepods/besteffort/pod114df326-c9ab-11e7-89f4-c6b053eac242: /kubepod/besteffort/pod114df326-c9ab-11e7-89f4-c6b053eac242 not handled by systemd handler
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.475667 27751 factory.go:116] Factory "systemd" was unable to handle container "/kubepods/besteffort/pod114df326-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.475677 27751 factory.go:112] Using factory "raw" for container "/kubepods/besteffort/pod114df326-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.480816 27751 container.go:409] Start housekeeping for container "/kubepods/besteffort/pod10c857c2-c9ab-11e7-89f4-c6b053eac242/e96d5281f165594493aeaeeae76261f1190dfd41d2f7d754f19ce7b72309658"
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.481736 27751 cni.go:326] Got netns path /proc/2335/ns/net
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.481754 27751 cni.go:327] Using netns path default
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.482135 27751 cni.go:298] About to add CNI network weave (type=weave-net)
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.529495 27751 kuberuntime_manager.go:654] Determined the ip "10.32.0.7" for pod "snginx-1_default(106570f0-c9ab-11e7-89f4-c6b053eac242)" aftersandbox changed
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.545901 27751 kuberuntime_manager.go:705] Creating container &Container{Name:nginx,Image:nginx,Command:[],Args:[],WorkingDir:,Ports:[{ 0 80 TC }],Env:[],Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[{default-token-qjbsf true /var/run/secrets/kubernetes.io/serviceaccount <nil>}],LivenessProb:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[],TerminationMessagePolic:File,} in pod snginx-1_default(106570f0-c9ab-11e7-89f4-c6b053eac242)
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.620780 27751 eviction_manager.go:221] eviction manager: synchronize housekeeping
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.622049 27751 manager.go:932] Added container: "/kubepods/besteffort/pod114df326-c9ab-11e7-89f4-c6b053eac242" (aliases: [], namespace: "")
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.626649 27751 handler.go:325] Added event &{/kubepods/besteffort/pod114df326-c9ab-11e7-89f4-c6b053eac242 2017-11-15 02:17:05.945194711 +0000 UC containerCreation {<nil>}}
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.664752 27751 server.go:227] Event(v1.ObjectReference{Kind:"Pod", Namespace:"default", Name:"snginx-1", UID:"106570f0-c9ab-11e7-89f4-c6b053eac42", APIVersion:"v1", ResourceVersion:"1337", FieldPath:"spec.containers{nginx}"}): type: 'Normal' reason: 'Pulling' pulling image "nginx"
Nov 15 02:17:06 af867b kubelet[27751]: I1115 02:17:06.665245 27751 container.go:409] Start housekeeping for container "/kubepods/besteffort/pod114df326-c9ab-11e7-89f4-c6b053eac242"
Nov 15 02:17:07 af867b kubelet[27751]: I1115 02:17:07.016699 27751 kuberuntime_manager.go:640] Created PodSandbox "e96d5281f1655945493aeaeeae76261f1190dfd41d2f7d754f19ce7b72309658" for pod "sngin-11_default(10c857c2-c9ab-11e7-89f4-c6b053eac242)"
Nov 15 02:17:07 af867b kubelet[27751]: I1115 02:17:07.600828 27751 prober.go:160] HTTP-Probe Host: http://10.32.0.2, Port: 8081, Path: /readiness
Nov 15 02:17:07 af867b kubelet[27751]: I1115 02:17:07.600869 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:17:07 af867b kubelet[27751]: I1115 02:17:07.609885 27751 http.go:96] Probe succeeded for http://10.32.0.2:8081/readiness, Response: {200 OK 200 HTTP/1.1 1 1 map[Date:[Wed, 15 Nov 2017 0:17:07 GMT] Content-Length:[3] Content-Type:[text/plain; charset=utf-8]] 0xc421a252a0 3 [] true false map[] 0xc420c90400 <nil>}
Nov 15 02:17:07 af867b kubelet[27751]: I1115 02:17:07.609945 27751 prober.go:113] Readiness probe for "kube-dns-545bc4bfd4-zvfqd_kube-system(97270c63-c9a8-11e7-89f4-c6b053eac242):kubedns" succeedd
Nov 15 02:17:07 af867b kubelet[27751]: I1115 02:17:07.682899 27751 prober.go:160] HTTP-Probe Host: http://127.0.0.1, Port: 10251, Path: /healthz
Nov 15 02:17:07 af867b kubelet[27751]: I1115 02:17:07.682944 27751 prober.go:163] HTTP-Probe Headers: map[]
Nov 15 02:17:07 af867b kubelet[27751]: I1115 02:17:07.703933 27751 ht
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment