Skip to content

Instantly share code, notes, and snippets.

@jseguillon
Last active July 11, 2022 18:53
Show Gist options
  • Save jseguillon/e1cc4f500ac81f2b40fc7528407cf451 to your computer and use it in GitHub Desktop.
Save jseguillon/e1cc4f500ac81f2b40fc7528407cf451 to your computer and use it in GitHub Desktop.
last ok molecule-kubevirt action
Starting virtctl console
Script started, file is typescript
Successfully connected to instance console. The escape sequence is ^]
[ 0.000000] Linux version 5.6.6-300.fc32.x86_64 (mockbuild@bkernel03.phx2.fedoraproject.org) (gcc version 10.0.1 20200328 (Red Hat 10.0.1-0.11) (GCC)) #1 SMP Tue Apr 21 13:44:19 UTC 2020
[ 0.000000] Command line: BOOT_IMAGE=(hd0,msdos1)/boot/vmlinuz-5.6.6-300.fc32.x86_64 root=UUID=d1b37ed4-3bbb-40b2-a6ba-f377f0c90217 ro no_timer_check net.ifnames=0 console=tty1 console=ttyS0,115200n8
[ 0.000000] x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers'
[ 0.000000] x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers'
[ 0.000000] x86/fpu: Supporting XSAVE feature 0x008: 'MPX bounds registers'
[ 0.000000] x86/fpu: Supporting XSAVE feature 0x010: 'MPX CSR'
[ 0.000000] x86/fpu: Supporting XSAVE feature 0x200: 'Protection Keys User registers'
[ 0.000000] x86/fpu: xstate_offset[3]: 960, xstate_sizes[3]: 64
[ 0.000000] x86/fpu: xstate_offset[4]: 1024, xstate_sizes[4]: 64
[ 0.000000] x86/fpu: xstate_offset[9]: 2688, xstate_sizes[9]: 8
[ 0.000000] x86/fpu: Enabled xstate features 0x21b, context size is 2696 bytes, using 'standard' format.
[ 0.000000] BIOS-provided physical RAM map:
[ 0.000000] BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable
[ 0.000000] BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved
[ 0.000000] BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved
[ 0.000000] BIOS-e820: [mem 0x0000000000100000-0x000000007ffdcfff] usable
[ 0.000000] BIOS-e820: [mem 0x000000007ffdd000-0x000000007fffffff] reserved
[ 0.000000] BIOS-e820: [mem 0x00000000b0000000-0x00000000bfffffff] reserved
[ 0.000000] BIOS-e820: [mem 0x00000000fed1c000-0x00000000fed1ffff] reserved
[ 0.000000] BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved
[ 0.000000] NX (Execute Disable) protection: active
[ 0.000000] SMBIOS 2.8 present.
[ 0.000000] DMI: KubeVirt None/RHEL-AV, BIOS 1.14.0-1.el8s 04/01/2014
[ 0.000000] last_pfn = 0x7ffdd max_arch_pfn = 0x10000000000
[ 0.000000] x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT
[ 0.000000] found SMP MP-table at [mem 0x000f5c20-0x000f5c2f]
[ 0.000000] Using GB pages for direct mapping
[ 0.000000] RAMDISK: [mem 0x34f54000-0x367a1fff]
[ 0.000000] ACPI: Early table checksum verification disabled
[ 0.000000] ACPI: RSDP 0x00000000000F59E0 000014 (v00 BOCHS )
[ 0.000000] ACPI: RSDT 0x000000007FFE1FF9 000034 (v01 BOCHS BXPC 00000001 BXPC 00000001)
[ 0.000000] ACPI: FACP 0x000000007FFE1E29 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001)
[ 0.000000] ACPI: DSDT 0x000000007FFE0040 001DE9 (v01 BOCHS BXPC 00000001 BXPC 00000001)
[ 0.000000] ACPI: FACS 0x000000007FFE0000 000040
[ 0.000000] ACPI: APIC 0x000000007FFE1F1D 000078 (v01 BOCHS BXPC 00000001 BXPC 00000001)
[ 0.000000] ACPI: MCFG 0x000000007FFE1F95 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001)
[ 0.000000] ACPI: WAET 0x000000007FFE1FD1 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001)
[ 0.000000] No NUMA configuration found
[ 0.000000] Faking a node at [mem 0x0000000000000000-0x000000007ffdcfff]
[ 0.000000] NODE_DATA(0) allocated [mem 0x7ffb2000-0x7ffdcfff]
[ 0.000000] Zone ranges:
[ 0.000000] DMA [mem 0x0000000000001000-0x0000000000ffffff]
[ 0.000000] DMA32 [mem 0x0000000001000000-0x000000007ffdcfff]
[ 0.000000] Normal empty
[ 0.000000] Device empty
[ 0.000000] Movable zone start for each node
[ 0.000000] Early memory node ranges
[ 0.000000] node 0: [mem 0x0000000000001000-0x000000000009efff]
[ 0.000000] node 0: [mem 0x0000000000100000-0x000000007ffdcfff]
[ 0.000000] Zeroed struct page in unavailable ranges: 133 pag
[ 0.000000] Initmem setup node 0 [mem 0x0000000000001000-0x000000007ffdcfff]
[ 0.000000] ACPI: PM-Timer IO Port: 0x608
[ 0.000000] ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1])
[ 0.000000] IOAPIC[0]: apic_id 0, version 32, address 0xfec00000, GSI 0-23
[ 0.000000] ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl)
[ 0.000000] ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level)
[ 0.000000] ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level)
[ 0.000000] ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level)
[ 0.000000] ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level)
[ 0.000000]ng ACPI (MADT) for SMP configuration information
[ 0.000000] smpboot: Allowing 1 CPUs, 0 hotplug CPUs
[ 0.000000] PM: hibernation: Registered nosave memory: [mem 0x00000000-0x00000fff]
[ 0.000000] PM: hibernation: Registered nosave memory: [mem 0x0009f000-0x0009ffff 0.000000] PM: hibernation: Registered nosave memory: [mem 0x000a0000-0x000effff]
[ 0.000000] PM: hibernation: Registered nosave memory: [mem 0x000f0000-0x000fffff]
[ 0.000000] [mem 0xc0000000-0xfed1bfff] available for PCI devices
[ 0.000000] Booting paravirtualized kernel on bare hardware
[ 0.000000] clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns
[ 0.000000] setup_percpu: NR_CPUS:8192 nr_cpumask_bits:1 nr_cpu_ids:1 nr_node_ids:1
[ 0.000000] percpu: Embedded 55 pages/cpu s188416 r8192 d28672 u2097152
[ 0.000000] Built 1 zonelists, mobility grouping on. Total pages: 515942
[ 0.000000] Policy zone: DMA32
[ 0.000000] Kernel command line: BOOT_IMAGE=(hd0,msdos1)/boot/vmlinuz-5.6.6-300.fc32.x86_64 root=UUID=d1b37ed4-3bbb-40b2-a6ba-f377f0c90217 ro no_k net.ifnames=0 console=tty1 console=ttyS0,115200n8
[ 0.000000] Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear)
[ 0.000000] Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear)
[ 0.000000] mem auto-init: stack:off, heap alloc:off, heap free:off
[ 0.000000] Memory: 1993340K/2096620K available (14339K kernel code, 2400K rwdata, 4868K rodata, 2452K init, 6136K bss, 103280K reserved, 0K cma-reserved)
[ 0.000000] random: get_random_u64 called from __kmem_cache_create+0x3e/0x620 with crng_init=0
[ 0.000000] SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=1, Nodes=1
[ 0.000000] ftrace: allocating 40698 entries in 159 pages
[ 0.000000] ftrace: allocated 159 pages with 6 groups
[ 0.000000] rcu: Hierarchical RCU implementation.
[ 0.000000] rcu: RCU restricting CPUs from NR_CPUS=8192 to nr_cpu_ids=1.
[ 0.000000] Tasks RCU enabled.
[ .000000] rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies.
[ 0.000000] rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=1
[ 0.000000] NR_IRQS: 524544, nr_irqs: 256, preallocated irqs: 16
[ 0.000000] random: crng done (trusting CPU's manufacturer)
[ 0.000000] Console: colour VGA+ 80x25
[ 0.000000] printk: console [tty1] enabled
[ 0.000000] printk: console [ttyS0] enabled
[ 0.000000] ACPI: Core revision 20200110
[ 0.003000] APIC: Switch to symmetric I/O mode setup
[ 0.008000] ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1
[ 0.012000] tsc: Unable to calibrate against PIT
[ 0.013000] tsc: using PMTIMER reference calibration
[ 0.013000] tsc: Detected 2095.032 MHz processor
[ 0.001524] clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x1e32db87ee8, max_idle_ns: 440795252272 ns
[ 0.002876] Calibrating delay loop (skipped), value calculated using timer frequency.. 4190.06 BogoMIPS (lpj=2095032)
[ 0.004142] pid_max: default: 32768 minimum: 301
[ 0.006042] LSM: Security Framework initializing
[ 0.007905] Yama: becoming mindful.
[ 0.009428] SELinux: Initializing.
[ 0.010244] *** VALIDATE selinux ***
[ 0.011456] Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear)
[ 0.011683] Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear)
[ 0.015044] *** VALIDATE tmpfs ***
[ 0.027687] *** VALIDATE proc ***
[ 0.036078] *** VALIDATE cgroup ***
[ 0.036343] *** VALIDATE cgroup2 ***
[ 0.051583] Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127
[ 0.053681] Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0
[ 0.055155] Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization
[ 0.056144] Spectre V2 : Mitigation: Full AMD retpoline
[ 0.056624] Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch
[ 0.057164] Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl and seccomp
[ 0.519777] Freeing SMP alternatives memory: 36K
[ 0.669583] smpboot: CPU0: AMD EPYC Processor (family: 0x17, model: 0x1, stepping: 0x2)
[ 0.685021] Performance Events: PMU not available due to virtualization, using software events only.
[ 0.688857] rcu: Hierarchical SRCU implementation.
[ 0.701650] NMI watchdog: Perf NMI watchdog permanently disabled
[ 0.704287] smp: Bringing up secondary CPUs ...
[ 0.704583] smp: Brought up 1 node, 1 CPU
[ 0.704583] smpboot: Max logical packages: 1
[ 0.704737] smpboot: Total of 1 processors activated (4190.06 BogoMIPS)
[ 0.733318] devtmpfs: initialized
[ 0.740627] x86/mm: Memory block size: 128MB
[ 0.755484] clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns
[ 0.756937] futex hash table entries: 256 (order: 2, 16384 bytes, linear)
[ 0.764211] pinctrl core: initialized pinctrl subsystem
[ 0.775894] PM: RTC time: 00:18:59, date: 2022-06-20
[ 0.776560] thermal_sys: Registered thermal governor 'fair_share'
[ 0.776622] thermal_sys: Registered thermal governor 'bang_bang'
[ 0.776921] thermal_sys: Registered thermal governor 'step_wise'
[ 0.777163] thermal_sys: Registered thermal governor 'user_space'
[ 0.787345] NET: Registered protocol family 16
[ 0.791609] audit: initializing netlink subsys (disabled)
[ 0.797174] audit: type=2000 audit(1655684338.806:1): state=initialized audit_enabled=0 res=1
[ 0.798177] cpuidle: using governor menu
[ 0.802260] ACPI: bus type PCI registered
[ 0.802651] acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5
[ 0.807303] PCI: MMCONFIG for domain 0000 [bus 00-ff] at [mem 0xb0000000-0xbfffffff] (base 0xb0000000)
[ 0.808167] PCI: MMCONFIG at [mem 0xb0000000-0xbfffffff] reserved in E820
[ 0.810896] PCI: Using configuration type 1 for base access
[ 0.843803] HugeTLB registered 1.00 GiB page size, pre-allocated 0 pages
[ 0.844155] HugeTLB registered 2.00 MiB page size, pre-allocated 0 pages
[ 1.355744] cryptd: max_cpu_qlen set to 1000
[ 1.400141] alg: No test for 842 (842-generic)
[ 1.401374] alg: No test for 842 (842-scomp)
[ 1.462085] ACPI: Added _OSI(Module Device)
[ 1.462371] ACPI: Added _OSI(Processor Device)
[ 1.462565] ACPI: Added _OSI(3.0 _SCP Extensions)
[ 1.462609] ACPI: Added _OSI(Processor Aggregator Device)
[ 1.462953] ACPI: Added _OSI(Linux-Dell-Video)
[ 1.463149] ACPI: Added _OSI(Linux-Lenovo-NV-HDMI-Audio)
[ 1.463384] ACPI: Added _OSI(Linux-HPI-Hybrid-Graphics)
[ 1.485583] ACPI: 1 ACPI AML tables successfully acquired and loaded
[ 1.506679] clocksource: timekeeping watchdog on CPU0: Marking clocksource 'tsc-early' as unstable because the skew is too large:
[ 1.507249] clocksource: 'refined-jiffies' wd_now: fffb7210 wd_last: fffb7020 mask: ffffffff
[ 1.507625] clocksource: 'tsc-early' cs_now: 5fb996a17 cs_last: 5b062e6d3 mask: ffffffffffffffff
[ 1.508202] tsc: Marking TSC unstable due to clocksource watchdog
[ 1.510108] ACPI: Interpreter enabled
[ 1.512334] ACPI: (supports S0 S5)
[ 1.512678] ACPI: Using IOAPIC for interrupt routing
[ 1.514583] PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug
[ 1.515583] ACPI: Enabled 1 GPEs in block 00 to 3F
[ 1.546079] ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff])
[ 1.546891] acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3]
[ 1.550156] acpi PNP0A08:00: _OSC: platform does not support [LTR]
[ 1.552397] acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug SHPCHotplug PME AER PCIeCapability]
[ 1.552923] PCI host bridge to bus 0000:00
[ 1.554042] pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window]
[ 1.554358] pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window]
[ 1.554612] pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window]
[ 1.555635] pci_bus 0000:00: root bus resource [mem 0x80000000-0xafffffff window]
[ 1.555983] pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window]
[ 1.556618] pci_bus 0000:00: root bus resource [mem 0x100000000-0x8ffffffff window]
[ 1.557760] pci_bus 0000:00: root bus resource [bus 00-ff]
[ 1.559003] pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000
[ 1.565876] pci 0000:00:01.0: [1234:1111] type 00 class 0x030000
[ 1.568965] pci 0000:00:01.0: reg 0x10: [mem 0xfb000000-0xfbffffff pref]
[ 1.572659] pci 0000:00:01.0: reg 0x18: [mem 0xfea10000-0xfea10fff]
[ 1.580663] pci 0000:00:01.0: reg 0x30: [mem 0xfea00000-0xfea0ffff pref]
[ 1.582847] pci 0000:00:02.0: [1b36:000c] type 01 class 0x060400
[ 1.583583] pci 0000:00:02.0: reg 0x10: [mem 0xfea11000-0xfea11fff]
[ 1.584678] pci 0000:00:02.1: [1b36:000c] type 01 class 0x060400
[ 1.587271] pci 0000:00:02.1: reg 0x10: [mem 0xfea12000-0xfea12fff]
[ 1.592410] pci 0000:00:02.2: [1b36:000c] type 01 class 0x060400
[ 1.596281] pci 0000:00:02.2: reg 0x10: [mem 0xfea13000-0xfea13fff]
[ 1.600397] pci 0000:00:02.3: [1b36:000c] type 01 class 0x060400
[ 1.602657] pci 0000:00:02.3: reg 0x10: [mem 0xfea14000-0xfea14fff]
[ 1.610252] pci 0000:00:02.4: [1b36:000c] type 01 class 0x060400
[ 1.612660] pci 0000:00:02.4: reg 0x10: [mem 0xfea15000-0xfea15fff]
[ 1.618801] pci 0000:00:02.5: [1b36:000c] type 01 class 0x060400
[ 1.621287] pci 0000:00:02.5: reg 0x10: [mem 0xfea16000-0xfea16fff]
[ 1.625949] pci 0000:00:02.6: [1b36:000c] type 01 class 0x060400
[ 1.628277] pci 0000:00:02.6: reg 0x10: [mem 0xfea17000-0xfea17fff]
[ 1.633827] pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100
[ 1.634583] pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO
[ 1.635849] pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601
[ 1.643270] pci 0000:00:1f.2: reg 0x20: [io 0xc040-0xc05f]
[ 1.644652] pci 0000:00:1f.2: reg 0x24: [mem 0xfea18000-0xfea18fff]
[ 1.647329] pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500
[ 1.650431] pci 0000:00:1f.3: reg 0x20: [io 0x0700-0x073f]
[ 1.656901] pci 0000:01:00.0: [1af4:1041] type 00 class 0x020000
[ 1.659677] pci 0000:01:00.0: reg 0x14: [mem 0xfe800000-0xfe800fff]
[ 1.664766] pci 0000:01:00.0: reg 0x20: [mem 0xfcc00000-0xfcc03fff 64bit pref]
[ 1.670183] pci 0000:00:02.0: PCI bridge to [bus 01]
[ 1.670583] pci 0000:00:02.0: bridge window [mem 0xfe800000-0xfe9fffff]
[ 1.670754] pci 0000:00:02.0: bridge window [mem 0xfcc00000-0xfcdfffff 64bit pref]
[ 1.672848] pci 0000:02:00.0: [1af4:1048] type 00 class 0x010000
[ 1.675668] pci 0000:02:00.0: reg 0x14: [mem 0xfe600000-0xfe600fff]
[ 1.679662] pci 0000:02:00.0: reg 0x20: [mem 0xfca00000-0xfca03fff 64bit pref]
[ 1.684713] pci 0000:00:02.1: PCI bridge to [bus 02]
[ 1.685017] pci 0000:00:02.1: bridge window [mem 0xfe600000-0xfe7fffff]
[ 1.685614] pci 0000:00:02.1: bridge window [mem 0xfca00000-0xfcbfffff 64bit pref]
[ 1.688638] pci 0000:03:00.0: [1af4:1043] type 00 class 0x078000
[ 1.691643] pci 0000:03:00.0: reg 0x14: [mem 0xfe400000-0xfe400fff]
[ 1.695671] pci 0000:03:00.0: reg 0x20: [mem 0xfc800000-0xfc803fff 64bit pref]
[ 1.700226] pci 0000:00:02.2: PCI bridge to [bus 03]
[ 1.700507] pci 0000:00:02.2: bridge window [mem 0xfe400000-0xfe5fffff]
[ 1.700619] pci 0000:00:02.2: bridge window [mem 0xfc800000-0xfc9fffff 64bit pref]
[ 1.702749] pci 0000:04:00.0: [1af4:1042] type 00 class 0x010000
[ 1.706662] pci 0000:04:00.0: reg 0x14: [mem 0xfe200000-0xfe200fff]
[ 1.712163] pci 0000:04:00.0: reg 0x20: [mem 0xfc600000-0xfc603fff 64bit pref]
[ 1.714949] pci 0000:00:02.3: PCI bridge to [bus 04]
[ 1.715251] pci 0000:00:02.3: bridge window [mem 0xfe200000-0xfe3fffff]
[ 1.715583] pci 0000:00:02.3: bridge window [mem 0xfc600000-0xfc7fffff 64bit pref]
[ 1.717010] pci 0000:05:00.0: [1af4:1042] type 00 class 0x010000
[ 1.720193] pci 0000:05:00.0: reg 0x14: [mem 0xfe000000-0xfe000fff]
[ 1.725157] pci 0000:05:00.0: reg 0x20: [mem 0xfc400000-0xfc403fff 64bit pref]
[ 1.727968] pci 0000:00:02.4: PCI bridge to [bus 05]
[ 1.728277] pci 0000:00:02.4: bridge window [mem 0xfe000000-0xfe1fffff]
[ 1.728628] pci 0000:00:02.4: bridge window [mem 0xfc400000-0xfc5fffff 64bit pref]
[ 1.730519] pci 0000:06:00.0: [1af4:1045] type 00 class 0x00ff00
[ 1.735955] pci 0000:06:00.0: reg 0x20: [mem 0xfc200000-0xfc203fff 64bit pref]
[ 1.738267] pci 0000:00:02.5: PCI bridge to [bus 06]
[ 1.738633] pci 0000:00:02.5: bridge window [mem 0xfde00000-0xfdffffff]
[ 1.738944] pci 0000:00:02.5: bridge window [mem 0xfc200000-0xfc3fffff 64bit pref]
[ 1.740892] pci 0000:00:02.6: PCI bridge to [bus 07]
[ 1.741189] pci 0000:00:02.6: bridge window [mem 0xfdc00000-0xfddfffff]
[ 1.741481] pci 0000:00:02.6: bridge window [mem 0xfc000000-0xfc1fffff 64bit pref]
[ 1.752616] ACPI: PCI Interrupt Link [LNKA] (IRQs 5 *10 11)
[ 1.754336] ACPI: PCI Interrupt Link [LNKB] (IRQs 5 *10 11)
[ 1.755237] ACPI: PCI Interrupt Link [LNKC] (IRQs 5 10 *11)
[ 1.756256] ACPI: PCI Interrupt Link [LNKD] (IRQs 5 10 *11)
[ 1.758385] ACPI: PCI Interrupt Link [LNKE] (IRQs 5 *10 11)
[ 1.759283] ACPI: PCI Interrupt Link [LNKF] (IRQs 5 *10 11)
[ 1.760298] ACPI: PCI Interrupt Link [LNKG] (IRQs 5 10 *11)
[ 1.762441] ACPI: PCI Interrupt Link [LNKH] (IRQs 5 10 *11)
[ 1.763674] ACPI: PCI Interrupt Link [GSIA] (IRQs *16)
[ 1.764091] ACPI: PCI Interrupt Link [GSIB] (IRQs *17)
[ 1.764445] ACPI: PCI Interrupt Link [GSIC] (IRQs *18)
[ 1.764825] ACPI: PCI Interrupt Link [GSID] (IRQs *19)
[ 1.765810] ACPI: PCI Interrupt Link [GSIE] (IRQs *20)
[ 1.766296] ACPI: PCI Interrupt Link [GSIF] (IRQs *21)
[ 1.766767] ACPI: PCI Interrupt Link [GSIG] (IRQs *22)
[ 1.767614] ACPI: PCI Interrupt Link [GSIH] (IRQs *23)
[ 1.773221] iommu: Default domain type: Translated
[ 1.777827] pci 0000:00:01.0: vgaarb: setting as boot VGA device
[ 1.778284] pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none
[ 1.778753] pci 0000:00:01.0: vgaarb: bridge control possible
[ 1.779638] vgaarb: loaded
[ 1.783097] SCSI subsystem initialized
[ 1.785955] ACPI: bus type USB registered
[ 1.787008] usbcore: registered new interface driver usbfs
[ 1.787867] usbcore: registered new interface driver hub
[ 1.788932] usbcore: registered new device driver usb
[ 1.790397] pps_core: LinuxPPS API ver. 1 registered
[ 1.790627] pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti <giometti@linux.it>
[ 1.791303] PTP clock support registered
[ 1.793725] EDAC MC: Ver: 3.0.0
[ 1.797905] PCI: Using ACPI for IRQ routing
[ 1.820583] NetLabel: Initializing
[ 1.820583] NetLabel: domain hash size = 128
[ 1.820583] NetLabel: protocols = UNLABELED CIPSOv4 CALIPSO
[ 1.821829] NetLabel: unlabeled traffic allowed by default
[ 1.829677] clocksource: Switched to clocksource refined-jiffies
[ 1.967326] *** VALIDATE bpf ***
[ 1.969384] VFS: Disk quotas dquot_6.6.0
[ 1.969920] VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes)
[ 1.972235] *** VALIDATE ramfs ***
[ 1.972562] *** VALIDATE hugetlbfs ***
[ 1.974800] pnp: PnP ACPI init
[ 1.980331] system 00:04: [mem 0xb0000000-0xbfffffff window] has been reserved
[ 1.983087] pnp: PnP ACPI: found 5 devices
[ 2.002557] clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns
[ 2.002557] clocksource: Switched to clocksource acpi_pm
[ 2.002557] pci 0000:00:02.0: bridge window [io 0x1000-0x0fff] to [bus 01] add_size 1000
[ 2.002557] pci 0000:00:02.1: bridge window [io 0x1000-0x0fff] to [bus 02] add_size 1000
[ 2.002557] pci 0000:00:02.2: bridge window [io 0x1000-0x0fff] to [bus 03] add_size 1000
[ 2.006995] pci 0000:00:02.3: bridge window [io 0x1000-0x0fff] to [bus 04] add_size 1000
[ 2.014770] pci 0000:00:02.4: bridge window [io 0x1000-0x0fff] to [bus 05] add_size 1000
[ 2.015126] pci 0000:00:02.5: bridge window [io 0x1000-0x0fff] to [bus 06] add_size 1000
[ 2.015512] pci 0000:00:02.6: bridge window [io 0x1000-0x0fff] to [bus 07] add_size1000
[ 2.042525] pci 0000:00:02.0: BAR 13: assigned [io 0x1000-0x1fff]
[ 2.042969] pci 0000:00:02.1: BAR 13: assigned [io 0x2000-0x2fff]
[ 2.043223] pci 0000:00:02.2: BAR 13: assigned [io 0x3000-0x3fff]
[ 2.043468] pci 0000:00:02.3: BAR 13: assigned [io 0x4000-0x4fff]
[ 2.044150] pci 0000:00:02.4: BAR 13: assigned [io 0x5000-0x5fff]
[ 2.044404] pci 0000:00:02.5: BAR 13: assigned [io 0x6000-0x6fff]
[ 2.045416] pci 0000:00:02.6: BAR 13: assigned [io 0x7000-0x7fff]
[ 2.046929] pci 0000:00:02.0: PCI bridge to [bus 01]
[ 2.053457] pci 0000:00:02.0: bridge window [io 0x1000-0x1fff]
[ 2.057205] pci 0000:00:02.0: bridge window [mem 0xfe800000-0xfe9fffff]
[ 2.059131] pci 0000:00:02.0: bridge window [mem 0xfcc00000-0xfcdfffff 64bit pref]
[ 2.061736] pci 0000:00:02.1: PCI bridge to [bus 02]
[ 2.063849] pci 0000:00:02.1: bridge window [io 0x2000-0x2fff]
[ 2.066161] pci 0000:00:02.1: bridge window [mem 0xfe600000-0xfe7fffff]
[ 2.072015] pci 0000:00:02.1: bridge window [mem 0xfca00000-0xfcbfffff 64bit pref]
[ 2.076185] pci 0000:00:02.2: PCI bridge to [bus 03]
[ 2.076465] pci 0000:00:02.2: bridge window [io 0x3000-0x3fff]
[ 2.078959] pci 0000:00:02.2: bridge window [mem 0xfe400000-0xfe5fffff]
[ 2.080599] pci 0000:00:02.2: bridge window [mem 0xfc800000-0xfc9fffff 64bit pref]
[ 2.084032] pci 0000:00:02.3: PCI bridge to [bus 04]
[ 2.084318] pci 0000:00:02.3: bridge window [io 0x4000-0x4fff]
[ 2.087977] pci 0000:00:02.3: bridge window [mem 0xfe200000-0xfe3fffff]
[ 2.089981] pci 0000:00:02.3: bridge window [mem 0xfc600000-0xfc7fffff 64bit pref]
[ 2.092431] pci 0000:00:02.4: PCI bridge to [bus 05]
[ 2.093025] pci 0000:00:02.4: bridge window [io 0x5000-0x5fff]
[ 2.095009] pci 0000:00:02.4: bridge window [mem 0xfe000000-0xfe1fffff]
[ 2.096629] pci 0000:00:02.4: bridge window [mem 0xfc400000-0xfc5fffff 64bit pref]
[ 2.100137] pci 0000:00:02.5: PCI bridge to [bus 06]
[ 2.100447] pci 0000:00:02.5: bridge window [io 0x6000-0x6fff]
[ 2.102288] pci 0000:00:02.5: bridge window [mem 0xfde00000-0xfdffffff]
[ 2.104069] pci 0000:00:02.5: bridge window [mem 0xfc200000-0xfc3fffff 64bit pref]
[ 2.106524] pci 0000:00:02.6: PCI bridge to [bus 07]
[ 2.107205] pci 0000:00:02.6: bridge window [io 0x7000-0x7fff]
[ 2.109424] pci 0000:00:02.6: bridge window [mem 0xfdc00000-0xfddfffff]
[ 2.112461] pci 0000:00:02.6: bridge window [mem 0xfc000000-0xfc1fffff 64bit pref]
[ 2.115121] pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window]
[ 2.115461] pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window]
[ 2.116133] pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window]
[ 2.116405] pci_bus 0000:00: resource 7 [mem 0x80000000-0xafffffff window]
[ 2.116956] pci_bus 0000:00: resource 8 [mem 0xc0000000-0xfebfffff window]
[ 2.117512] pci_bus 0000:00: resource 9 [mem 0x100000000-0x8ffffffff window]
[ 2.118068] pci_bus 0000:01: resource 0 [io 0x1000-0x1fff]
[ 2.118522] pci_bus 0000:01: resource 1 [mem 0xfe800000-0xfe9fffff]
[ 2.118985] pci_bus 0000:01: resource 2 [mem 0xfcc00000-0xfcdfffff 64bit pref]
[ 2.119486] pci_bus 0000:02: resource 0 [io 0x2000-0x2fff]
[ 2.120145] pci_bus 0000:02: resource 1 [mem 0xfe600000-0xfe7fffff]
[ 2.120426] pci_bus 0000:02: resource 2 [mem 0xfca00000-0xfcbfffff 64bit pref]
[ 2.121104] pci_bus 0000:03: resource 0 [io 0x3000-0x3fff]
[ 2.121511] pci_bus 0000:03: resource 1 [mem 0xfe400000-0xfe5fffff]
[ 2.122493] pci_bus 0000:03: resource 2 [mem 0xfc800000-0xfc9fffff 64bit pref]
[ 2.123214] pci_bus 0000:04: resource 0 [io 0x4000-0x4fff]
[ 2.123457] pci_bus 0000:04: resource 1 [mem 0xfe200000-0xfe3fffff]
[ 2.124040] pci_bus 0000:04: resource 2 [mem 0xfc600000-0xfc7fffff 64bit pref]
[ 2.124910] pci_bus 0000:05: resource 0 [io 0x5000-0x5fff]
[ 2.125350] pci_bus 0000:05: resource 1 [mem 0xfe000000-0xfe1fffff]
[ 2.125779] pci_bus 0000:05: resource 2 [mem 0xfc400000-0xfc5fffff 64bit pref]
[ 2.126370] pci_bus 0000:06: resource 0 [io 0x6000-0x6fff]
[ 2.126730] pci_bus 0000:06: resource 1 [mem 0xfde00000-0xfdffffff]
[ 2.127167] pci_bus 0000:06: resource 2 [mem 0xfc200000-0xfc3fffff 64bit pref]
[ 2.127680] pci_bus 0000:07: resource 0 [io 0x7000-0x7fff]
[ 2.128061] pci_bus 0000:07: resource 1 [mem 0xfdc00000-0xfddfffff]
[ 2.128502] pci_bus 0000:07: resource 2 [mem 0xfc000000-0xfc1fffff 64bit pref]
[ 2.131196] NET: Registered protocol family 2
[ 2.146990] tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear)
[ 2.148054] TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear)
[ 2.149405] TCP bind hash table entries: 16384 (order: 6, 262144 bytes, linear)
[ 2.151881] TCP: Hash tables configured (established 16384 bind 16384)
[ 2.155134] UDP hash table entries: 1024 (order: 3, 32768 bytes, linear)
[ 2.156342] UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear)
[ 2.159870] NET: Registered protocol family 1
[ 2.160780] NET: Registered protocol family 44
[ 2.161983] pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff]
[ 2.163097] PCI: CLS 0 bytes, default 64
[ 2.169352] Trying to unpack rootfs image as initramfs...
[ 4.542382] Freeing initrd memory: 24888K
[ 4.553360] Initialise system trusted keyrings
[ 4.556745] Key type blacklist registered
[ 4.570830] workingset: timestamp_bits=36 max_order=19 bucket_order=0
[ 4.590533] zbud: loaded
[ 4.607530] Platform Keyring initialized
[ 4.779329] NET: Registered protocol family 38
[ 4.780051] Key type asymmetric registered
[ 4.780361] Asymmetric key parser 'x509' registered
[ 4.781031] Block layer SCSI generic (bsg) driver version 0.4 loaded (major 246)
[ 4.783074] io scheduler mq-deadline registered
[ 4.783377] io scheduler kyber registered
[ 4.784312] io scheduler bfq registered
[ 4.790248] atomic64_test: passed for x86-64 platform with CX8 and with SSE
[ 4.799221] PCI Interrupt Link [GSIG] enabled at IRQ 22
[ 4.810033] pcieport 0000:00:02.0: PME: Signaling with IRQ 24
[ 4.814322] pcieport 0000:00:02.0: AER: enabled with IRQ 24
[ 4.815530] pcieport 0000:00:02.0: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- LLActRep+
[ 4.828382] pcieport 0000:00:02.1: PME: Signaling with IRQ 25
[ 4.833004] pcieport 0000:00:02.1: AER: enabled with IRQ 25
[ 4.833567] pcieport 0000:00:02.1: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- LLActRep+
[ 4.840281] pcieport 0000:00:02.2: PME: Signaling with IRQ 26
[ 4.842580] pcieport 0000:00:02.2: AER: enabled with IRQ 26
[ 4.842747] pcieport 0000:00:02.2: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- LLActRep+
[ 4.851160] pcieport 0000:00:02.3: PME: Signaling with IRQ 27
[ 4.852792] pcieport 0000:00:02.3: AER: enabled with IRQ 27
[ 4.853347] pcieport 0000:00:02.3: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- LLActRep+
[ 4.861147] pcieport 0000:00:02.4: PME: Signaling with IRQ 28
[ 4.862527] pcieport 0000:00:02.4: AER: enabled with IRQ 28
[ 4.862737] pcieport 0000:00:02.4: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- LLActRep+
[ 4.870268] pcieport 0000:00:02.5: PME: Signaling with IRQ 29
[ 4.872779] pcieport 0000:00:02.5: AER: enabled with IRQ 29
[ 4.873369] pcieport 0000:00:02.5: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- LLActRep+
[ 4.882007] pcieport 0000:00:02.6: PME: Signaling with IRQ 30
[ 4.883826] pcieport 0000:00:02.6: AER: enabled with IRQ 30
[ 4.884780] pcieport 0000:00:02.6: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- LLActRep+
[ 4.888042] shpchp: Standard Hot Plug PCI Controller Driver version: 0.4
[ 4.892086] input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input0
[ 4.898164] ACPI: Power Button [PWRF]
[ 4.935197] Serial: 8250/16550 driver, 32 ports, IRQ sharing enabled
[ 4.938058] 00:00: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A
[ 4.963893] Non-volatile memory driver v1.3
[ 4.964804] Linux agpgart interface v0.103
[ 4.976341] PCI Interrupt Link [GSIA] enabled at IRQ 16
[ 4.983075] ahci 0000:00:1f.2: AHCI 0001.0000 32 slots 6 ports 1.5 Gbps 0x3f impl SATA mode
[ 4.983620] ahci 0000:00:1f.2: flags: 64bit ncq only
[ 4.996057] scsi host0: ahci
[ 5.002047] scsi host1: ahci
[ 5.004075] scsi host2: ahci
[ 5.005918] scsi host3: ahci
[ 5.007537] scsi host4: ahci
[ 5.009383] scsi host5: ahci
[ 5.010769] ata1: SATA max UDMA/133 abar m4096@0xfea18000 port 0xfea18100 irq 31
[ 5.011347] ata2: SATA max UDMA/133 abar m4096@0xfea18000 port 0xfea18180 irq 31
[ 5.011738] ata3: SATA max UDMA/133 abar m4096@0xfea18000 port 0xfea18200 irq 31
[ 5.012176] ata4: SATA max UDMA/133 abar m4096@0xfea18000 port 0xfea18280 irq 31
[ 5.012555] ata5: SATA max UDMA/133 abar m4096@0xfea18000 port 0xfea18300 irq 31
[ 5.019811] ata6: SATA max UDMA/133 abar m4096@0xfea18000 port 0xfea18380 irq 31
[ 5.034078] libphy: Fixed MDIO Bus: probed
[ 5.037211] ehci_hcd: USB 2.0 'Enhanced' Host Controller (EHCI) Driver
[ 5.037633] ehci-pci: EHCI PCI platform driver
[ 5.039214] ohci_hcd: USB 1.1 'Open' Host Controller (OHCI) Driver
[ 5.039723] ohci-pci: OHCI PCI platform driver
[ 5.039939] uhci_hcd: USB Universal Host Controller Interface driver
[ 5.041973] usbcore: registered new interface driver usbserial_generic
[ 5.042729] usbserial: USB Serial support registered for generic
[ 5.044602] i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12
[ 5.054848] serio: i8042 KBD port at 0x60,0x64 irq 1
[ 5.055883] serio: i8042 AUX port at 0x60,0x64 irq 12
[ 5.060250] mousedev: PS/2 mouse device common for all mice
[ 5.065959] input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input1
[ 5.068818] rtc_cmos 00:03: RTC can wake from S4
[ 5.081750] rtc_cmos 00:03: registered as rtc0
[ 5.082339] rtc_cmos 00:03: alarms up to one day, y3k, 242 bytes nvram
[ 5.083993] device-mapper: uevent: version 1.0.3
[ 5.086593] device-mapper: ioctl: 4.42.0-ioctl (2020-02-27) initialised: dm-devel@redhat.com
[ 5.091527] hid: raw HID events driver (C) Jiri Kosina
[ 5.093060] usbcore: registered new interface driver usbhid
[ 5.093480] usbhid: USB HID core driver
[ 5.094910] drop_monitor: Initializing network drop monitor service
[ 5.097096] Initializing XFRM netlink socket
[ 5.099219] NET: Registered protocol family 10
[ 5.347547] ata2: SATA link down (SStatus 0 SControl 300)
[ 5.351228] ata1: SATA link down (SStatus 0 SControl 300)
[ 5.351564] ata6: SATA link down (SStatus 0 SControl 300)
[ 5.352244] ata5: SATA link down (SStatus 0 SControl 300)
[ 5.352945] ata4: SATA link down (SStatus 0 SControl 300)
[ 5.353352] ata3: SATA link down (SStatus 0 SControl 300)
[ 5.425106] Segment Routing with IPv6
[ 5.426313] mip6: Mobile IPv6
[ 5.426785] NET: Registered protocol family 17
[ 5.435711] RAS: Correctable Errors collector initialized.
[ 5.436999] IPI shorthand broadcast: enabled
[ 5.437778] SSE version of gcm_enc/dec engaged.
[ 5.940596] input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input3
[ 7.826578] registered taskstats version 1
[ 7.827736] Loading compiled-in X.509 certificates
[ 8.039529] Loaded X.509 cert 'Fedora kernel signing key: 0d97d514d1651893ca89cc557d3830fffcdfaba2'
[ 8.039630] zswap: loaded using pool lzo/zbud
[ 8.044752] Key type ._fscrypt registered
[ 8.044752] Key type .fscrypt registered
[ 8.044752] Key type fscrypt-provisioning registered
[ 8.514128] Key type big_key registered
[ 8.738185] Key type encrypted registered
[ 8.739272] ima: No TPM chip found, activating TPM-bypass!
[ 8.739925] ima: Allocated hash algorithm: sha256
[ 8.742586] ima: No architecture policies found
[ 8.748284] PM: Magic number: 10:769:304
[ 8.749048] tty tty33: hash matches
[ 8.750310] rtc_cmos 00:03: setting system clock to 2022-06-20T00:19:08 UTC (1655684348)
[ 8.752066] Unstable clock detected, switching default tracing clock to "global"
[ 8.752066] If you want to keep using the local clock, then add:
[ 8.752066] "trace_clock=local"
[ 8.752066] on the kernel command line
[ 8.797816] Freeing unused decrypted memory: 2040K
[ 8.865830] Freeing unused kernel image (initmem) memory: 2452K
[ 8.865830] Write protecting the kernel read-only data: 22528k
[ 8.873768] Freeing unused kernel image (text/rodata gap) memory: 2044K
[ 8.876135] Freeing unused kernel image (rodata/data gap) memory: 1276K
[ 8.998234] x86/mm: Checked W+X mappings: passed, no W+X pages found.
[ 8.999174] rodata_test: all tests were successful
[ 8.999545] Run /init as init process
[ 9.286619] systemd[1]: systemd v245.4-1.fc32 running in system mode. (+PAM +AUDIT +SELINUX +IMA -APPARMOR +SMACK +SYSVINIT +UTMP +LIBCRYPTSETUP +GCRYPT +GNUTLS +ACL +XZ +LZ4 +SECCOMP +BLKID +ELFUTILS +KMOD +IDN2 -IDN +PCRE2 default-hierarchy=unified)
[ 9.290468] systemd[1]: Detected virtualization qemu.
[ 9.291337] systemd[1]: Detected architecture x86-64.
[ 9.291927] systemd[1]: Running in initial RAM disk.
Welcome to Fedora 32 (Cloud Edition) dracut-050-26.git20200316.fc32 (Initramfs)!
[ 9.303064] systemd[1]: No hostname configured.
[ 9.304165] systemd[1]: Set hostname to <localhost>.
[ 9.306222] systemd[1]: Initializing machine ID from random generator.
[ 12.443824] systemd[1]: Started Dispatch Password Requests to Console Directory Watch.
[ OK ] Started Dispatch Password …ts to Console Directory Watch.
[ 12.458862] systemd[1]: Reached target Local Encrypted Volumes.
[ OK ] Reached target Local Encrypted Volumes.
[ 12.466188] systemd[1]: Reached target Local File Systems.
[ OK ] Reached target Local File Systems.
[ 12.469030] systemd[1]: Reached target Paths.
[ OK ] Reached target Paths.
[ 12.477579] systemd[1]: Reached target Slices.
[ OK ] Reached target Slices.
[ 12.479424] systemd[1]: Reached target Swap.
[ OK ] Reached target Swap.
[ 12.482183] systemd[1]: Reached target Timers.
[ OK ] Reached target Timers.
[ 12.497348] systemd[1]: Listening on Journal Audit Socket.
[ OK ] Listening on Journal Audit Socket.
[ 12.502584] systemd[1]: Listening on Journal Socket (/dev/log).
[ OK ] Listening on Journal Socket (/dev/log).
[ 12.507332] systemd[1]: Listening on Journal Socket.
[ OK ] Listening on Journal Socket.
[ 12.512609] systemd[1]: Listening on udev Control Socket.
[ OK ] Listening on udev Control Socket.
[ 12.516363] systemd[1]: Listening on udev Kernel Socket.
[ OK ] Listening on udev Kernel Socket.
[ 12.519117] systemd[1]: Reached target Sockets.
[ OK ] Reached target Sockets.
[ 12.586364] systemd[1]: Starting Create list of static device nodes for the current kernel...
Starting Create list of st…odes for the current kernel...
[ 12.698054] systemd[1]: Starting Journal Service...
Starting Journal Service...
[ 12.698054] systemd[1]: Condition check resulted in Load Kernel Modules being skipped.
[ 12.846193] systemd[1]: Starting Apply Kernel Variables...
Starting Apply Kernel Variables...
[ 13.004523] systemd[1]: Starting Setup Virtual Console...
Starting Setup Virtual Console...
[ 13.149762] systemd[1]: Finished Create list of static device nodes for the current kernel.
[ OK ] Finished Create list of st… nodes for the current kernel.
[ 13.276304] systemd[1]: Starting Create Static Device Nodes in /dev...
Starting Create Static Device Nodes in /dev...
[ 13.892550] systemd[1]: Finished Setup Virtual Console.
[ OK ] Finished Setup Virtual Console.
[ 13.935556] systemd[1]: Finished Apply Kernel Variables.
[ OK ] Finished Apply Kernel Variables.
[ 13.963808] systemd[1]: Condition check resulted in dracut ask for additional cmdline parameters being skipped.
[ 14.040974] systemd[1]: Starting dracut cmdline hook...
Starting dracut cmdline hook...
[ 14.282539] systemd[1]: Finished Create Static Device Nodes in /dev.
[ OK ] Finished Create Static Device Nodes in /dev.
[ 16.612703] systemd[1]: Finished dracut cmdline hook.
[ OK ] Finished dracut cmdline hook.
[ 16.677752] systemd[1]: Starting dracut pre-udev hook...
Starting dracut pre-udev hook...
[ 18.238582] systemd[1]: Finished dracut pre-udev hook.
[ OK ] Finished dracut pre-udev hook.
[ 18.302243] systemd[1]: Starting udev Kernel Device Manager...
Starting udev Kernel Device Manager...
[** ] (1 of 3) A start job is running for…-a6ba-f377f0c90217 (7s / no limit)
M
[*** ] (1 of 3) A start job is running for…-a6ba-f377f0c90217 (8s / no limit)
M
[ *** ] (2 of 3) A start job is running for Journal Service (8s / 1min 30s)
M
[ *** ] (2 of 3) A start job is running for Journal Service (9s / 1min 30s)
M
[ ***] (2 of 3) A start job is running for Journal Service (9s / 1min 30s)
M
[ **] (3 of 3) A start job is running for…el Device Manager (10s / 1min 35s)
M
[ *] (3 of 3) A start job is running for…el Device Manager (10s / 1min 35s)
M
[ **] (3 of 3) A start job is running for…el Device Manager (11s / 1min 35s)
M
[ ***] (1 of 3) A start job is running for…a6ba-f377f0c90217 (11s / no limit)
M
[ *** ] (1 of 3) A start job is running for…a6ba-f377f0c90217 (12s / no limit)
M
[ *** ] (1 of 3) A start job is running for…a6ba-f377f0c90217 (12s / no limit)
[ 25.713599] systemd[1]: Started Journal Service.
M
[ OK ] Started Journal Service.
[ 25.731118] audit: type=1130 audit(1655684365.480:2): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
Starting Create Volatile Files and Directories...
[ OK ] Finished Create Volatile Files and Directories.
[ 26.590432] audit: type=1130 audit(1655684366.339:3): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
[ OK ] Started udev Kernel Device Manager.
[ 27.724179] audit: type=1130 audit(1655684367.473:4): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
Starting udev Coldplug all Devices...
[*** ] (1 of 2) A start job is running for…a6ba-f377f0c90217 (17s / no limit)
M
[** ] (1 of 2) A start job is running for…a6ba-f377f0c90217 (17s / no limit)
M
[* ] (1 of 2) A start job is running for…a6ba-f377f0c90217 (18s / no limit)
M
[ OK ] Finished udev Coldplug all Devices.
[ 30.902222] audit: type=1130 audit(1655684370.651:5): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
[ OK ] Reached target System Initialization.
[ OK ] Reached target Basic System.
Starting dracut initqueue hook...
Mounting Kernel Configuration File System...
[ OK ] Mounted Kernel Configuration File System.
[ 33.488601] scsi host6: Virtio SCSI HBA
[ 34.035854] virtio_blk virtio3: [vda] 8388608 512-byte logical blocks (4.29 GB/4.00 GiB)
[ 34.216126] vda: vda1
[ 34.402440] virtio_blk virtio4: [vdb] 2048 512-byte logical blocks (1.05 MB/1.00 MiB)
[ OK ] Found device /dev/disk/by-…4-3bbb-40b2-a6ba-f377f0c90217.
[ OK ] Reached target Initrd Root Device.
[ OK ] Finished dracut initqueue hook.
[ 39.252589] audit: type=1130 audit(1655684379.001:6): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
[ OK ] Reached target Remote File Systems (Pre).
[ OK ] Reached target Remote File Systems.
Starting File System Check…3bbb-40b2-a6ba-f377f0c90217...
[ OK ] Finished File System Check…4-3bbb-40b2-a6ba-f377f0c90217.
[ 39.897307] audit: type=1130 audit(1655684379.644:7): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
Mounting /sysroot...
[ 40.195422] EXT4-fs (vda1): mounted filesystem with ordered data mode. Opts: (null)
[ OK ] Mounted /sysroot.
[ OK ] Reached target Initrd Root File System.
Starting Reload Configuration from the Real Root...
[ 40.617082] audit: type=1334 audit(1655684380.366:8): prog-id=5 op=UNLOAD
[ 40.623104] audit: type=1334 audit(1655684380.371:9): prog-id=4 op=UNLOAD
[ 40.624866] audit: type=1334 audit(1655684380.373:10): prog-id=3 op=UNLOAD
[ 40.664188] audit: type=1334 audit(1655684380.412:11): prog-id=7 op=UNLOAD
[ 40.675901] audit: type=1334 audit(1655684380.424:12): prog-id=6 op=UNLOAD
[ 44.358489] audit: type=1334 audit(1655684384.107:13): prog-id=8 op=LOAD
[ 44.371157] audit: type=1334 audit(1655684384.119:14): prog-id=9 op=LOAD
[ 44.390667] audit: type=1334 audit(1655684384.139:15): prog-id=10 op=LOAD
[ 44.399038] audit: type=1334 audit(1655684384.148:16): prog-id=11 op=LOAD
[ 44.403304] audit: type=1334 audit(1655684384.152:17): prog-id=12 op=LOAD
[ OK ] Finished Reload Configuration from the Real Root.
[ 45.039677] audit: type=1130 audit(1655684384.788:18): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
[ 45.045343] audit: type=1131 audit(1655684384.794:19): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
[ OK ] Reached target Initrd File Systems.
[ OK ] Reached target Initrd Default Target.
Starting dracut mount hook...
[ OK ] Finished dracut mount hook.
[ 45.614919] audit: type=1130 audit(1655684385.363:20): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
Starting Cleaning Up and Shutting Down Daemons...
[ OK ] Stopped target Initrd Default Target.
[ OK ] Stopped target Basic System.
[ OK ] Stopped target Initrd Root Device.
[ OK ] Stopped target Paths.
[ OK ] Stopped target Remote File Systems.
[ OK ] Stopped target Remote File Systems (Pre).
[ OK ] Stopped target Slices.
[ OK ] Stopped target Sockets.
[ OK ] Stopped target System Initialization.
[ OK ] Stopped target Local Encrypted Volumes.
[ OK ] Stopped Dispatch Password …ts to Console Directory Watch.
[ OK ] Stopped target Swap.
[ OK ] Stopped target Timers.
[ OK ] Stopped dracut mount hook.
[ 46.113349] audit: type=1131 audit(1655684385.862:21): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
[ OK ] Stopped dracut initqueue hook.
[ 46.143524] audit: type=1131 audit(1655684385.892:22): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
[ OK ] Stopped Apply Kernel Variables.
[ OK ] Stopped Create Volatile Files and Directories.
[ OK ] Stopped target Local File Systems.
[ OK ] Stopped udev Coldplug all Devices.
Stopping udev Kernel Device Manager...
[ OK ] Stopped Setup Virtual Console.
[ OK ] Stopped udev Kernel Device Manager.
[ OK ] Closed udev Control Socket.
[ OK ] Closed udev Kernel Socket.
[ OK ] Stopped dracut pre-udev hook.
[ OK ] Stopped dracut cmdline hook.
Starting Cleanup udevd DB...
[ OK ] Stopped Create Static Device Nodes in /dev.
[ OK ] Stopped Create list of sta… nodes for the current kernel.
[ OK ] Finished Cleaning Up and Shutting Down Daemons.
[ OK ] Finished Cleanup udevd DB.
[ OK ] Reached target Switch Root.
Starting Switch Root...
[ 47.111377] systemd-journald[293]: Received SIGTERM from PID 1 (systemd).
[ 51.412571] SELinux: Permission watch in class filesystem not defined in policy.
[ 51.413259] SELinux: Permission watch in class file not defined in policy.
[ 51.413592] SELinux: Permission watch_mount in class file not defined in policy.
[ 51.413911] SELinux: Permission watch_sb in class file not defined in policy.
[ 51.414344] SELinux: Permission watch_with_perm in class file not defined in policy.
[ 51.427091] SELinux: Permission watch_reads in class file not defined in policy.
[ 51.427438] SELinux: Permission watch in class dir not defined in policy.
[ 51.427692] SELinux: Permission watch_mount in class dir not defined in policy.
[ 51.429703] SELinux: Permission watch_sb in class dir not defined in policy.
[ 51.430091] SELinux: Permission watch_with_perm in class dir not defined in policy.
[ 51.430543] SELinux: Permission watch_reads in class dir not defined in policy.
[ 51.430653] SELinux: Permission watch in class lnk_file not defined in policy.
[ 51.431603] SELinux: Permission watch_mount in class lnk_file not defined in policy.
[ 51.432072] SELinux: Permission watch_sb in class lnk_file not defined in policy.
[ 51.432493] SELinux: Permission watch_with_perm in class lnk_file not defined in policy.
[ 51.432872] SELinux: Permission watch_reads in class lnk_file not defined in policy.
[ 51.433405] SELinux: Permission watch in class chr_file not defined in policy.
[ 51.433856] SELinux: Permission watch_mount in class chr_file not defined in policy.
[ 51.434932] SELinux: Permission watch_sb in class chr_file not defined in policy.
[ 51.435261] SELinux: Permission watch_with_perm in class chr_file not defined in policy.
[ 51.435616] SELinux: Permission watch_reads in class chr_file not defined in policy.
[ 51.436677] SELinux: Permission watch in class blk_file not defined in policy.
[ 51.437639] SELinux: Permission watch_mount in class blk_file not defined in policy.
[ 51.437639] SELinux: Permission watch_sb in class blk_file not defined in policy.
[ 51.439279] SELinux: Permission watch_with_perm in class blk_file not defined in policy.
[ 51.439901] SELinux: Permission watch_reads in class blk_file not defined in policy.
[ 51.440285] SELinux: Permission watch in class sock_file not defined in policy.
[ 51.442313] SELinux: Permission watch_mount in class sock_file not defined in policy.
[ 51.442788] SELinux: Permission watch_sb in class sock_file not defined in policy.
[ 51.443114] SELinux: Permission watch_with_perm in class sock_file not defined in policy.
[ 51.443447] SELinux: Permission watch_reads in class sock_file not defined in policy.
[ 51.449286] SELinux: Permission watch in class fifo_file not defined in policy.
[ 51.449860] SELinux: Permission watch_mount in class fifo_file not defined in policy.
[ 51.450273] SELinux: Permission watch_sb in class fifo_file not defined in policy.
[ 51.450823] SELinux: Permission watch_with_perm in class fifo_file not defined in policy.
[ 51.451192] SELinux: Permission watch_reads in class fifo_file not defined in policy.
[ 51.453112] SELinux: Class perf_event not defined in policy.
[ 51.453493] SELinux: Class lockdown not defined in policy.
[ 51.454050] SELinux: the above unknown classes and permissions will be allowed
[ 51.456135] SELinux: policy capability network_peer_controls=1
[ 51.456441] SELinux: policy capability open_perms=1
[ 51.456701] SELinux: policy capability extended_socket_class=1
[ 51.457185] SELinux: policy capability always_check_network=0
[ 51.457425] SELinux: policy capability cgroup_seclabel=1
[ 51.457856] SELinux: policy capability nnp_nosuid_transition=1
[ 51.579868] kauditd_printk_skb: 19 callbacks suppressed
[ 51.579868] audit: type=1403 audit(1655684391.328:42): auid=4294967295 ses=4294967295 lsm=selinux res=1
[ 51.626430] systemd[1]: Successfully loaded SELinux policy in 3.613504s.
[ 52.332011] systemd[1]: Relabelled /dev, /dev/shm, /run, /sys/fs/cgroup in 430.677ms.
[ 52.383481] systemd[1]: systemd v245.4-1.fc32 running in system mode. (+PAM +AUDIT +SELINUX +IMA -APPARMOR +SMACK +SYSVINIT +UTMP +LIBCRYPTSETUP +GCRYPT +GNUTLS +ACL +XZ +LZ4 +SECCOMP +BLKID +ELFUTILS +KMOD +IDN2 -IDN +PCRE2 default-hierarchy=unified)
[ 52.387052] systemd[1]: Detected virtualization qemu.
[ 52.388075] systemd[1]: Detected architecture x86-64.
Welcome to Fedora 32 (Cloud Edition)!
[ 52.405181] systemd[1]: Set hostname to <localhost.localdomain>.
[ 52.417556] systemd[1]: Initializing machine ID from random generator.
[ 52.420931] systemd[1]: Installed transient /etc/machine-id file.
[ 52.513109] audit: type=1334 audit(1655684392.262:43): prog-id=13 op=LOAD
[ 52.514849] audit: type=1334 audit(1655684392.263:44): prog-id=13 op=UNLOAD
[ 52.517277] audit: type=1334 audit(1655684392.266:45): prog-id=14 op=LOAD
[ 52.518959] audit: type=1334 audit(1655684392.267:46): prog-id=14 op=UNLOAD
[ 58.685476] systemd[1]: /usr/lib/systemd/system/sssd.service:13: PIDFile= references a path below legacy directory /var/run/, updating /var/run/sssd.pid → /run/sssd.pid; please update the unit file accordingly.
[ 59.223216] systemd[1]: /usr/lib/systemd/system/sssd-kcm.socket:7: ListenStream= references a path below legacy directory /var/run/, updating /var/run/.heim_org.h5l.kcm-socket → /run/.heim_org.h5l.kcm-socket; please update the unit file accordingly.
[ 59.995146] audit: type=1334 audit(1655684399.744:47): prog-id=15 op=LOAD
[ 60.009201] audit: type=1334 audit(1655684399.758:48): prog-id=16 op=LOAD
[ 60.011298] audit: type=1334 audit(1655684399.760:49): prog-id=17 op=LOAD
[ 60.062073] audit: type=1334 audit(1655684399.801:50): prog-id=15 op=UNLOAD
[ 60.066034] audit: type=1131 audit(1655684399.814:51): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
[ 60.086563] systemd[1]: initrd-switch-root.service: Succeeded.
[ 60.090604] systemd[1]: Stopped Switch Root.
[ OK ] Stopped Switch Root.
[ 60.099805] audit: type=1130 audit(1655684399.848:52): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
[ 60.100693] audit: type=1131 audit(1655684399.849:53): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
[ 60.109386] systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1.
[ 60.117509] systemd[1]: Created slice system-getty.slice.
[ OK ] Created slice system-getty.slice.
[ 60.123920] systemd[1]: Created slice system-modprobe.slice.
[ OK ] Created slice system-modprobe.slice.
[ 60.129385] systemd[1]: Created slice system-serial\x2dgetty.slice.
[ OK ] Created slice system-serial\x2dgetty.slice.
[ 60.134938] systemd[1]: Created slice system-sshd\x2dkeygen.slice.
[ OK ] Created slice system-sshd\x2dkeygen.slice.
[ 60.143397] systemd[1]: Created slice User and Session Slice.
[ OK ] Created slice User and Session Slice.
[ 60.149282] systemd[1]: Started Dispatch Password Requests to Console Directory Watch.
[ OK ] Started Dispatch Password …ts to Console Directory Watch.
[ 60.156143] systemd[1]: Started Forward Password Requests to Wall Directory Watch.
[ OK ] Started Forward Password R…uests to Wall Directory Watch.
[ 60.169664] systemd[1]: Set up automount Arbitrary Executable File Formats File System Automount Point.
[ OK ] Set up automount Arbitrary…s File System Automount Point.
[ 60.172255] systemd[1]: Reached target Local Encrypted Volumes.
[ OK ] Reached target Local Encrypted Volumes.
[ 60.174585] systemd[1]: Stopped target Switch Root.
[ OK ] Stopped target Switch Root.
[ 60.177190] systemd[1]: Stopped target Initrd File Systems.
[ OK ] Stopped target Initrd File Systems.
[ 60.179072] systemd[1]: Stopped target Initrd Root File System.
[ OK ] Stopped target Initrd Root File System.
[ 60.181303] systemd[1]: Reached target Paths.
[ OK ] Reached target Paths.
[ 60.183237] systemd[1]: Reached target Remote File Systems.
[ OK ] Reached target Remote File Systems.
[ 60.185549] systemd[1]: Reached target Slices.
[ OK ] Reached target Slices.
[ 60.188017] systemd[1]: Reached target Swap.
[ OK ] Reached target Swap.
[ 60.229576] systemd[1]: Listening on Process Core Dump Socket.
[ OK ] Listening on Process Core Dump Socket.
[ 60.235388] systemd[1]: Listening on initctl Compatibility Named Pipe.
[ OK ] Listening on initctl Compatibility Named Pipe.
[ 60.249958] systemd[1]: Listening on udev Control Socket.
[ OK ] Listening on udev Control Socket.
[ 60.256085] systemd[1]: Listening on udev Kernel Socket.
[ OK ] Listening on udev Kernel Socket.
[ 60.264138] systemd[1]: Listening on User Database Manager Socket.
[ OK ] Listening on User Database Manager Socket.
[ 60.288194] systemd[1]: Mounting Huge Pages File System...
Mounting Huge Pages File System...
[ 60.344667] systemd[1]: Mounting POSIX Message Queue File System...
Mounting POSIX Message Queue File System...
[ 60.441022] systemd[1]: Mounting Kernel Debug File System...
Mounting Kernel Debug File System...
[ 60.592650] systemd[1]: Mounting Kernel Trace File System...
Mounting Kernel Trace File System...
[ 60.726841] systemd[1]: Starting Create list of static device nodes for the current kernel...
Starting Create list of st…odes for the current kernel...
[ 60.996336] systemd[1]: Starting Load Kernel Module drm...
Starting Load Kernel Module drm...
[ 61.105979] systemd[1]: Condition check resulted in Set Up Additional Binary Formats being skipped.
[ 61.116371] systemd[1]: Stopped Journal Service.
[ OK ] Stopped Journal Service.
[ 61.135271] audit: type=1130 audit(1655684400.884:54): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
[ 61.142346] audit: type=1131 audit(1655684400.891:55): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
[ 61.149591] systemd[1]: systemd-journald.service: Consumed 6.411s CPU time.
[ 61.239703] audit: type=1334 audit(1655684400.988:56): prog-id=18 op=LOAD
[ 61.407372] systemd[1]: Starting Journal Service...
Starting Journal Service...
[ 61.514430] systemd[1]: Condition check resulted in Load Kernel Modules being skipped.
[ 61.537665] systemd[1]: Condition check resulted in FUSE Control File System being skipped.
[ 61.654629] systemd[1]: Starting Remount Root and Kernel File Systems...
Starting Remount Root and Kernel File Systems...
[ 61.770073] systemd[1]: Starting Repartition Root Disk...
Starting Repartition Root Disk...
[ 61.933522] systemd[1]: Starting Apply Kernel Variables...
Starting Apply Kernel Variables...
[ 62.158853] systemd[1]: Starting udev Coldplug all Devices...
Starting udev Coldplug all Devices...
[ 62.274640] systemd[1]: sysroot.mount: Succeeded.
[ 62.463312] systemd[1]: Mounted Huge Pages File System.
[ OK ] Mounted Huge Pages File System.
[ 62.523217] systemd[1]: Mounted POSIX Message Queue File System.
[ OK ] Mounted POSIX Message Queue File System.
[ 62.570208] systemd[1]: Mounted Kernel Debug File System.
[ OK ] Mounted Kernel Debug File System.
[ 62.623430] systemd[1]: Mounted Kernel Trace File System.
[ OK ] Mounted Kernel Trace File System.
[ 62.673160] systemd[1]: Finished Create list of static device nodes for the current kernel.
[ OK ] Finished Create list of st… nodes for the current kernel.
[ 62.750298] systemd[1]: modprobe@drm.service: Succeeded.
[ 62.825547] systemd[1]: Finished Load Kernel Module drm.
[ OK ] Finished Load Kernel Module drm.
[ 63.196136] systemd[1]: Finished Repartition Root Disk.
[ OK ] Finished Repartition Root Disk.
[ 63.623615] systemd[1]: Finished Apply Kernel Variables.
[ OK ] Finished Apply Kernel Variables.
[ 63.675053] EXT4-fs (vda1): re-mounted. Opts: (null)
[ 63.780488] systemd[1]: Finished Remount Root and Kernel File Systems.
[ OK ] Finished Remount Root and Kernel File Systems.
[ 63.800083] systemd[1]: Condition check resulted in First Boot Wizard being skipped.
[ 63.865295] systemd[1]: Starting Rebuild Hardware Database...
Starting Rebuild Hardware Database...
[ 63.965791] systemd[1]: Starting Load/Save Random Seed...
Starting Load/Save Random Seed...
[ 64.078676] systemd[1]: Starting Create System Users...
Starting Create System Users...
[ 65.374536] systemd[1]: Finished Load/Save Random Seed.
[ OK ] Finished Load/Save Random Seed.
[ 65.389273] kauditd_printk_skb: 10 callbacks suppressed
[ 65.389293] audit: type=1130 audit(1655684405.138:67): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=systemd-random-seed comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
[ 65.681487] systemd[1]: Finished Create System Users.
[ OK ] Finished Create System Users.
[ 65.695950] audit: type=1130 audit(1655684405.444:68): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=systemd-sysusers comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
[ 65.811570] systemd[1]: Starting Create Static Device Nodes in /dev...
Starting Create Static Device Nodes in /dev...
[* ] (1 of 4) A start job is running for… Hardware Database (7s / 1min 33s)
[ 68.034234] systemd[1]: Finished Create Static Device Nodes in /dev.
M
[ OK ] Finished Create Static Device Nodes in /dev.
[ 68.042766] audit: type=1130 audit(1655684407.791:69): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
[ 68.059114] systemd[1]: Reached target Local File Systems (Pre).
[ OK ] Reached target Local File Systems (Pre).
[ 68.072518] systemd[1]: Reached target Local File Systems.
[ OK ] Reached target Local File Systems.
[ 68.198951] systemd[1]: Starting Restore /run/initramfs on shutdown...
Starting Restore /run/initramfs on shutdown...
[ 68.223491] systemd[1]: Condition check resulted in Import network configuration from initramfs being skipped.
[ 68.314356] systemd[1]: Starting Rebuild Dynamic Linker Cache...
Starting Rebuild Dynamic Linker Cache...
[ 68.357516] systemd[1]: Condition check resulted in Mark the need to relabel after reboot being skipped.
[ 68.379277] systemd[1]: Condition check resulted in Store a System Token in an EFI Variable being skipped.
[ 68.545526] systemd[1]: Starting Commit a transient machine-id on disk...
Starting Commit a transient machine-id on disk...
[ 68.642512] systemd[1]: Finished Restore /run/initramfs on shutdown.
[ OK ] Finished Restore /run/initramfs on shutdown.
[ 68.652065] audit: type=1130 audit(1655684408.400:70): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=dracut-shutdown comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
[ 69.566924] systemd[1]: etc-machine\x2did.mount: Succeeded.
[ 69.621642] systemd[1]: Finished Commit a transient machine-id on disk.
[ OK ] Finished Commit a transient machine-id on disk.
[ 69.636569] audit: type=1130 audit(1655684409.385:71): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=systemd-machine-id-commit comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
[ 69.926137] systemd[1]: Finished Rebuild Dynamic Linker Cache.
[ OK ] Finished Rebuild Dynamic Linker Cache.
[ 69.939655] audit: type=1130 audit(1655684409.688:72): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=ldconfig comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
[ 71.162282] systemd[1]: Finished udev Coldplug all Devices.
[ OK ] Finished udev Coldplug all Devices.
[ 71.174820] audit: type=1130 audit(1655684410.922:73): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
[** ] (1 of 2) A start job is running for…Hardware Database (13s / 1min 33s)
M
[*** ] (1 of 2) A start job is running for…Hardware Database (13s / 1min 33s)
M
[ *** ] (2 of 2) A start job is running for Journal Service (14s / 1min 31s)
M
[ *** ] (2 of 2) A start job is running for Journal Service (14s / 1min 31s)
M
[ ***] (2 of 2) A start job is running for Journal Service (15s / 1min 31s)
M
[ **] (1 of 2) A start job is running for…Hardware Database (15s / 1min 33s)
M
[ *] (1 of 2) A start job is running for…Hardware Database (16s / 1min 33s)
M
[ **] (1 of 2) A start job is running for…Hardware Database (16s / 1min 33s)
M
[ ***] (2 of 2) A start job is running for Journal Service (17s / 1min 31s)
[ 77.364384] audit: type=1305 audit(1655684417.112:74): op=set audit_enabled=1 old=1 auid=4294967295 ses=4294967295 subj=system_u:system_r:syslogd_t:s0 res=1
[ 77.791443] systemd[1]: Started Journal Service.
M
[ OK ] Started Journal Service.
[ 77.809454] audit: type=1130 audit(1655684417.558:75): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
Starting Flush Journal to Persistent Storage...
[ 78.684495] systemd-journald[507]: Received client request to flush runtime journal.
[ OK ] Finished Flush Journal to Persistent Storage.
[ 79.346328] audit: type=1130 audit(1655684419.094:76): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=systemd-journal-flush comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
Starting Create Volatile Files and Directories...
[ *** ] (2 of 2) A start job is running for…s and Directories (21s / no limit)
M
[ OK ] Finished Create Volatile Files and Directories.
[ 81.644526] audit: type=1130 audit(1655684421.393:77): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
Starting Security Auditing Service...
Starting Rebuild Journal Catalog...
[ 82.823657] audit: type=1305 audit(1655684422.572:78): op=set audit_enabled=1 old=1 auid=4294967295 ses=4294967295 subj=system_u:system_r:auditd_t:s0 res=1
[ OK ] Finished Rebuild Journal Catalog.
[ *** ] (2 of 2) A start job is running for… Auditing Service (25s / 1min 52s)
M
[*** ] (1 of 2) A start job is running for…Hardware Database (26s / 1min 33s)
M
[ OK ] Started Security Auditing Service.
 Starting Update UTMP about System Boot/Shutdown...
[ OK ] Finished Rebuild Hardware Database.
[ OK ] Finished Update UTMP about System Boot/Shutdown.
Starting udev Kernel Device Manager...
Starting Update is Completed...
[ OK ] Finished Update is Completed.
[** ] A start job is running for udev Kernel Device Manager (29s / 1min 57s)
M
[* ] A start job is running for udev Kernel Device Manager (30s / 1min 57s)
M
[** ] A start job is running for udev Kernel Device Manager (30s / 1min 57s)
M
[*** ] A start job is running for udev Kernel Device Manager (31s / 1min 57s)
M
[ *** ] A start job is running for udev Kernel Device Manager (31s / 1min 57s)
M
[ OK ] Started udev Kernel Device Manager.
[ OK ] Reached target System Initialization.
[ OK ] Started dnf makecache --timer.
[ OK ] Started Discard unused blocks once a week.
[ OK ] Started Daily Cleanup of Temporary Directories.
[ OK ] Started daily update of the root trust anchor for DNSSEC.
[ OK ] Reached target Timers.
[ OK ] Listening on D-Bus System Message Bus Socket.
[ OK ] Listening on SSSD Kerberos…ache Manager responder socket.
[ OK ] Reached target Sockets.
[ OK ] Reached target Basic System.
Starting NTP client/server...
Starting Initial cloud-init job (pre-networking)...
Starting OpenSSH ecdsa Server Key Generation...
Starting OpenSSH ed25519 Server Key Generation...
Starting OpenSSH rsa Server Key Generation...
Starting System Security Services Daemon...
Starting Home Area Manager...
[ OK ] Started NTP client/server.
[ OK ] Finished OpenSSH ecdsa Server Key Generation.
[ OK ] Finished OpenSSH ed25519 Server Key Generation.
[ OK ] Started System Security Services Daemon.
[ OK ] Reached target User and Group Name Lookups.
Starting Login Service...
[ OK ] Finished OpenSSH rsa Server Key Generation.
[ OK ] Reached target sshd-keygen.target.
Starting D-Bus System Message Bus...
[ OK ] Started D-Bus System Message Bus.
[ 155.350644] i801_smbus 0000:00:1f.3: SMBus using PCI interrupt
[ OK ] Started Home Area Manager.
[ OK ] Started Login Service.
[ 175.683511] kvm: Nested Virtualization enabled
[ 175.683703] kvm: Nested Paging enabled
[ 176.049319] Decoding supported only on Scalable MCA processors.
Starting Hostname Service...
[ 183.084313] bochs-drm 0000:00:01.0: vgaarb: deactivate vga console
[ 183.134087] Console: switching to colour dummy device 80x25
[ 183.206934] [drm] Found bochs VGA, ID 0xb0c0.
[ 183.207236] [drm] Framebuffer size 16384 kB @ 0xfb000000, mmio @ 0xfea10000.
[ 183.239507] [TTM] Zone kernel: Available graphics memory: 1013038 KiB
[ 183.239775] [TTM] Initializing pool allocator
[ 183.240872] [TTM] Initializing DMA pool allocator
[ 183.262295] [drm] Found EDID data blob.
[ 183.317152] [drm] Initialized bochs-drm 1.0.0 20130925 for 0000:00:01.0 on minor 0
[ 183.459526] fbcon: bochs-drmdrmfb (fb0) is primary device
[ 183.698445] Console: switching to colour frame buffer device 128x48
[ 183.837045] bochs-drm 0000:00:01.0: fb0: bochs-drmdrmfb frame buffer device
[ OK ] Started Hostname Service.
[ 192.644591] cloud-init[605]: Cloud-init v. 19.4 running 'init-local' at Mon, 20 Jun 2022 00:21:45 +0000. Up 166.25 seconds.
[ OK ] Finished Initial cloud-init job (pre-networking).
[ OK ] Reached target Network (Pre).
Starting Network Manager...
[ OK ] Started Network Manager.
[ OK ] Reached target Network.
Starting Network Manager Wait Online...
Starting Initial cloud-ini… (metadata service crawler)...
Starting Network Manager Script Dispatcher Service...
[ OK ] Started Network Manager Script Dispatcher Service.
[ OK ] Finished Network Manager Wait Online.
[ 211.249669] cloud-init[696]: Cloud-init v. 19.4 running 'init' at Mon, 20 Jun 2022 00:22:28 +0000. Up 208.62 seconds.
[ 211.252987] cloud-init[696]: ci-info: ++++++++++++++++++++++++++++++++++++Net device info++++++++++++++++++++++++++++++++++++
[ 211.257780] cloud-init[696]: ci-info: +--------+------+--------------------------+-------------+--------+-------------------+
[ 211.264377] cloud-init[696]: ci-info: | Device | Up | Address | Mask | Scope | Hw-Address |
[ 211.273743] cloud-init[696]: ci-info: +--------+------+--------------------------+-------------+--------+-------------------+
[ 211.276902] cloud-init[696]: ci-info: | eth0 | True | 172.17.0.16 | 255.255.0.0 | global | 02:42:ac:11:00:10 |
[ 211.283679] cloud-init[696]: ci-info: | eth0 | True | fe80::42:acff:fe11:10/64 | . | link | 02:42:ac:11:00:10 |
[ 211.286122] cloud-init[696]: ci-info: | lo | True | 127.0.0.1 | 255.0.0.0 | host | . |
[ 211.294989] cloud-init[696]: ci-info: | lo | True | ::1/128 | . | host | . |
[ 211.301436] cloud-init[696]: ci-info: +--------+------+--------------------------+-------------+--------+-------------------+
[ 211.305397] cloud-init[696]: ci-info: +++++++++++++++++++++++++++Route IPv4 info++++++++++++++++++++++++++++
[ 211.314464] cloud-init[696]: ci-info: +-------+-------------+------------+-------------+-----------+-------+
[ 211.322486] cloud-init[696]: ci-info: | Route | Destination | Gateway | Genmask | Interface | Flags |
[ 211.330524] cloud-init[696]: ci-info: +-------+-------------+------------+-------------+-----------+-------+
[ 211.334003] cloud-init[696]: ci-info: | 0 | 0.0.0.0 | 172.17.0.1 | 0.0.0.0 | eth0 | UG |
[ 211.340915] cloud-init[696]: ci-info: | 1 | 172.17.0.0 | 0.0.0.0 | 255.255.0.0 | eth0 | U |
[ 211.347664] cloud-init[696]: ci-info: +-------+-------------+------------+-------------+-----------+-------+
[ 211.353410] cloud-init[696]: ci-info: +++++++++++++++++++Route IPv6 info+++++++++++++++++++
[ 211.357932] cloud-init[696]: ci-info: +-------+-------------+---------+-----------+-------+
[ 211.366702] cloud-init[696]: ci-info: | Route | Destination | Gateway | Interface | Flags |
[ 211.373357] cloud-init[696]: ci-info: +-------+-------------+---------+-----------+-------+
[ 211.376593] cloud-init[696]: ci-info: | 1 | fe80::/64 | :: | eth0 | U |
[ 211.382974] cloud-init[696]: ci-info: | 3 | local | :: | eth0 | U |
[ 211.386883] cloud-init[696]: ci-info: | 4 | ff00::/8 | :: | eth0 | U |
[ 211.393891] cloud-init[696]: ci-info: +-------+-------------+---------+-----------+-------+
[ 228.053205] cloud-init[696]: Generating public/private rsa key pair.
[ 228.056391] cloud-init[696]: Your identification has been saved in /etc/ssh/ssh_host_rsa_key
[ 228.058211] cloud-init[696]: Your public key has been saved in /etc/ssh/ssh_host_rsa_key.pub
[ 228.059580] cloud-init[696]: The key fingerprint is:
[ 228.060906] cloud-init[696]: SHA256:6XzsvI9ulumJLWDV361hGT8xZW7UyA5U4K+zEGolFIs root@instance
[ 228.066589] cloud-init[696]: The key's randomart image is:
[ 228.070742] cloud-init[696]: +---[RSA 3072]----+
[ 228.072383] cloud-init[696]: | . .o+...|
[ 228.078655] cloud-init[696]: | . o .. o =|
[ 228.086587] cloud-init[696]: | E o. .o +.|
[ 228.089644] cloud-init[696]: | .... .ooo|
[ 228.096450] cloud-init[696]: | .S o. ..*o|
[ 228.098011] cloud-init[696]: | oo = ...=.o|
[ 228.102917] cloud-init[696]: | . .= +oo. o.|
[ 228.113240] cloud-init[696]: | ..*=+ o. |
[ 228.115570] cloud-init[696]: | .*Xoo |
[ 228.125228] cloud-init[696]: +----[SHA256]-----+
[ 228.126985] cloud-init[696]: Generating public/private dsa key pair.
[ 228.128208] cloud-init[696]: Your identification has been saved in /etc/ssh/ssh_host_dsa_key
[ 228.134466] cloud-init[696]: Your public key has been saved in /etc/ssh/ssh_host_dsa_key.pub
[ 228.142757] cloud-init[696]: The key fingerprint is:
[ 228.150589] cloud-init[696]: SHA256:TphuMvMaRDyebSBD8MmbnfqPBcnGnPI0jpLGlF9Hseo root@instance
[ 228.152614] cloud-init[696]: The key's randomart image is:
[ 228.160591] cloud-init[696]: +---[DSA 1024]----+
[ 228.163855] cloud-init[696]: |... . |
[ 228.170796] cloud-init[696]: | + o o |
[ 228.172422] cloud-init[696]: | * = o |
[ 228.175751] cloud-init[696]: | .% Ooo |
[ 228.182588] cloud-init[696]: | o+ ^o=.S |
[ 228.184749] cloud-init[696]: |o..Xo=.o |
[ 228.188936] cloud-init[696]: |ooo.BE+ . |
[ 228.195169] cloud-init[696]: |.. . X |
[ 228.196702] cloud-init[696]: | +oo |
[ 228.205330] cloud-init[696]: +----[SHA256]-----+
[ 228.208756] cloud-init[696]: Generating public/private ecdsa key pair.
[ 228.215630] cloud-init[696]: Your identification has been saved in /etc/ssh/ssh_host_ecdsa_key
[ 228.220951] cloud-init[696]: Your public key has been saved in /etc/ssh/ssh_host_ecdsa_key.pub
[ 228.231922] cloud-init[696]: The key fingerprint is:
[ 228.233762] cloud-init[696]: SHA256:ddm+K9kzJLXQf07Ev01eZPBb9SXPefu+LH7YhxESuAQ root@instance
[ 228.241696] cloud-init[696]: The key's randomart image is:
[ 228.243487] cloud-init[696]: +---[ECDSA 256]---+
[ 228.247896] cloud-init[696]: | E. . |
[ 228.255854] cloud-init[696]: | o .oo o|
[ 228.257762] cloud-init[696]: | ...oo.B=|
[ 228.277929] cloud-init[696]: | ...o.+.%|
[ 228.292897] cloud-init[696]: | S +.B*|
[ 228.299672] cloud-init[696]: | . ++B|
[ 228.307614] cloud-init[696]: | =+B*|
[ 228.315523] cloud-init[696]: | oo*=*|
[ 228.317793] cloud-init[696]: | .o+*+|
[ 228.325756] cloud-init[696]: +----[SHA256]-----+
[ 228.327469] cloud-init[696]: Generating public/private ed25519 key pair.
[ 228.332977] cloud-init[696]: Your identification has been saved in /etc/ssh/ssh_host_ed25519_key
[ 228.342000] cloud-init[696]: Your public key has been saved in /etc/ssh/ssh_host_ed25519_key.pub
[ 228.349997] cloud-init[696]: The key fingerprint is:
[ 228.359360] cloud-init[696]: SHA256:KZ0gHgUJF+YjMFqFGvqEuY83YMUdpK+dlfUVRmkzfpQ root@instance
[ 228.362776] cloud-init[696]: The key's randomart image is:
[ 228.367659] cloud-init[696]: +--[ED25519 256]--+
[ 228.374724] cloud-init[696]: |o o+B=. .+. .|
[ 228.378868] cloud-init[696]: |o+.+oo .=.E |
[ 228.385407] cloud-init[696]: |o=o.* o . o.+ |
[ 228.389592] cloud-init[696]: |= .=.= ooo. .. . |
[ 228.396464] cloud-init[696]: | +. ...oS . . |
[ 228.400734] cloud-init[696]: |o.. o o. |
[ 228.408527] cloud-init[696]: |.+ . o |
[ 228.417773] cloud-init[696]: |. + |
[ 228.419556] cloud-init[696]: | . . |
[ 228.420977] cloud-init[696]: +----[SHA256]-----+
[ OK ] Finished Initial cloud-ini…ob (metadata service crawler).
[ OK ] Reached target Cloud-config availability.
[ OK ] Reached target Network is Online.
Starting Apply the settings specified in cloud-config...
Starting OpenSSH server daemon...
Starting Permit User Sessions...
[ OK ] Finished Permit User Sessions.
[ OK ] Started Getty on tty1.
[ OK ] Started Serial Getty on ttyS0.
[ OK ] Reached target Login Prompts.
[ OK ] Started OpenSSH server daemon.
[ OK ] Reached target Multi-User System.
Starting Update UTMP about System Runlevel Changes...
[ OK ] Finished Update UTMP about System Runlevel Changes.
Fedora 32 (Cloud Edition)
Kernel 5.6.6-300.fc32.x86_64 on an x86_64 (ttyS0)
SSH host key: SHA256:kjxnLTfBHJQLAcnMjXRITIA08S0J79ZvSXrW0v4OA6w (RSA)
SSH host key: SHA256:Twuyhfy53eQToF+rt/loDZlCR9aXpkvgbw1INRq2aHc (ECDSA)
SSH host key: SHA256:IsYIQTIPPrlv+yomMB8FKR1fq7Z68gbp42+29e4OJS4 (ED25519)
eth0: 172.17.0.16 fe80::42:acff:fe11:10
instance login: [ 250.922800] cloud-init[763]: Cloud-init v. 19.4 running 'modules:config' at Mon, 20 Jun 2022 00:23:02 +0000. Up 242.91 seconds.
ci-info: +++Authorized keys from /home/molecule/.ssh/authorized_keys for user molecule++++
ci-info: +---------+-------------------------------------------------+---------+---------+
ci-info: | Keytype | Fingerprint (md5) | Options | Comment |
ci-info: +---------+-------------------------------------------------+---------+---------+
ci-info: | ssh-rsa | 77:db:f2:17:d6:4f:8e:a8:19:3f:07:1e:f1:eb:40:0d | - | - |
ci-info: +---------+-------------------------------------------------+---------+---------+
<14>Jun 20 00:23:34 ec2:
<14>Jun 20 00:23:35 ec2: #############################################################
<14>Jun 20 00:23:35 ec2: -----BEGIN SSH HOST KEY FINGERPRINTS-----
<14>Jun 20 00:23:35 ec2: 1024 SHA256:TphuMvMaRDyebSBD8MmbnfqPBcnGnPI0jpLGlF9Hseo root@instance (DSA)
<14>Jun 20 00:23:35 ec2: 256 SHA256:ddm+K9kzJLXQf07Ev01eZPBb9SXPefu+LH7YhxESuAQ root@instance (ECDSA)
<14>Jun 20 00:23:36 ec2: 256 SHA256:KZ0gHgUJF+YjMFqFGvqEuY83YMUdpK+dlfUVRmkzfpQ root@instance (ED25519)
<14>Jun 20 00:23:36 ec2: 3072 SHA256:6XzsvI9ulumJLWDV361hGT8xZW7UyA5U4K+zEGolFIs root@instance (RSA)
<14>Jun 20 00:23:36 ec2: -----END SSH HOST KEY FINGERPRINTS-----
<14>Jun 20 00:23:36 ec2: #############################################################
-----BEGIN SSH HOST KEY KEYS-----
ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBAifR1N6BXi1FXOnfex6xZoMyJYlZ/+DduZkcX/DKT0FVSDla4eOJsW7eHFJAgVol9DfVnYnni0Rrqq+xn/wW/0= root@instance
ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIB4b1KZNAYqGeGZUTl779GZSMKdCphTac2B4lznUbvxm root@instance
ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDZCfdLBkI388MYx9r3/0l0T/Xvlh4AFWbeCZoTLhcaWZpPxMazf0r1VuxR4MEHKNiuIRJp06+jjA6dvaBEq+v5WhOhwTeGmhoGUfk5CRlUMTWDwK6O/OngHOKi4CpJwX0WrO3ul6Lkb449KtQjHyC15wo1Ju7FHtEJaPirbQfy8IAHY7+KV35pOT2Njb11YLBOsMWOiQipTPExiGbrB7ds445TKSuaKesa1iIcTNNnCCxUcalfeeYeQPuWQjRA9Gry+jlW/qR/TChAfhM/OMw9odpjISC/R8zXJkiSjeSks+wn//WwsYTxd+SW0m2s5ErS3rP/gm156oYzX5KMJ5B1tmWniMXaX7BKYJKqb+L6pT94HWGlDjAmJ5mn8qWuWw21fBaNIvUymZoc93//Qsh9DfspPHWpPQ8IBkoV77HVY2zXz9Z8xJl79LUe9LLHgDIL/EzVjNeF0tl3L3XUPEpUELjpmDcXuexHLZom2EUPwAwz6BDEurmCo2IoK0y5iTE= root@instance
-----END SSH HOST KEY KEYS-----
[ 278.747618] cloud-init[836]: Cloud-init v. 19.4 running 'modules:final' at Mon, 20 Jun 2022 00:23:29 +0000. Up 269.34 seconds.
[ 278.846704] cloud-init[836]: Cloud-init v. 19.4 finished at Mon, 20 Jun 2022 00:23:37 +0000. Datasource DataSourceNoCloud [seed=/dev/vdb][dsmode=net]. Up 277.52 seconds
You were disconnected from the console. This has one of the following reasons:
- another user connected to the console of the target vm
- network issues
websocket: close 1006 (abnormal closure): unexpected EOF
Script done, file is typescript
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","hostname":"minikube","level":"info","pos":"virt-handler.go:194","timestamp":"2022-06-20T00:14:46.839928Z"}
kubevirt/virt-handler-49ttn[virt-handler]: W0620 00:14:46.842013 7737 client_config.go:617] Neither --kubeconfig nor --master was specified. Using the inClusterConfig. This might not work.
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"set verbosity to 2","pos":"virt-handler.go:453","timestamp":"2022-06-20T00:14:46.857054Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"set verbosity to 2","pos":"virt-handler.go:453","timestamp":"2022-06-20T00:14:46.857222Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"setting rate limiter to 5 QPS and 10 Burst","pos":"virt-handler.go:462","timestamp":"2022-06-20T00:14:46.857326Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"CPU features of a minimum baseline CPU model: map[apic:true clflush:true cmov:true cx16:true cx8:true de:true fpu:true fxsr:true lahf_lm:true lm:true mca:true mce:true mmx:true msr:true mtrr:true nx:true pae:true pat:true pge:true pni:true pse:true pse36:true sep:true sse:true sse2:true sse4.1:true ssse3:true syscall:true tsc:true]","pos":"cpu_plugin.go:96","timestamp":"2022-06-20T00:14:46.897739Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"warning","msg":"host model mode is expected to contain only one model","pos":"cpu_plugin.go:103","timestamp":"2022-06-20T00:14:46.897777Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"error","msg":"something happened during opening kvm file: open /dev/kvm: no such file or directory","pos":"kvm-caps-info-plugin_amd64.go:226","timestamp":"2022-06-20T00:14:46.898461Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"node-labeller is running","pos":"node_labeller.go:94","timestamp":"2022-06-20T00:14:46.898580Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"set verbosity to 2","pos":"virt-handler.go:453","timestamp":"2022-06-20T00:14:46.898719Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"setting rate limiter to 5 QPS and 10 Burst","pos":"virt-handler.go:462","timestamp":"2022-06-20T00:14:46.898748Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"Starting domain stats collector: node name=minikube","pos":"prometheus.go:446","timestamp":"2022-06-20T00:14:46.899992Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"metrics: max concurrent requests=3","pos":"virt-handler.go:477","timestamp":"2022-06-20T00:14:46.900244Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"certificate with common name 'kubevirt.io:system:client:virt-handler' retrieved.","pos":"cert-manager.go:198","timestamp":"2022-06-20T00:14:46.900936Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"STARTING informer vmiInformer-sources","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:14:46.900340Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"STARTING informer vmiInformer-targets","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:14:46.901164Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"STARTING informer extensionsKubeVirtCAConfigMapInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:14:46.901200Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"STARTING informer CRDInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:14:46.901260Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"STARTING informer kubeVirtInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:14:46.901294Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"certificate with common name 'kubevirt.io:system:node:virt-handler' retrieved.","pos":"cert-manager.go:198","timestamp":"2022-06-20T00:14:46.908244Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"Updating cluster config from KubeVirt to resource version '1485'","pos":"configuration.go:320","timestamp":"2022-06-20T00:14:46.989055Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"set verbosity to 2","pos":"virt-handler.go:453","timestamp":"2022-06-20T00:14:46.989316Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"setting rate limiter to 5 QPS and 10 Burst","pos":"virt-handler.go:462","timestamp":"2022-06-20T00:14:46.989551Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"set verbosity to 2","pos":"virt-handler.go:453","timestamp":"2022-06-20T00:14:47.098928Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"setting rate limiter to 5 QPS and 10 Burst","pos":"virt-handler.go:462","timestamp":"2022-06-20T00:14:47.099140Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"set verbosity to 2","pos":"virt-handler.go:453","timestamp":"2022-06-20T00:14:47.099300Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"setting rate limiter to 5 QPS and 10 Burst","pos":"virt-handler.go:462","timestamp":"2022-06-20T00:14:47.099368Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"Starting virt-handler controller.","pos":"vm.go:1298","timestamp":"2022-06-20T00:14:47.190673Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"Starting a device plugin for device: kvm","pos":"device_controller.go:56","timestamp":"2022-06-20T00:14:47.190797Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"Starting a device plugin for device: tun","pos":"device_controller.go:56","timestamp":"2022-06-20T00:14:47.190827Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"Starting a device plugin for device: vhost-net","pos":"device_controller.go:56","timestamp":"2022-06-20T00:14:47.190842Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"Starting a device plugin for device: sev","pos":"device_controller.go:56","timestamp":"2022-06-20T00:14:47.190856Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"refreshed device plugins for permitted/forbidden host devices","pos":"device_controller.go:292","timestamp":"2022-06-20T00:14:47.190881Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"enabled device-plugins for: []","pos":"device_controller.go:293","timestamp":"2022-06-20T00:14:47.190895Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"disabled device-plugins for: []","pos":"device_controller.go:294","timestamp":"2022-06-20T00:14:47.190905Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"refreshed device plugins for permitted/forbidden host devices","pos":"device_controller.go:292","timestamp":"2022-06-20T00:14:47.190923Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"enabled device-plugins for: []","pos":"device_controller.go:293","timestamp":"2022-06-20T00:14:47.190935Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"disabled device-plugins for: []","pos":"device_controller.go:294","timestamp":"2022-06-20T00:14:47.190944Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"set verbosity to 2","pos":"virt-handler.go:453","timestamp":"2022-06-20T00:14:47.200265Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"setting rate limiter to 5 QPS and 10 Burst","pos":"virt-handler.go:462","timestamp":"2022-06-20T00:14:47.200506Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"set verbosity to 2","pos":"virt-handler.go:453","timestamp":"2022-06-20T00:14:47.214510Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"setting rate limiter to 5 QPS and 10 Burst","pos":"virt-handler.go:462","timestamp":"2022-06-20T00:14:47.214573Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"kvm device plugin started","pos":"generic_device.go:158","timestamp":"2022-06-20T00:14:47.216899Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"warning","msg":"device '/dev/kvm' is not present, the device plugin can't expose it.","pos":"generic_device.go:304","timestamp":"2022-06-20T00:14:47.217138Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"sev device plugin started","pos":"generic_device.go:158","timestamp":"2022-06-20T00:14:47.228590Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"warning","msg":"device '/dev/sev' is not present, the device plugin can't expose it.","pos":"generic_device.go:304","timestamp":"2022-06-20T00:14:47.228847Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"device '/dev/sev' is present.","pos":"generic_device.go:307","timestamp":"2022-06-20T00:14:47.228885Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"tun device plugin started","pos":"generic_device.go:158","timestamp":"2022-06-20T00:14:47.231777Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"device '/dev/net/tun' is present.","pos":"generic_device.go:307","timestamp":"2022-06-20T00:14:47.231949Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"device '/dev/kvm' is present.","pos":"generic_device.go:307","timestamp":"2022-06-20T00:14:47.233278Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"vhost-net device plugin started","pos":"generic_device.go:158","timestamp":"2022-06-20T00:14:47.304835Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"device '/dev/vhost-net' is present.","pos":"generic_device.go:307","timestamp":"2022-06-20T00:14:47.305142Z"}
kubevirt/virt-handler-49ttn[virt-handler]: W0620 00:15:00.289516 7737 client_config.go:617] Neither --kubeconfig nor --master was specified. Using the inClusterConfig. This might not work.
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"Updating cluster config from KubeVirt to resource version '1805'","pos":"configuration.go:320","timestamp":"2022-06-20T00:15:20.407939Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"setting rate limiter to 5 QPS and 10 Burst","pos":"virt-handler.go:462","timestamp":"2022-06-20T00:15:20.409130Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"set verbosity to 2","pos":"virt-handler.go:453","timestamp":"2022-06-20T00:15:20.409212Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"refreshed device plugins for permitted/forbidden host devices","pos":"device_controller.go:292","timestamp":"2022-06-20T00:15:20.408631Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"enabled device-plugins for: []","pos":"device_controller.go:293","timestamp":"2022-06-20T00:15:20.409494Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"disabled device-plugins for: []","pos":"device_controller.go:294","timestamp":"2022-06-20T00:15:20.409548Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"Updating cluster config from KubeVirt to resource version '1823'","pos":"configuration.go:320","timestamp":"2022-06-20T00:15:25.444165Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"refreshed device plugins for permitted/forbidden host devices","pos":"device_controller.go:292","timestamp":"2022-06-20T00:15:25.450904Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"enabled device-plugins for: []","pos":"device_controller.go:293","timestamp":"2022-06-20T00:15:25.452892Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"disabled device-plugins for: []","pos":"device_controller.go:294","timestamp":"2022-06-20T00:15:25.452968Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"setting rate limiter to 5 QPS and 10 Burst","pos":"virt-handler.go:462","timestamp":"2022-06-20T00:15:25.451671Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"set verbosity to 2","pos":"virt-handler.go:453","timestamp":"2022-06-20T00:15:25.456420Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"Generic Allocate: resourceName: tun","pos":"generic_device.go:244","timestamp":"2022-06-20T00:18:31.856150Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"Generic Allocate: request: [\u0026ContainerAllocateRequest{DevicesIDs:[tun11],}]","pos":"generic_device.go:245","timestamp":"2022-06-20T00:18:31.856249Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"VMI is in phase: Scheduled | Domain does not exist","name":"instance","namespace":"default","pos":"vm.go:1553","timestamp":"2022-06-20T00:18:46.349222Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"migration is block migration because of boot volume","name":"instance","namespace":"default","pos":"vm.go:2215","timestamp":"2022-06-20T00:18:46.349713Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"migration is block migration because of cloudinit volume","name":"instance","namespace":"default","pos":"vm.go:2215","timestamp":"2022-06-20T00:18:46.350156Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"VMI is in phase: Scheduled | Domain does not exist","name":"instance","namespace":"default","pos":"vm.go:1553","timestamp":"2022-06-20T00:18:46.376942Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"migration is block migration because of boot volume","name":"instance","namespace":"default","pos":"vm.go:2215","timestamp":"2022-06-20T00:18:46.377292Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"migration is block migration because of cloudinit volume","name":"instance","namespace":"default","pos":"vm.go:2215","timestamp":"2022-06-20T00:18:46.377324Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"VMI is in phase: Scheduled | Domain does not exist","name":"instance","namespace":"default","pos":"vm.go:1553","timestamp":"2022-06-20T00:18:47.349712Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"Bind mounting container disk at /var/lib/docker/overlay2/9a21a86f77d9475b4c53b9880213d326d4219fcc1c77190a3ef23791ed1da8c7/merged/disk/downloaded to /var/lib/kubelet/pods/c82e2a23-8bb4-45f8-a2b3-c33b5d2618a2/volumes/kubernetes.io~empty-dir/container-disks/disk_0.img","name":"instance","namespace":"default","pos":"mount.go:274","timestamp":"2022-06-20T00:18:47.361175Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"mounting kernel artifacts","name":"instance","namespace":"default","pos":"mount.go:407","timestamp":"2022-06-20T00:18:47.468223Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"kernel boot not defined - nothing to mount","name":"instance","namespace":"default","pos":"mount.go:410","timestamp":"2022-06-20T00:18:47.468286Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"updated MAC for eth0-nic interface: old: 02:42:ac:11:00:10 -\u003e new: 02:42:ac:9b:ab:f8","pos":"common.go:347","timestamp":"2022-06-20T00:18:47.534415Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"[ContextExecutor]: Executing... Switching from original () to desired () context","pos":"context_executor.go:75","timestamp":"2022-06-20T00:18:47.563404Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"[ContextExecutor]: Execution ended successfully","pos":"context_executor.go:90","timestamp":"2022-06-20T00:18:47.627247Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"Created tap device: tap0 in PID: 12563","pos":"common.go:418","timestamp":"2022-06-20T00:18:47.627305Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"Successfully configured tap device: tap0","pos":"common.go:457","timestamp":"2022-06-20T00:18:47.629036Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"Accepted new notify pipe connection for vmi","name":"instance","namespace":"default","pos":"vm.go:348","timestamp":"2022-06-20T00:18:47.938605Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"Domain","level":"info","msg":"Domain is in state Shutoff reason Unknown","name":"instance","namespace":"default","pos":"vm.go:2728","timestamp":"2022-06-20T00:18:47.947302Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"Domain","level":"info","msg":"Domain is in state Paused reason StartingUp","name":"instance","namespace":"default","pos":"vm.go:2758","timestamp":"2022-06-20T00:18:48.226073Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"migration is block migration because of boot volume","name":"instance","namespace":"default","pos":"vm.go:2215","timestamp":"2022-06-20T00:18:48.695717Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"migration is block migration because of cloudinit volume","name":"instance","namespace":"default","pos":"vm.go:2215","timestamp":"2022-06-20T00:18:48.695784Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"VMI is in phase: Scheduled | Domain status: Paused, reason: StartingUp","name":"instance","namespace":"default","pos":"vm.go:1551","timestamp":"2022-06-20T00:18:48.695903Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"migration is block migration because of boot volume","name":"instance","namespace":"default","pos":"vm.go:2215","timestamp":"2022-06-20T00:18:48.696048Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"migration is block migration because of cloudinit volume","name":"instance","namespace":"default","pos":"vm.go:2215","timestamp":"2022-06-20T00:18:48.696089Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"Domain","level":"info","msg":"Domain is in state Running reason Unknown","name":"instance","namespace":"default","pos":"vm.go:2758","timestamp":"2022-06-20T00:18:48.708812Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"VMI is in phase: Running | Domain status: Running, reason: Unknown","name":"instance","namespace":"default","pos":"vm.go:1551","timestamp":"2022-06-20T00:18:48.734405Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"migration is block migration because of boot volume","name":"instance","namespace":"default","pos":"vm.go:2215","timestamp":"2022-06-20T00:18:48.740358Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"migration is block migration because of cloudinit volume","name":"instance","namespace":"default","pos":"vm.go:2215","timestamp":"2022-06-20T00:18:48.740418Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"error","msg":"Updating the VirtualMachineInstance status failed.","name":"instance","namespace":"default","pos":"vm.go:1698","reason":"Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"instance\": the object has been modified; please apply your changes to the latest version and try again","timestamp":"2022-06-20T00:18:48.841012Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"re-enqueuing VirtualMachineInstance default/instance","pos":"vm.go:1344","reason":"Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"instance\": the object has been modified; please apply your changes to the latest version and try again","timestamp":"2022-06-20T00:18:48.841087Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"VMI is in phase: Running | Domain status: Running, reason: Unknown","name":"instance","namespace":"default","pos":"vm.go:1551","timestamp":"2022-06-20T00:18:48.841334Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"migration is block migration because of boot volume","name":"instance","namespace":"default","pos":"vm.go:2215","timestamp":"2022-06-20T00:18:48.853944Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"migration is block migration because of cloudinit volume","name":"instance","namespace":"default","pos":"vm.go:2215","timestamp":"2022-06-20T00:18:48.853998Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"VMI is in phase: Running | Domain status: Running, reason: Unknown","name":"instance","namespace":"default","pos":"vm.go:1551","timestamp":"2022-06-20T00:18:48.876748Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"migration is block migration because of boot volume","name":"instance","namespace":"default","pos":"vm.go:2215","timestamp":"2022-06-20T00:18:48.892174Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"migration is block migration because of cloudinit volume","name":"instance","namespace":"default","pos":"vm.go:2215","timestamp":"2022-06-20T00:18:48.892226Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"CA update in configmap kubevirt/kubevirt-ca detected. Updating from resource version -1 to 1178","pos":"ca-manager.go:96","timestamp":"2022-06-20T00:18:49.263574Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"Websocket connection upgraded","name":"instance","namespace":"default","pos":"console.go:220","timestamp":"2022-06-20T00:18:49.282146Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"Connecting to proc/12563/root/var/run/kubevirt-private/ea7e77f5-c65e-4a05-8537-bb5d108bb5b0/virt-serial0","name":"instance","namespace":"default","pos":"console.go:221","timestamp":"2022-06-20T00:18:49.282200Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"Connected to proc/12563/root/var/run/kubevirt-private/ea7e77f5-c65e-4a05-8537-bb5d108bb5b0/virt-serial0","name":"instance","namespace":"default","pos":"console.go:231","timestamp":"2022-06-20T00:18:49.284981Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"resyncing virt-launcher domains","pos":"cache.go:384","timestamp":"2022-06-20T00:19:46.986808Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"VMI is in phase: Running | Domain status: Running, reason: Unknown","name":"instance","namespace":"default","pos":"vm.go:1551","timestamp":"2022-06-20T00:19:46.994102Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"migration is block migration because of boot volume","name":"instance","namespace":"default","pos":"vm.go:2215","timestamp":"2022-06-20T00:19:47.011600Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"migration is block migration because of cloudinit volume","name":"instance","namespace":"default","pos":"vm.go:2215","timestamp":"2022-06-20T00:19:47.011650Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"VMI is in phase: Running | Domain status: Running, reason: Unknown","name":"instance","namespace":"default","pos":"vm.go:1551","timestamp":"2022-06-20T00:24:17.803075Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"ACPI feature not available, killing deleted VirtualMachineInstance instance","name":"instance","namespace":"default","pos":"vm.go:2006","timestamp":"2022-06-20T00:24:17.808827Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"error","msg":"error encountered reading from unix socket","name":"instance","namespace":"default","pos":"console.go:236","timestamp":"2022-06-20T00:24:17.897751Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"error","msg":"error encountered reading from client (virt-api) websocket","name":"instance","namespace":"default","pos":"console.go:242","reason":"read tcp 172.17.0.13:8186-\u003e172.17.0.9:51042: use of closed network connection","timestamp":"2022-06-20T00:24:17.898002Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"migration is block migration because of boot volume","name":"instance","namespace":"default","pos":"vm.go:2215","timestamp":"2022-06-20T00:24:18.038953Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"migration is block migration because of cloudinit volume","name":"instance","namespace":"default","pos":"vm.go:2215","timestamp":"2022-06-20T00:24:18.039028Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"VMI is in phase: Running | Domain status: Running, reason: Unknown","name":"instance","namespace":"default","pos":"vm.go:1551","timestamp":"2022-06-20T00:24:18.039191Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"ACPI feature not available, killing deleted VirtualMachineInstance instance","name":"instance","namespace":"default","pos":"vm.go:2006","timestamp":"2022-06-20T00:24:18.040386Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"migration is block migration because of boot volume","name":"instance","namespace":"default","pos":"vm.go:2215","timestamp":"2022-06-20T00:24:18.057000Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"migration is block migration because of cloudinit volume","name":"instance","namespace":"default","pos":"vm.go:2215","timestamp":"2022-06-20T00:24:18.057074Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"Domain","level":"info","msg":"Domain is in state Shutoff reason Destroyed","name":"instance","namespace":"default","pos":"vm.go:2758","timestamp":"2022-06-20T00:24:18.061789Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"VMI is in phase: Running | Domain status: Shutoff, reason: Destroyed","name":"instance","namespace":"default","pos":"vm.go:1551","timestamp":"2022-06-20T00:24:18.062112Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"Signaled deletion for instance","name":"instance","namespace":"default","pos":"vm.go:2074","timestamp":"2022-06-20T00:24:18.074703Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"migration is block migration because of boot volume","name":"instance","namespace":"default","pos":"vm.go:2215","timestamp":"2022-06-20T00:24:18.095673Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"migration is block migration because of cloudinit volume","name":"instance","namespace":"default","pos":"vm.go:2215","timestamp":"2022-06-20T00:24:18.095755Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"Domain","level":"info","msg":"Domain is in state NoState reason NonExistent","name":"instance","namespace":"default","pos":"vm.go:2758","timestamp":"2022-06-20T00:24:18.099194Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"Domain","level":"info","msg":"Domain is marked for deletion","name":"instance","namespace":"default","pos":"vm.go:2762","timestamp":"2022-06-20T00:24:18.099254Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"VMI is in phase: Succeeded | Domain does not exist","name":"instance","namespace":"default","pos":"vm.go:1553","timestamp":"2022-06-20T00:24:18.151647Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"Performing final local cleanup for vmi with uid ea7e77f5-c65e-4a05-8537-bb5d108bb5b0","name":"instance","namespace":"default","pos":"vm.go:1809","timestamp":"2022-06-20T00:24:18.151727Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"Found container disk mount entries","name":"instance","namespace":"default","pos":"mount.go:361","timestamp":"2022-06-20T00:24:18.151782Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"Looking to see if containerdisk is mounted at path /var/lib/kubelet/pods/c82e2a23-8bb4-45f8-a2b3-c33b5d2618a2/volumes/kubernetes.io~empty-dir/container-disks/disk_0.img","name":"instance","namespace":"default","pos":"mount.go:364","timestamp":"2022-06-20T00:24:18.151850Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"unmounting container disk at path /var/lib/kubelet/pods/c82e2a23-8bb4-45f8-a2b3-c33b5d2618a2/volumes/kubernetes.io~empty-dir/container-disks/disk_0.img","name":"instance","namespace":"default","pos":"mount.go:368","timestamp":"2022-06-20T00:24:18.152025Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"Cleaning up remaining hotplug volumes","name":"instance","namespace":"default","pos":"mount.go:708","timestamp":"2022-06-20T00:24:18.225433Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"Domain","level":"info","msg":"Removing domain from cache during final cleanup","name":"instance","namespace":"default","pos":"vm.go:1843","timestamp":"2022-06-20T00:24:18.226252Z","uid":""}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"VMI is in phase: Succeeded | Domain does not exist","name":"instance","namespace":"default","pos":"vm.go:1553","timestamp":"2022-06-20T00:24:18.226348Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"Performing final local cleanup for vmi with uid ea7e77f5-c65e-4a05-8537-bb5d108bb5b0","name":"instance","namespace":"default","pos":"vm.go:1809","timestamp":"2022-06-20T00:24:18.226381Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"No container disk mount entries found to unmount","name":"instance","namespace":"default","pos":"mount.go:357","timestamp":"2022-06-20T00:24:18.226431Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"Cleaning up remaining hotplug volumes","name":"instance","namespace":"default","pos":"mount.go:708","timestamp":"2022-06-20T00:24:18.226450Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"Domain","level":"info","msg":"Removing domain from cache during final cleanup","name":"instance","namespace":"default","pos":"vm.go:1843","timestamp":"2022-06-20T00:24:18.226498Z","uid":""}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"closing notify pipe listener for vmi","name":"instance","namespace":"default","pos":"vm.go:304","timestamp":"2022-06-20T00:24:18.226540Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"","level":"info","msg":"gracefully closed notify pipe connection for vmi","name":"instance","namespace":"default","pos":"vm.go:365","timestamp":"2022-06-20T00:24:19.105285Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"VMI does not exist | Domain does not exist","pos":"vm.go:1558","timestamp":"2022-06-20T00:24:20.138091Z"}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"VirtualMachineInstance","level":"info","msg":"Performing final local cleanup for vmi with uid ","name":"instance","namespace":"default","pos":"vm.go:1809","timestamp":"2022-06-20T00:24:20.138151Z","uid":""}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","kind":"Domain","level":"info","msg":"Removing domain from cache during final cleanup","name":"instance","namespace":"default","pos":"vm.go:1843","timestamp":"2022-06-20T00:24:20.138185Z","uid":""}
kubevirt/virt-handler-49ttn[virt-handler]: {"component":"virt-handler","level":"info","msg":"resyncing virt-launcher domains","pos":"cache.go:384","timestamp":"2022-06-20T00:24:46.984165Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: W0620 00:12:53.960037 1 client_config.go:617] Neither --kubeconfig nor --master was specified. Using the inClusterConfig. This might not work.
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: W0620 00:12:53.960992 1 client_config.go:617] Neither --kubeconfig nor --master was specified. Using the inClusterConfig. This might not work.
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"we are on kubernetes","pos":"application.go:236","timestamp":"2022-06-20T00:12:54.027930Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"servicemonitor is not defined","pos":"application.go:252","timestamp":"2022-06-20T00:12:54.050515Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: I0620 00:12:55.179831 1 request.go:665] Waited for 1.126192874s due to client-side throttling, not priority and fairness, request: GET:https://10.96.0.1:443/apis/flowcontrol.apiserver.k8s.io/v1beta2?timeout=32s
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"prometheusrule is not defined","pos":"application.go:267","timestamp":"2022-06-20T00:12:56.183102Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Operator image: quay.io/kubevirt/virt-operator:v0.52.0","pos":"application.go:281","timestamp":"2022-06-20T00:12:56.671403Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer operatorMutatingWebhookInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.672136Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer FakeOperatorPrometheusRuleInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.672880Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer CRDInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.672895Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer kubeVirtInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.672906Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer OperatorClusterRoleInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.672918Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer operatorPodsInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.711391Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer operatorPodDisruptionBudgetInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.712795Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer FakeOperatorSCC","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.712813Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer OperatorRoleBindingInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.712825Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer OperatorServiceInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.712839Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer OperatorRoleInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.712849Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer OperatorCRDInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.712867Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer operatorValidatingWebhookInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.712888Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer namespaceInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.712897Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer extensionsConfigMapInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.712912Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer OperatorServiceAccountInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.712930Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer OperatorClusterRoleBindingInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.712939Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer operatorAPIServiceInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.712954Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer installStrategyConfigMapInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.712970Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer installStrategyJobsInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.712980Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer secretsInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.712995Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer OperatorConfigMapInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.713011Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer FakeOperatorServiceMonitor","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.713028Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer OperatorDeploymentInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.713039Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer OperatorDaemonSetInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.713052Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Attempting to acquire leader status","pos":"application.go:406","timestamp":"2022-06-20T00:12:56.813591Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: I0620 00:12:56.813648 1 leaderelection.go:248] attempting to acquire leader lease kubevirt/virt-operator...
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: I0620 00:12:56.825213 1 leaderelection.go:258] successfully acquired lease kubevirt/virt-operator
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Started leading","pos":"application.go:386","timestamp":"2022-06-20T00:12:56.835049Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Starting KubeVirt controller.","pos":"kubevirt.go:545","timestamp":"2022-06-20T00:12:56.835107Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Handling KubeVirt resource","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:641","timestamp":"2022-06-20T00:13:06.796381Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Handling deployment","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:995","timestamp":"2022-06-20T00:13:06.796559Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Install strategy config map not loaded. reason: no install strategy configmap found for version v0.52.0 with registry quay.io/kubevirt","pos":"kubevirt.go:879","timestamp":"2022-06-20T00:13:06.797135Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Created job to generate install strategy configmap for version v0.52.0","pos":"kubevirt.go:946","timestamp":"2022-06-20T00:13:06.803694Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Updating cluster config from KubeVirt to resource version '807'","pos":"configuration.go:320","timestamp":"2022-06-20T00:13:06.812402Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Handling KubeVirt resource","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:641","timestamp":"2022-06-20T00:13:11.808316Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Handling deployment","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:995","timestamp":"2022-06-20T00:13:11.808623Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Install strategy config map not loaded. reason: no install strategy configmap found for version v0.52.0 with registry quay.io/kubevirt","pos":"kubevirt.go:879","timestamp":"2022-06-20T00:13:11.808903Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"error","msg":"Waiting on install strategy to be posted from job kubevirt-72d62fe25180ebc296d7a30b4ba2508933d9c2fe-jobk2jkc","name":"kubevirt-72d62fe25180ebc296d7a30b4ba2508933d9c2fe-jobk2jkc","namespace":"kubevirt","pos":"kubevirt.go:931","timestamp":"2022-06-20T00:13:11.808929Z","uid":"57b82828-0632-46b0-9a54-c54aaa0ef81f"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Handling KubeVirt resource","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:641","timestamp":"2022-06-20T00:13:18.474801Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Handling deployment","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:995","timestamp":"2022-06-20T00:13:18.474892Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"ServiceAccount loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.488718Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"ServiceAccount loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.488857Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"ServiceAccount loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.488951Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"ClusterRole loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.489253Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"ClusterRole loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.489866Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"ClusterRole loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.490494Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"ClusterRole loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.490930Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"ClusterRole loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.491623Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"ClusterRole loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.492703Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"ClusterRole loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.493106Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"ClusterRoleBinding loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.493397Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"ClusterRoleBinding loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.493555Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"ClusterRoleBinding loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.493704Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"ClusterRoleBinding loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.493850Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"ClusterRoleBinding loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.493995Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Role loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.494191Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Role loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.494331Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"RoleBinding loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.494513Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"RoleBinding loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.494668Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"CustomResourceDefinition loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.542140Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"CustomResourceDefinition loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.558247Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"CustomResourceDefinition loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.622247Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"CustomResourceDefinition loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.683908Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"CustomResourceDefinition loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.685502Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"CustomResourceDefinition loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.686905Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"CustomResourceDefinition loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.719476Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"CustomResourceDefinition loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.720725Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"CustomResourceDefinition loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.721942Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"CustomResourceDefinition loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.723118Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"CustomResourceDefinition loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.750084Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"CustomResourceDefinition loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.751383Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Service loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.751758Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Service loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.751933Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Service loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.752091Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Secret loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.752208Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Secret loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.752285Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Secret loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.752364Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Secret loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.752436Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Secret loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.752512Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Secret loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.752585Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"ValidatingWebhookConfiguration loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.753097Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"ValidatingWebhookConfiguration loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.755260Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"MutatingWebhookConfiguration loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.755904Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"APIService loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.756159Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"APIService loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.756309Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Deployment loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.757483Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Deployment loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.758181Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"DaemonSet loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.759684Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"SecurityContextConstraints loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.760165Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"SecurityContextConstraints loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.760470Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"ConfigMap loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.760613Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"ConfigMap loaded","pos":"strategy.go:712","timestamp":"2022-06-20T00:13:18.760682Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Loaded install strategy for kubevirt version v0.52.0 into cache","pos":"kubevirt.go:875","timestamp":"2022-06-20T00:13:18.760696Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Garbage collected completed install strategy job","name":"kubevirt-72d62fe25180ebc296d7a30b4ba2508933d9c2fe-jobk2jkc","namespace":"kubevirt","pos":"kubevirt.go:795","timestamp":"2022-06-20T00:13:18.772192Z","uid":"57b82828-0632-46b0-9a54-c54aaa0ef81f"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Waiting on deployment virt-api to roll over to latest version","pos":"reconcile.go:293","timestamp":"2022-06-20T00:13:18.777118Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Waiting on deployment virt-controller to roll over to latest version","pos":"reconcile.go:305","timestamp":"2022-06-20T00:13:18.777158Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Waiting on daemonset virt-handler to roll over to latest version","pos":"reconcile.go:317","timestamp":"2022-06-20T00:13:18.777172Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Validation webhook created for image v0.52.0 and registry quay.io/kubevirt","pos":"reconcile.go:409","timestamp":"2022-06-20T00:13:18.982696Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"crd virtualmachineinstances.kubevirt.io created","pos":"crds.go:94","timestamp":"2022-06-20T00:13:19.317743Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"crd virtualmachineinstancepresets.kubevirt.io created","pos":"crds.go:94","timestamp":"2022-06-20T00:13:19.536710Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"crd virtualmachineinstancereplicasets.kubevirt.io created","pos":"crds.go:94","timestamp":"2022-06-20T00:13:19.880872Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"crd virtualmachines.kubevirt.io created","pos":"crds.go:94","timestamp":"2022-06-20T00:13:20.282100Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"crd virtualmachineinstancemigrations.kubevirt.io created","pos":"crds.go:94","timestamp":"2022-06-20T00:13:20.552499Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"crd virtualmachinesnapshots.snapshot.kubevirt.io created","pos":"crds.go:94","timestamp":"2022-06-20T00:13:20.793458Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"crd virtualmachinesnapshotcontents.snapshot.kubevirt.io created","pos":"crds.go:94","timestamp":"2022-06-20T00:13:21.043232Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"crd virtualmachinerestores.snapshot.kubevirt.io created","pos":"crds.go:94","timestamp":"2022-06-20T00:13:21.224301Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"crd virtualmachineflavors.flavor.kubevirt.io created","pos":"crds.go:94","timestamp":"2022-06-20T00:13:21.268041Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"crd virtualmachineclusterflavors.flavor.kubevirt.io created","pos":"crds.go:94","timestamp":"2022-06-20T00:13:21.319616Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"crd virtualmachinepools.pool.kubevirt.io created","pos":"crds.go:94","timestamp":"2022-06-20T00:13:21.535089Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"crd migrationpolicies.migrations.kubevirt.io created","pos":"crds.go:94","timestamp":"2022-06-20T00:13:21.641089Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"serviceaccount kubevirt-apiserver created","pos":"core.go:447","timestamp":"2022-06-20T00:13:21.673936Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"serviceaccount kubevirt-controller created","pos":"core.go:447","timestamp":"2022-06-20T00:13:21.762045Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"serviceaccount kubevirt-handler created","pos":"core.go:447","timestamp":"2022-06-20T00:13:21.778023Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"ClusterRole kubevirt.io:default created","pos":"rbac.go:59","timestamp":"2022-06-20T00:13:21.867900Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"ClusterRole kubevirt.io:admin created","pos":"rbac.go:59","timestamp":"2022-06-20T00:13:21.912819Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"ClusterRole kubevirt.io:edit created","pos":"rbac.go:59","timestamp":"2022-06-20T00:13:22.099820Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"ClusterRole kubevirt.io:view created","pos":"rbac.go:59","timestamp":"2022-06-20T00:13:22.282788Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"ClusterRole kubevirt-apiserver created","pos":"rbac.go:59","timestamp":"2022-06-20T00:13:22.484042Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"ClusterRole kubevirt-controller created","pos":"rbac.go:59","timestamp":"2022-06-20T00:13:22.984203Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"ClusterRole kubevirt-handler created","pos":"rbac.go:59","timestamp":"2022-06-20T00:13:23.262433Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"ClusterRoleBinding kubevirt.io:default created","pos":"rbac.go:59","timestamp":"2022-06-20T00:13:23.544977Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"ClusterRoleBinding kubevirt-apiserver created","pos":"rbac.go:59","timestamp":"2022-06-20T00:13:24.002859Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"ClusterRoleBinding kubevirt-apiserver-auth-delegator created","pos":"rbac.go:59","timestamp":"2022-06-20T00:13:24.175330Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"ClusterRoleBinding kubevirt-controller created","pos":"rbac.go:59","timestamp":"2022-06-20T00:13:24.244556Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"ClusterRoleBinding kubevirt-handler created","pos":"rbac.go:59","timestamp":"2022-06-20T00:13:24.292473Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Role kubevirt-apiserver created","pos":"rbac.go:59","timestamp":"2022-06-20T00:13:24.349686Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Role kubevirt-handler created","pos":"rbac.go:59","timestamp":"2022-06-20T00:13:24.397545Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"RoleBinding kubevirt-apiserver created","pos":"rbac.go:59","timestamp":"2022-06-20T00:13:24.427911Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"RoleBinding kubevirt-handler created","pos":"rbac.go:59","timestamp":"2022-06-20T00:13:24.472032Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"mutatingwebhoookconfiguration virt-api-mutator created","pos":"admissionregistration.go:311","timestamp":"2022-06-20T00:13:25.005892Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"poddisruptionbudget virt-api-pdb created","pos":"apps.go:351","timestamp":"2022-06-20T00:13:29.982546Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Processed deployment for this round","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:1072","timestamp":"2022-06-20T00:13:29.982618Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Updating cluster config from KubeVirt to resource version '1233'","pos":"configuration.go:320","timestamp":"2022-06-20T00:13:30.027088Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"error","msg":"Could not patch the KubeVirt finalizers.","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:683","reason":"Internal error occurred: failed calling webhook \"kubevirt-update-validator.kubevirt.io\": failed to call webhook: Post \"https://kubevirt-operator-webhook.kubevirt.svc:443/kubevirt-validate-update?timeout=10s\": dial tcp 10.110.244.254:443: i/o timeout","timestamp":"2022-06-20T00:13:40.023724Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"error","msg":"reenqueuing KubeVirt kubevirt/kubevirt","pos":"kubevirt.go:593","reason":"Internal error occurred: failed calling webhook \"kubevirt-update-validator.kubevirt.io\": failed to call webhook: Post \"https://kubevirt-operator-webhook.kubevirt.svc:443/kubevirt-validate-update?timeout=10s\": dial tcp 10.110.244.254:443: i/o timeout","timestamp":"2022-06-20T00:13:40.023786Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Handling KubeVirt resource","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:641","timestamp":"2022-06-20T00:13:40.023836Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Handling deployment","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:995","timestamp":"2022-06-20T00:13:40.023901Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Waiting on deployment virt-api to roll over to latest version","pos":"reconcile.go:293","timestamp":"2022-06-20T00:13:40.026925Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Waiting on deployment virt-controller to roll over to latest version","pos":"reconcile.go:305","timestamp":"2022-06-20T00:13:40.026978Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Waiting on daemonset virt-handler to roll over to latest version","pos":"reconcile.go:317","timestamp":"2022-06-20T00:13:40.026992Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"service kubevirt-prometheus-metrics patched","pos":"core.go:141","timestamp":"2022-06-20T00:13:40.047926Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"service virt-api patched","pos":"core.go:141","timestamp":"2022-06-20T00:13:40.050630Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"service kubevirt-operator-webhook patched","pos":"core.go:141","timestamp":"2022-06-20T00:13:40.053409Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Processed deployment for this round","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:1072","timestamp":"2022-06-20T00:13:40.065709Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Updating cluster config from KubeVirt to resource version '1293'","pos":"configuration.go:320","timestamp":"2022-06-20T00:13:40.085897Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Handling KubeVirt resource","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:641","timestamp":"2022-06-20T00:13:45.024705Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Handling deployment","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:995","timestamp":"2022-06-20T00:13:45.024819Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Waiting on deployment virt-api to roll over to latest version","pos":"reconcile.go:293","timestamp":"2022-06-20T00:13:45.028036Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Waiting on deployment virt-controller to roll over to latest version","pos":"reconcile.go:305","timestamp":"2022-06-20T00:13:45.028075Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Waiting on daemonset virt-handler to roll over to latest version","pos":"reconcile.go:317","timestamp":"2022-06-20T00:13:45.028103Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"service kubevirt-prometheus-metrics patched","pos":"core.go:141","timestamp":"2022-06-20T00:13:45.046171Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"service virt-api patched","pos":"core.go:141","timestamp":"2022-06-20T00:13:45.049913Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"service kubevirt-operator-webhook patched","pos":"core.go:141","timestamp":"2022-06-20T00:13:45.054340Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Processed deployment for this round","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:1072","timestamp":"2022-06-20T00:13:45.065679Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Handling KubeVirt resource","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:641","timestamp":"2022-06-20T00:13:54.422616Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Handling deployment","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:995","timestamp":"2022-06-20T00:13:54.436866Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Waiting on deployment virt-api to roll over to latest version","pos":"reconcile.go:293","timestamp":"2022-06-20T00:13:54.443625Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Waiting on deployment virt-controller to roll over to latest version","pos":"reconcile.go:305","timestamp":"2022-06-20T00:13:54.443660Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Waiting on daemonset virt-handler to roll over to latest version","pos":"reconcile.go:317","timestamp":"2022-06-20T00:13:54.443671Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"service kubevirt-prometheus-metrics patched","pos":"core.go:141","timestamp":"2022-06-20T00:13:54.463862Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"service virt-api patched","pos":"core.go:141","timestamp":"2022-06-20T00:13:54.468022Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"service kubevirt-operator-webhook patched","pos":"core.go:141","timestamp":"2022-06-20T00:13:54.472174Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Processed deployment for this round","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:1072","timestamp":"2022-06-20T00:13:54.554843Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Handling KubeVirt resource","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:641","timestamp":"2022-06-20T00:14:15.071583Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Handling deployment","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:995","timestamp":"2022-06-20T00:14:15.071705Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Waiting on deployment virt-controller to roll over to latest version","pos":"reconcile.go:305","timestamp":"2022-06-20T00:14:15.074901Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Waiting on daemonset virt-handler to roll over to latest version","pos":"reconcile.go:317","timestamp":"2022-06-20T00:14:15.074935Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Temporary blocking validation webhook virt-operator-tmp-webhooksrchc deleted","pos":"delete.go:85","timestamp":"2022-06-20T00:14:15.085903Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"service kubevirt-prometheus-metrics patched","pos":"core.go:141","timestamp":"2022-06-20T00:14:15.107584Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"service virt-api patched","pos":"core.go:141","timestamp":"2022-06-20T00:14:15.111005Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"service kubevirt-operator-webhook patched","pos":"core.go:141","timestamp":"2022-06-20T00:14:15.114144Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"poddisruptionbudget virt-controller-pdb created","pos":"apps.go:351","timestamp":"2022-06-20T00:14:15.146635Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Patching namespace kubevirt with {\"openshift.io/cluster-monitoring\":\"true\"}","pos":"core.go:67","timestamp":"2022-06-20T00:14:15.164759Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"kubevirt namespace labels patched","pos":"core.go:78","timestamp":"2022-06-20T00:14:15.175683Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Processed deployment for this round","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:1072","timestamp":"2022-06-20T00:14:15.181878Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Updating cluster config from KubeVirt to resource version '1485'","pos":"configuration.go:320","timestamp":"2022-06-20T00:14:15.370484Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Handling KubeVirt resource","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:641","timestamp":"2022-06-20T00:14:20.086477Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Handling deployment","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:995","timestamp":"2022-06-20T00:14:20.087645Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Waiting on deployment virt-controller to roll over to latest version","pos":"reconcile.go:305","timestamp":"2022-06-20T00:14:20.103330Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Waiting on daemonset virt-handler to roll over to latest version","pos":"reconcile.go:317","timestamp":"2022-06-20T00:14:20.103385Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"service kubevirt-prometheus-metrics patched","pos":"core.go:141","timestamp":"2022-06-20T00:14:20.173223Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"service virt-api patched","pos":"core.go:141","timestamp":"2022-06-20T00:14:20.336763Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"service kubevirt-operator-webhook patched","pos":"core.go:141","timestamp":"2022-06-20T00:14:20.341523Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Kubevirt namespace (kubevirt) labels are in sync","pos":"core.go:57","timestamp":"2022-06-20T00:14:20.353512Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Processed deployment for this round","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:1072","timestamp":"2022-06-20T00:14:20.353545Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Handling KubeVirt resource","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:641","timestamp":"2022-06-20T00:14:26.610510Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Handling deployment","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:995","timestamp":"2022-06-20T00:14:26.610634Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Waiting on deployment virt-controller to roll over to latest version","pos":"reconcile.go:305","timestamp":"2022-06-20T00:14:26.614027Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Waiting on daemonset virt-handler to roll over to latest version","pos":"reconcile.go:317","timestamp":"2022-06-20T00:14:26.614062Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"service kubevirt-prometheus-metrics patched","pos":"core.go:141","timestamp":"2022-06-20T00:14:26.623731Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"service virt-api patched","pos":"core.go:141","timestamp":"2022-06-20T00:14:26.628303Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"service kubevirt-operator-webhook patched","pos":"core.go:141","timestamp":"2022-06-20T00:14:26.631271Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Kubevirt namespace (kubevirt) labels are in sync","pos":"core.go:57","timestamp":"2022-06-20T00:14:26.656615Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Processed deployment for this round","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:1072","timestamp":"2022-06-20T00:14:26.656654Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Handling KubeVirt resource","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:641","timestamp":"2022-06-20T00:14:38.874846Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Handling deployment","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:995","timestamp":"2022-06-20T00:14:38.874983Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Waiting on daemonset virt-handler to roll over to latest version","pos":"reconcile.go:317","timestamp":"2022-06-20T00:14:38.878146Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"service kubevirt-prometheus-metrics patched","pos":"core.go:141","timestamp":"2022-06-20T00:14:38.887361Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"service virt-api patched","pos":"core.go:141","timestamp":"2022-06-20T00:14:38.891475Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"service kubevirt-operator-webhook patched","pos":"core.go:141","timestamp":"2022-06-20T00:14:38.894089Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Kubevirt namespace (kubevirt) labels are in sync","pos":"core.go:57","timestamp":"2022-06-20T00:14:38.905822Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Processed deployment for this round","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:1072","timestamp":"2022-06-20T00:14:38.905862Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Handling KubeVirt resource","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:641","timestamp":"2022-06-20T00:14:49.080642Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Handling deployment","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:995","timestamp":"2022-06-20T00:14:49.081163Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Waiting on daemonset virt-handler to roll over to latest version","pos":"reconcile.go:317","timestamp":"2022-06-20T00:14:49.084572Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"service kubevirt-prometheus-metrics patched","pos":"core.go:141","timestamp":"2022-06-20T00:14:49.098282Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"service virt-api patched","pos":"core.go:141","timestamp":"2022-06-20T00:14:49.106829Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"service kubevirt-operator-webhook patched","pos":"core.go:141","timestamp":"2022-06-20T00:14:49.110500Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Kubevirt namespace (kubevirt) labels are in sync","pos":"core.go:57","timestamp":"2022-06-20T00:14:49.130736Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Processed deployment for this round","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:1072","timestamp":"2022-06-20T00:14:49.130776Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Handling KubeVirt resource","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:641","timestamp":"2022-06-20T00:15:00.310897Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Handling deployment","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:995","timestamp":"2022-06-20T00:15:00.311082Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Waiting on daemonset virt-handler to roll over to latest version","pos":"reconcile.go:317","timestamp":"2022-06-20T00:15:00.314717Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"service kubevirt-prometheus-metrics patched","pos":"core.go:141","timestamp":"2022-06-20T00:15:00.325264Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"service virt-api patched","pos":"core.go:141","timestamp":"2022-06-20T00:15:00.330324Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"service kubevirt-operator-webhook patched","pos":"core.go:141","timestamp":"2022-06-20T00:15:00.333260Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Kubevirt namespace (kubevirt) labels are in sync","pos":"core.go:57","timestamp":"2022-06-20T00:15:00.344460Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Processed deployment for this round","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:1072","timestamp":"2022-06-20T00:15:00.344494Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Handling KubeVirt resource","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:641","timestamp":"2022-06-20T00:15:20.326846Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Handling deployment","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:995","timestamp":"2022-06-20T00:15:20.326965Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"service kubevirt-prometheus-metrics patched","pos":"core.go:141","timestamp":"2022-06-20T00:15:20.342655Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"service virt-api patched","pos":"core.go:141","timestamp":"2022-06-20T00:15:20.348014Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"service kubevirt-operator-webhook patched","pos":"core.go:141","timestamp":"2022-06-20T00:15:20.355080Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Kubevirt namespace (kubevirt) labels are in sync","pos":"core.go:57","timestamp":"2022-06-20T00:15:20.371880Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"All KubeVirt resources created","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:1060","timestamp":"2022-06-20T00:15:20.379264Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"All KubeVirt components ready","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:1064","timestamp":"2022-06-20T00:15:20.379320Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Updating cluster config from KubeVirt to resource version '1805'","pos":"configuration.go:320","timestamp":"2022-06-20T00:15:20.413226Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Handling KubeVirt resource","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:641","timestamp":"2022-06-20T00:15:25.420068Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Handling deployment","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:995","timestamp":"2022-06-20T00:15:25.420206Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Updating cluster config from KubeVirt to resource version '1823'","pos":"configuration.go:320","timestamp":"2022-06-20T00:15:25.450459Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"service kubevirt-prometheus-metrics patched","pos":"core.go:141","timestamp":"2022-06-20T00:15:25.458068Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"service virt-api patched","pos":"core.go:141","timestamp":"2022-06-20T00:15:25.465327Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"service kubevirt-operator-webhook patched","pos":"core.go:141","timestamp":"2022-06-20T00:15:25.470587Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Kubevirt namespace (kubevirt) labels are in sync","pos":"core.go:57","timestamp":"2022-06-20T00:15:25.488446Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"All KubeVirt resources created","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:1060","timestamp":"2022-06-20T00:15:25.534440Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"All KubeVirt components ready","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:1064","timestamp":"2022-06-20T00:15:25.534514Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Handling KubeVirt resource","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:641","timestamp":"2022-06-20T00:15:30.451860Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"Handling deployment","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:995","timestamp":"2022-06-20T00:15:30.452241Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"service kubevirt-prometheus-metrics patched","pos":"core.go:141","timestamp":"2022-06-20T00:15:30.488250Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"service virt-api patched","pos":"core.go:141","timestamp":"2022-06-20T00:15:30.495500Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"service kubevirt-operator-webhook patched","pos":"core.go:141","timestamp":"2022-06-20T00:15:30.500165Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","level":"info","msg":"Kubevirt namespace (kubevirt) labels are in sync","pos":"core.go:57","timestamp":"2022-06-20T00:15:30.517646Z"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"All KubeVirt resources created","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:1060","timestamp":"2022-06-20T00:15:30.525701Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-operator-564f568975-9kbf5[virt-operator]: {"component":"virt-operator","kind":"","level":"info","msg":"All KubeVirt components ready","name":"kubevirt","namespace":"kubevirt","pos":"kubevirt.go:1064","timestamp":"2022-06-20T00:15:30.526338Z","uid":"5139ce79-3891-4bba-a9bf-6fa86b259028"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: W0620 00:13:56.242396 1 client_config.go:617] Neither --kubeconfig nor --master was specified. Using the inClusterConfig. This might not work.
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: W0620 00:13:56.243379 1 client_config.go:617] Neither --kubeconfig nor --master was specified. Using the inClusterConfig. This might not work.
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"STARTING informer limitrangeInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:13:56.257173Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"STARTING informer vmRestoreInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:13:56.257228Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"STARTING informer kubeVirtInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:13:56.257241Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"STARTING informer extensionsConfigMapInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:13:56.257251Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"STARTING informer extensionsKubeVirtCAConfigMapInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:13:56.257261Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"STARTING informer CRDInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:13:56.257272Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"STARTING informer vmiPresetInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:13:56.257283Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"Waiting for cache sync of informer CRDInformer","pos":"virtinformers.go:317","timestamp":"2022-06-20T00:13:56.257294Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"Waiting for cache sync of informer vmiPresetInformer","pos":"virtinformers.go:317","timestamp":"2022-06-20T00:13:56.257304Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"Waiting for cache sync of informer limitrangeInformer","pos":"virtinformers.go:317","timestamp":"2022-06-20T00:13:56.257313Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"Waiting for cache sync of informer vmRestoreInformer","pos":"virtinformers.go:317","timestamp":"2022-06-20T00:13:56.257322Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"Waiting for cache sync of informer kubeVirtInformer","pos":"virtinformers.go:317","timestamp":"2022-06-20T00:13:56.257331Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"Waiting for cache sync of informer extensionsConfigMapInformer","pos":"virtinformers.go:317","timestamp":"2022-06-20T00:13:56.257339Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"Waiting for cache sync of informer extensionsKubeVirtCAConfigMapInformer","pos":"virtinformers.go:317","timestamp":"2022-06-20T00:13:56.257350Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"CDI detected, DataSource integration enabled","pos":"api.go:932","timestamp":"2022-06-20T00:13:56.658067Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"SKIPPING informer extensionsConfigMapInformer","pos":"virtinformers.go:302","timestamp":"2022-06-20T00:13:56.658134Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"SKIPPING informer extensionsKubeVirtCAConfigMapInformer","pos":"virtinformers.go:302","timestamp":"2022-06-20T00:13:56.658150Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"STARTING informer dataSourceInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:13:56.658159Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"SKIPPING informer kubeVirtInformer","pos":"virtinformers.go:302","timestamp":"2022-06-20T00:13:56.658170Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"SKIPPING informer CRDInformer","pos":"virtinformers.go:302","timestamp":"2022-06-20T00:13:56.658178Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"SKIPPING informer vmiPresetInformer","pos":"virtinformers.go:302","timestamp":"2022-06-20T00:13:56.658186Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"SKIPPING informer limitrangeInformer","pos":"virtinformers.go:302","timestamp":"2022-06-20T00:13:56.658195Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"SKIPPING informer vmRestoreInformer","pos":"virtinformers.go:302","timestamp":"2022-06-20T00:13:56.658203Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"STARTING informer vmFlavorInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:13:56.658212Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"STARTING informer vmClusterFlavorInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:13:56.658222Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"Waiting for cache sync of informer dataSourceInformer","pos":"virtinformers.go:317","timestamp":"2022-06-20T00:13:56.658233Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"Waiting for cache sync of informer extensionsConfigMapInformer","pos":"virtinformers.go:317","timestamp":"2022-06-20T00:13:56.658242Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"Waiting for cache sync of informer extensionsKubeVirtCAConfigMapInformer","pos":"virtinformers.go:317","timestamp":"2022-06-20T00:13:56.658254Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"Waiting for cache sync of informer vmiPresetInformer","pos":"virtinformers.go:317","timestamp":"2022-06-20T00:13:56.658263Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"Waiting for cache sync of informer limitrangeInformer","pos":"virtinformers.go:317","timestamp":"2022-06-20T00:13:56.658272Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"Waiting for cache sync of informer vmRestoreInformer","pos":"virtinformers.go:317","timestamp":"2022-06-20T00:13:56.658280Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"Waiting for cache sync of informer vmFlavorInformer","pos":"virtinformers.go:317","timestamp":"2022-06-20T00:13:56.658288Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"Waiting for cache sync of informer vmClusterFlavorInformer","pos":"virtinformers.go:317","timestamp":"2022-06-20T00:13:56.658297Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"Waiting for cache sync of informer kubeVirtInformer","pos":"virtinformers.go:317","timestamp":"2022-06-20T00:13:56.658306Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"Waiting for cache sync of informer CRDInformer","pos":"virtinformers.go:317","timestamp":"2022-06-20T00:13:56.658315Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"Updating cluster config from KubeVirt to resource version '1293'","pos":"configuration.go:320","timestamp":"2022-06-20T00:13:56.658449Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"set log verbosity to 2","pos":"api.go:993","timestamp":"2022-06-20T00:13:56.658501Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"set log verbosity to 2","pos":"api.go:993","timestamp":"2022-06-20T00:13:56.658519Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"setting rate limiter for the API to 5 QPS and 10 Burst","pos":"api.go:1002","timestamp":"2022-06-20T00:13:56.658550Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"setting rate limiter for webhooks to 200 QPS and 400 Burst","pos":"api.go:1006","timestamp":"2022-06-20T00:13:56.658569Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"set log verbosity to 2","pos":"api.go:993","timestamp":"2022-06-20T00:13:56.658604Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"setting rate limiter for the API to 5 QPS and 10 Burst","pos":"api.go:1002","timestamp":"2022-06-20T00:13:56.658619Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"setting rate limiter for webhooks to 200 QPS and 400 Burst","pos":"api.go:1006","timestamp":"2022-06-20T00:13:56.658629Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"setting rate limiter for the API to 5 QPS and 10 Burst","pos":"api.go:1002","timestamp":"2022-06-20T00:13:56.658673Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"setting rate limiter for webhooks to 200 QPS and 400 Burst","pos":"api.go:1006","timestamp":"2022-06-20T00:13:56.658683Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"set log verbosity to 2","pos":"api.go:993","timestamp":"2022-06-20T00:13:56.658709Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"set log verbosity to 2","pos":"api.go:993","timestamp":"2022-06-20T00:13:56.658769Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"setting rate limiter for the API to 5 QPS and 10 Burst","pos":"api.go:1002","timestamp":"2022-06-20T00:13:56.658781Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"setting rate limiter for webhooks to 200 QPS and 400 Burst","pos":"api.go:1006","timestamp":"2022-06-20T00:13:56.658791Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"certificate with common name 'virt-api.kubevirt.pod.cluster.local' retrieved.","pos":"cert-manager.go:198","timestamp":"2022-06-20T00:13:56.758919Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"certificate with common name 'kubevirt.io:system:client:virt-handler' retrieved.","pos":"cert-manager.go:198","timestamp":"2022-06-20T00:13:56.759127Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"CA update in configmap kube-system/extension-apiserver-authentication detected. Updating from resource version -1 to 40","pos":"ca-manager.go:96","timestamp":"2022-06-20T00:14:10.031033Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: W0620 00:14:10.034371 1 client_config.go:617] Neither --kubeconfig nor --master was specified. Using the inClusterConfig. This might not work.
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:10.036676Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:10.373408Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:10.434277Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:10.436670Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:10.487577Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:10.473545Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:10.474700Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:10.441383Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:10.481178Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:10.478718Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:10.491863Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:10.492165Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:10.493071Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:10.495809Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:10.749319Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":67779,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:11.145492Z","url":"/openapi/v2","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:11.616424Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:13.041176Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:13.918223Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:14.950827Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:15.010372Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"Updating cluster config from KubeVirt to resource version '1485'","pos":"configuration.go:320","timestamp":"2022-06-20T00:14:15.290153Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"setting rate limiter for the API to 5 QPS and 10 Burst","pos":"api.go:1002","timestamp":"2022-06-20T00:14:15.290202Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"setting rate limiter for webhooks to 200 QPS and 400 Burst","pos":"api.go:1006","timestamp":"2022-06-20T00:14:15.290219Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"set log verbosity to 2","pos":"api.go:993","timestamp":"2022-06-20T00:14:15.336241Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:15.345431Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:16.216627Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:17.642884Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:18.815988Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:19.941153Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:20.044929Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:20.816646Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:22.245406Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:23.124889Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:23.652562Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:24.543659Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:26.840837Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:27.527329Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:27.603828Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:27.604926Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:27.609039Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:27.621722Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:29.140367Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:30.039463Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:31.440860Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:33.740478Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:36.044483Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:38.341212Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:40.034355Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:40.540443Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:44.972115Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:45.024412Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:50.035148Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:57.498921Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:57.502654Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:57.517083Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:57.542727Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:00.034381Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:10.042472Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":67779,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:11.225206Z","url":"/openapi/v2","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:15.008036Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:15.036203Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:20.035364Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"Updating cluster config from KubeVirt to resource version '1805'","pos":"configuration.go:320","timestamp":"2022-06-20T00:15:20.418734Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"setting rate limiter for the API to 5 QPS and 10 Burst","pos":"api.go:1002","timestamp":"2022-06-20T00:15:20.418795Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"setting rate limiter for webhooks to 200 QPS and 400 Burst","pos":"api.go:1006","timestamp":"2022-06-20T00:15:20.418811Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"set log verbosity to 2","pos":"api.go:993","timestamp":"2022-06-20T00:15:20.418840Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"Updating cluster config from KubeVirt to resource version '1823'","pos":"configuration.go:320","timestamp":"2022-06-20T00:15:25.442006Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"setting rate limiter for the API to 5 QPS and 10 Burst","pos":"api.go:1002","timestamp":"2022-06-20T00:15:25.442057Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"setting rate limiter for webhooks to 200 QPS and 400 Burst","pos":"api.go:1006","timestamp":"2022-06-20T00:15:25.442073Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"set log verbosity to 2","pos":"api.go:993","timestamp":"2022-06-20T00:15:25.442107Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:27.485126Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:27.494052Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:27.497284Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:27.497622Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:27.504188Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:27.844115Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:30.033880Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:30.143956Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:32.448776Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:34.693513Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:37.044087Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:39.343300Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:40.035306Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:45.066312Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:45.071115Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:50.035130Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:57.483865Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:57.495848Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:57.496104Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:57.503820Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:57.504511Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:57.511561Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:00.036294Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:10.035373Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":67779,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:11.250439Z","url":"/openapi/v2","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:15.111339Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:15.124702Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:20.035769Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:27.534585Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:27.537091Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:27.540571Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:27.540923Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:27.546138Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:30.036204Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:40.034699Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:40.310686Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:42.661552Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:44.911623Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:45.167691Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:45.167939Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:47.215106Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:49.510787Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:50.035191Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:51.861341Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:57.616449Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:57.704063Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:57.714595Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:57.718528Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:00.035144Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:10.034996Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":67779,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:11.284176Z","url":"/openapi/v2","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:15.211662Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:15.231427Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:20.049266Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:27.515864Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:27.518926Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:27.533729Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:27.548442Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:30.035609Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:40.035034Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:45.296240Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:45.311748Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:45.312772Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:50.034715Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:52.751254Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:52.800775Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:55.051879Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:55.101182Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:57.351714Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:57.411762Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:57.504017Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:57.524809Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:57.539274Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:57.557954Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:59.650915Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:59.701448Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:00.034610Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:01.951149Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:02.000972Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:04.251593Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:04.301746Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:10.035090Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":67779,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:11.314950Z","url":"/openapi/v2","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":67779,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:11.430424Z","url":"/openapi/v2","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:15.351675Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:15.352044Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:15.367626Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"http: TLS handshake error from 172.17.0.1:24671: EOF\n","pos":"server.go:3160","timestamp":"2022-06-20T00:18:15.371302Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:20.039677Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:27.505266Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:27.520608Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:27.523031Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:27.530523Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:27.535384Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:27.547787Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:30.038798Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:40.034511Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:45.407362Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:45.429003Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","kind":"VirtualMachineInstance","level":"error","msg":"Unable to establish connection to virt-handler","name":"instance","namespace":"default","pos":"dialers.go:63","reason":"Unable to connect to VirtualMachineInstance because phase is Scheduled instead of Running","timestamp":"2022-06-20T00:18:48.222351Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":225,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":400,"timestamp":"2022-06-20T00:18:48.223186Z","url":"/apis/subresources.kubevirt.io/v1alpha3/namespaces/default/virtualmachineinstances/instance/console","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"CA update in configmap kubevirt/kubevirt-ca detected. Updating from resource version -1 to 1178","pos":"ca-manager.go:96","timestamp":"2022-06-20T00:18:49.276073Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:50.035274Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:57.522357Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:57.573384Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:57.587644Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:57.618641Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:00.036840Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:05.493412Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:07.733255Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:10.051290Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:10.082220Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":67779,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:11.342065Z","url":"/openapi/v2","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:12.394994Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:14.684080Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:15.453738Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:15.487013Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:16.991323Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:20.045052Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:27.510999Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:30.035268Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:40.035246Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:45.495803Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"http: TLS handshake error from 172.17.0.1:39877: EOF\n","pos":"server.go:3160","timestamp":"2022-06-20T00:19:45.573872Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:45.596519Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","level":"info","msg":"http: TLS handshake error from 172.17.0.1:40548: EOF\n","pos":"server.go:3160","timestamp":"2022-06-20T00:19:45.619685Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:50.035153Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:57.544461Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:57.594596Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:57.628038Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:57.629320Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:57.700302Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:00.036559Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:10.035958Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":67779,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:11.394933Z","url":"/openapi/v2","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:15.565129Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:15.623399Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:18.290702Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:20.036200Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:20.540760Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:22.843792Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:24.991211Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:27.440992Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:27.564170Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:27.575748Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:27.596582Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:27.597021Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:27.614128Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:29.790486Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:30.034450Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:40.061239Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:45.636940Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:45.674340Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:50.035683Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:57.606621Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:57.616496Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:57.641140Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:57.646778Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:00.038274Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:10.035874Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":67779,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:11.432727Z","url":"/openapi/v2","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:15.661931Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:15.692842Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:20.095554Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:27.606194Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:27.643103Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:27.649217Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:27.675821Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:27.691572Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:27.707937Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:30.035976Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:30.728866Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:32.869638Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:35.328354Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:37.370077Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:39.919748Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:40.035209Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:42.221421Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:45.695475Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:45.726465Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:50.038088Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:57.475882Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:57.504731Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:57.587941Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:58.084548Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:58.100747Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:58.165703Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:58.186639Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:00.038382Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:10.034734Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":67779,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:11.477365Z","url":"/openapi/v2","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":67779,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:11.710245Z","url":"/openapi/v2","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:15.751754Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:15.755607Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:15.770097Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:15.775666Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:20.047756Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:27.509057Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:27.518401Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:27.537128Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:27.523083Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:27.545032Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:30.035045Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:40.035038Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:43.286504Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:43.335851Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:45.536365Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:45.588264Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:45.855795Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:45.877970Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:45.884558Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:47.837100Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:47.890685Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:50.036722Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:50.188318Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:50.243245Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:52.438086Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:52.487277Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:54.735663Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:54.788676Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:57.568011Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:57.583347Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:57.608420Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:57.647794Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:00.035361Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:10.034635Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":67779,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:11.507281Z","url":"/openapi/v2","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":67779,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:11.737805Z","url":"/openapi/v2","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:15.926133Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:15.954122Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:16.000007Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:20.051873Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:27.502684Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:27.536006Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:30.035833Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:40.038951Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:46.023870Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:46.027487Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:46.064051Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:46.069968Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:50.034681Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:55.736419Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:55.788687Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:57.611951Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:57.671526Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:57.687850Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:57.735607Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:57.744009Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:58.035879Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:58.087036Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:59.989676Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:00.036688Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:00.040987Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:02.635249Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:02.686213Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:04.635315Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:04.685557Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:07.237427Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:07.287137Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:10.036029Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":67779,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:11.564868Z","url":"/openapi/v2","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":67779,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:11.763921Z","url":"/openapi/v2","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:16.074387Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:16.079885Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:16.100222Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:16.104356Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:17.713297Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:17.747672Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:17.751453Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":0,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:17.900446Z","url":"/apis/subresources.kubevirt.io/v1alpha3/namespaces/default/virtualmachineinstances/instance/console","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":223,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:17.923543Z","url":"/apis/subresources.kubevirt.io/v1/version","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:19.657176Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:20.040834Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:21.173458Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:22.348273Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:22.349375Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:27.507004Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:27.516416Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:27.516884Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:27.519691Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:27.543807Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:30.035698Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:32.064850Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:33.453507Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:34.790618Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:40.035028Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:40.512957Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:41.870537Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:43.380437Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:46.105766Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:46.107256Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:46.122330Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:46.122649Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:50.035794Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:54.001538Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:54.063221Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:54.063681Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:55.481804Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:57.621876Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:57.632454Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:57.632658Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:57.633033Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:57.635838Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:57.650557Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:57.662079Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:57.789771Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:59.579089Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:25:00.036653Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:25:00.965432Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:25:02.350857Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:25:03.727426Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:25:03.775082Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:25:03.775125Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:25:05.094794Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:25:06.367544Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:25:08.193383Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:25:08.244277Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:25:08.773923Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:25:08.786169Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: W0620 00:12:53.993679 1 client_config.go:617] Neither --kubeconfig nor --master was specified. Using the inClusterConfig. This might not work.
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: W0620 00:12:53.993903 1 client_config.go:617] Neither --kubeconfig nor --master was specified. Using the inClusterConfig. This might not work.
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: {"component":"virt-operator","level":"info","msg":"we are on kubernetes","pos":"application.go:236","timestamp":"2022-06-20T00:12:54.024103Z"}
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: {"component":"virt-operator","level":"info","msg":"servicemonitor is not defined","pos":"application.go:252","timestamp":"2022-06-20T00:12:54.067776Z"}
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: I0620 00:12:55.195279 1 request.go:665] Waited for 1.123905892s due to client-side throttling, not priority and fairness, request: GET:https://10.96.0.1:443/apis/discovery.k8s.io/v1beta1?timeout=32s
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: {"component":"virt-operator","level":"info","msg":"prometheusrule is not defined","pos":"application.go:267","timestamp":"2022-06-20T00:12:56.196589Z"}
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: {"component":"virt-operator","level":"info","msg":"Operator image: quay.io/kubevirt/virt-operator:v0.52.0","pos":"application.go:281","timestamp":"2022-06-20T00:12:56.942069Z"}
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer OperatorCRDInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.942287Z"}
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer operatorValidatingWebhookInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.942318Z"}
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer kubeVirtInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.942330Z"}
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer OperatorServiceAccountInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.942351Z"}
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer OperatorClusterRoleBindingInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.942359Z"}
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer OperatorServiceInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.942372Z"}
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer installStrategyConfigMapInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.942384Z"}
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer namespaceInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.942395Z"}
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer extensionsConfigMapInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.942404Z"}
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer OperatorClusterRoleInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.942415Z"}
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer OperatorRoleBindingInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.942424Z"}
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer OperatorDeploymentInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.942432Z"}
kubevirt/virt-controller-749d8d99d4-wz5m4[virt-controller]: W0620 00:14:37.603922 1 client_config.go:617] Neither --kubeconfig nor --master was specified. Using the inClusterConfig. This might not work.
kubevirt/virt-controller-749d8d99d4-wz5m4[virt-controller]: {"component":"virt-controller","level":"info","msg":"STARTING informer CRDInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:14:37.604915Z"}
kubevirt/virt-controller-749d8d99d4-wz5m4[virt-controller]: {"component":"virt-controller","level":"info","msg":"STARTING informer kubeVirtInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:14:37.605022Z"}
kubevirt/virt-controller-749d8d99d4-wz5m4[virt-controller]: {"component":"virt-controller","level":"info","msg":"CDI detected, DataVolume integration enabled","pos":"application.go:347","timestamp":"2022-06-20T00:14:37.906174Z"}
kubevirt/virt-controller-749d8d99d4-wz5m4[virt-controller]: W0620 00:14:37.906243 1 client_config.go:617] Neither --kubeconfig nor --master was specified. Using the inClusterConfig. This might not work.
kubevirt/virt-controller-749d8d99d4-wz5m4[virt-controller]: {"component":"virt-controller","level":"info","msg":"Updating cluster config from KubeVirt to resource version '1485'","pos":"configuration.go:320","timestamp":"2022-06-20T00:14:37.906358Z"}
kubevirt/virt-controller-749d8d99d4-wz5m4[virt-controller]: {"component":"virt-controller","level":"info","msg":"setting rate limiter to 20 QPS and 30 Burst","pos":"application.go:398","timestamp":"2022-06-20T00:14:37.906408Z"}
kubevirt/virt-controller-749d8d99d4-wz5m4[virt-controller]: {"component":"virt-controller","level":"info","msg":"set log verbosity to 2","pos":"application.go:405","timestamp":"2022-06-20T00:14:37.906449Z"}
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer operatorMutatingWebhookInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.942441Z"}
kubevirt/virt-controller-749d8d99d4-wz5m4[virt-controller]: {"component":"virt-controller","level":"info","msg":"set log verbosity to 2","pos":"application.go:405","timestamp":"2022-06-20T00:14:37.906486Z"}
kubevirt/virt-controller-749d8d99d4-wz5m4[virt-controller]: {"component":"virt-controller","level":"info","msg":"setting rate limiter to 20 QPS and 30 Burst","pos":"application.go:398","timestamp":"2022-06-20T00:14:37.906499Z"}
kubevirt/virt-controller-749d8d99d4-wz5m4[virt-controller]: {"component":"virt-controller","level":"info","msg":"set log verbosity to 2","pos":"application.go:405","timestamp":"2022-06-20T00:14:37.906531Z"}
kubevirt/virt-controller-749d8d99d4-wz5m4[virt-controller]: {"component":"virt-controller","level":"info","msg":"setting rate limiter to 20 QPS and 30 Burst","pos":"application.go:398","timestamp":"2022-06-20T00:14:37.906543Z"}
kubevirt/virt-controller-749d8d99d4-wz5m4[virt-controller]: {"component":"virt-controller","level":"info","msg":"set log verbosity to 2","pos":"application.go:405","timestamp":"2022-06-20T00:14:37.906630Z"}
kubevirt/virt-controller-749d8d99d4-wz5m4[virt-controller]: {"component":"virt-controller","level":"info","msg":"set log verbosity to 2","pos":"application.go:405","timestamp":"2022-06-20T00:14:37.906643Z"}
kubevirt/virt-controller-749d8d99d4-wz5m4[virt-controller]: {"component":"virt-controller","level":"info","msg":"setting rate limiter to 20 QPS and 30 Burst","pos":"application.go:398","timestamp":"2022-06-20T00:14:37.906658Z"}
kubevirt/virt-controller-749d8d99d4-wz5m4[virt-controller]: W0620 00:14:37.908207 1 shared_informer.go:504] resyncPeriod 5m0s is smaller than resyncCheckPeriod 19h15m21.324440719s and the informer has already started. Changing it to 19h15m21.324440719s
kubevirt/virt-controller-749d8d99d4-wz5m4[virt-controller]: W0620 00:14:37.908481 1 client_config.go:617] Neither --kubeconfig nor --master was specified. Using the inClusterConfig. This might not work.
kubevirt/virt-controller-749d8d99d4-wz5m4[virt-controller]: {"component":"virt-controller","level":"info","msg":"certificate with common name 'virt-controller.kubevirt.pod.cluster.local' retrieved.","pos":"cert-manager.go:198","timestamp":"2022-06-20T00:14:37.908823Z"}
kubevirt/virt-controller-749d8d99d4-wz5m4[virt-controller]: I0620 00:14:37.909196 1 leaderelection.go:248] attempting to acquire leader lease kubevirt/virt-controller...
kubevirt/virt-controller-749d8d99d4-wz5m4[virt-controller]: {"action":"listening","component":"virt-controller","interface":"0.0.0.0","level":"info","port":8443,"pos":"application.go:417","service":"http","timestamp":"2022-06-20T00:14:37.909389Z"}
kubevirt/virt-controller-749d8d99d4-wz5m4[virt-controller]: {"component":"virt-controller","level":"info","msg":"Updating cluster config from KubeVirt to resource version '1805'","pos":"configuration.go:320","timestamp":"2022-06-20T00:15:20.410498Z"}
kubevirt/virt-controller-749d8d99d4-wz5m4[virt-controller]: {"component":"virt-controller","level":"info","msg":"setting rate limiter to 20 QPS and 30 Burst","pos":"application.go:398","timestamp":"2022-06-20T00:15:20.412673Z"}
kubevirt/virt-controller-749d8d99d4-wz5m4[virt-controller]: {"component":"virt-controller","level":"info","msg":"set log verbosity to 2","pos":"application.go:405","timestamp":"2022-06-20T00:15:20.413832Z"}
kubevirt/virt-controller-749d8d99d4-wz5m4[virt-controller]: {"component":"virt-controller","level":"info","msg":"Updating cluster config from KubeVirt to resource version '1823'","pos":"configuration.go:320","timestamp":"2022-06-20T00:15:25.443059Z"}
kubevirt/virt-controller-749d8d99d4-wz5m4[virt-controller]: {"component":"virt-controller","level":"info","msg":"setting rate limiter to 20 QPS and 30 Burst","pos":"application.go:398","timestamp":"2022-06-20T00:15:25.443123Z"}
kubevirt/virt-controller-749d8d99d4-wz5m4[virt-controller]: {"component":"virt-controller","level":"info","msg":"set log verbosity to 2","pos":"application.go:405","timestamp":"2022-06-20T00:15:25.443160Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: W0620 00:14:25.367142 1 client_config.go:617] Neither --kubeconfig nor --master was specified. Using the inClusterConfig. This might not work.
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"STARTING informer CRDInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:14:25.368489Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"STARTING informer kubeVirtInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:14:25.371876Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"CDI detected, DataVolume integration enabled","pos":"application.go:347","timestamp":"2022-06-20T00:14:25.573598Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: W0620 00:14:25.573672 1 client_config.go:617] Neither --kubeconfig nor --master was specified. Using the inClusterConfig. This might not work.
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"Updating cluster config from KubeVirt to resource version '1485'","pos":"configuration.go:320","timestamp":"2022-06-20T00:14:25.573762Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"setting rate limiter to 20 QPS and 30 Burst","pos":"application.go:398","timestamp":"2022-06-20T00:14:25.573812Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"set log verbosity to 2","pos":"application.go:405","timestamp":"2022-06-20T00:14:25.573848Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"setting rate limiter to 20 QPS and 30 Burst","pos":"application.go:398","timestamp":"2022-06-20T00:14:25.573862Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"set log verbosity to 2","pos":"application.go:405","timestamp":"2022-06-20T00:14:25.573901Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"set log verbosity to 2","pos":"application.go:405","timestamp":"2022-06-20T00:14:25.573964Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"setting rate limiter to 20 QPS and 30 Burst","pos":"application.go:398","timestamp":"2022-06-20T00:14:25.573976Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"set log verbosity to 2","pos":"application.go:405","timestamp":"2022-06-20T00:14:25.574043Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"set log verbosity to 2","pos":"application.go:405","timestamp":"2022-06-20T00:14:25.574060Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"setting rate limiter to 20 QPS and 30 Burst","pos":"application.go:398","timestamp":"2022-06-20T00:14:25.574072Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: W0620 00:14:25.575879 1 shared_informer.go:504] resyncPeriod 5m0s is smaller than resyncCheckPeriod 19h15m21.324440719s and the informer has already started. Changing it to 19h15m21.324440719s
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: W0620 00:14:25.576108 1 client_config.go:617] Neither --kubeconfig nor --master was specified. Using the inClusterConfig. This might not work.
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: I0620 00:14:25.576754 1 leaderelection.go:248] attempting to acquire leader lease kubevirt/virt-controller...
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"certificate with common name 'virt-controller.kubevirt.pod.cluster.local' retrieved.","pos":"cert-manager.go:198","timestamp":"2022-06-20T00:14:25.577378Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"action":"listening","component":"virt-controller","interface":"0.0.0.0","level":"info","port":8443,"pos":"application.go:417","service":"http","timestamp":"2022-06-20T00:14:25.577426Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: I0620 00:14:25.582782 1 leaderelection.go:258] successfully acquired lease kubevirt/virt-controller
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"STARTING informer vmInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:14:25.584687Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"STARTING informer vmimInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:14:25.584724Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"STARTING informer vmFlavorInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:14:25.584734Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"STARTING informer kubeVirtNodeInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:14:25.584745Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"STARTING informer vmpool","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:14:25.584754Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"STARTING informer persistentVolumeClaimInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:14:25.584763Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"STARTING informer controllerRevisionInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:14:25.584771Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"STARTING informer storageClassInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:14:25.584781Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"STARTING informer podInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:14:25.584789Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"STARTING informer dataVolumeInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:14:25.584798Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"STARTING informer migrationPolicyInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:14:25.584806Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"STARTING informer kubeVirtPodInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:14:25.584817Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"SKIPPING informer kubeVirtInformer","pos":"virtinformers.go:302","timestamp":"2022-06-20T00:14:25.584827Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"STARTING informer vmirsInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:14:25.584837Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"STARTING informer vmSnapshotInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:14:25.584846Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"STARTING informer vmSnapshotContentInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:14:25.584854Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"STARTING informer vmRestoreInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:14:25.584862Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"STARTING informer cdiInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:14:25.584872Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"STARTING informer vmClusterFlavorInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:14:25.584880Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"SKIPPING informer CRDInformer","pos":"virtinformers.go:302","timestamp":"2022-06-20T00:14:25.584888Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"STARTING informer cdiConfigInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:14:25.584897Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"STARTING informer vmiInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:14:25.584905Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"STARTING controllers with following threads : node 3, vmi 10, replicaset 3, vm 3, migration 3, evacuation 3, disruptionBudget 3\n","pos":"application.go:444","timestamp":"2022-06-20T00:14:25.584937Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"Starting vmi collector","pos":"collector.go:83","timestamp":"2022-06-20T00:14:25.584948Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"Starting performance and scale metrics","pos":"register.go:30","timestamp":"2022-06-20T00:14:25.584967Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"Starting evacuation controller.","pos":"evacuation.go:270","timestamp":"2022-06-20T00:14:25.612347Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"Starting disruption budget controller.","pos":"disruptionbudget.go:314","timestamp":"2022-06-20T00:14:25.612402Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"Starting node controller.","pos":"node.go:110","timestamp":"2022-06-20T00:14:25.612432Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"Starting vmi controller.","pos":"vmi.go:238","timestamp":"2022-06-20T00:14:25.612458Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"Starting VirtualMachineInstanceReplicaSet controller.","pos":"replicaset.go:112","timestamp":"2022-06-20T00:14:25.612488Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"Starting pool controller.","pos":"pool.go:424","timestamp":"2022-06-20T00:14:25.612513Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"Starting VirtualMachine controller.","pos":"vm.go:163","timestamp":"2022-06-20T00:14:25.612533Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"Starting migration controller.","pos":"migration.go:170","timestamp":"2022-06-20T00:14:25.618818Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"Starting snapshot controller.","pos":"snapshot_base.go:196","timestamp":"2022-06-20T00:14:25.618863Z"}
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer operatorAPIServiceInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.942449Z"}
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer operatorPodDisruptionBudgetInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.942457Z"}
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer secretsInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.942466Z"}
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer FakeOperatorSCC","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.942476Z"}
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer FakeOperatorPrometheusRuleInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.942488Z"}
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer CRDInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.942497Z"}
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer OperatorRoleInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.942506Z"}
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer OperatorDaemonSetInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.942514Z"}
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer installStrategyJobsInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.942523Z"}
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer operatorPodsInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.942532Z"}
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer OperatorConfigMapInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.942541Z"}
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: {"component":"virt-operator","level":"info","msg":"STARTING informer FakeOperatorServiceMonitor","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:12:56.942549Z"}
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: {"component":"virt-operator","level":"info","msg":"Attempting to acquire leader status","pos":"application.go:406","timestamp":"2022-06-20T00:12:57.042893Z"}
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: I0620 00:12:57.043000 1 leaderelection.go:248] attempting to acquire leader lease kubevirt/virt-operator...
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"Starting restore controller.","pos":"restore_base.go:93","timestamp":"2022-06-20T00:14:25.618878Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"Starting workload update controller.","pos":"workload-updater.go:220","timestamp":"2022-06-20T00:14:25.618896Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"error","msg":"Skipping TSC frequency updates on all nodes","pos":"nodetopologyupdater.go:54","reason":"failed to calculate lowest TSC frequency for nodes: no schedulable node exposes a tsc-frequency","timestamp":"2022-06-20T00:14:25.623859Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"TSC Freqency node update status: 0 updated, 0 skipped, 0 errors","pos":"nodetopologyupdater.go:47","timestamp":"2022-06-20T00:14:25.638731Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"error","msg":"Skipping TSC frequency updates on all nodes","pos":"nodetopologyupdater.go:54","reason":"failed to calculate lowest TSC frequency for nodes: no schedulable node exposes a tsc-frequency","timestamp":"2022-06-20T00:15:06.334928Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"TSC Freqency node update status: 0 updated, 0 skipped, 0 errors","pos":"nodetopologyupdater.go:47","timestamp":"2022-06-20T00:15:06.335028Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"Updating cluster config from KubeVirt to resource version '1805'","pos":"configuration.go:320","timestamp":"2022-06-20T00:15:20.410547Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"setting rate limiter to 20 QPS and 30 Burst","pos":"application.go:398","timestamp":"2022-06-20T00:15:20.411705Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"set log verbosity to 2","pos":"application.go:405","timestamp":"2022-06-20T00:15:20.413459Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"Updating cluster config from KubeVirt to resource version '1823'","pos":"configuration.go:320","timestamp":"2022-06-20T00:15:25.448786Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"setting rate limiter to 20 QPS and 30 Burst","pos":"application.go:398","timestamp":"2022-06-20T00:15:25.453274Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"set log verbosity to 2","pos":"application.go:405","timestamp":"2022-06-20T00:15:25.453412Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"error","msg":"Skipping TSC frequency updates on all nodes","pos":"nodetopologyupdater.go:54","reason":"failed to calculate lowest TSC frequency for nodes: no schedulable node exposes a tsc-frequency","timestamp":"2022-06-20T00:16:03.427824Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"TSC Freqency node update status: 0 updated, 0 skipped, 0 errors","pos":"nodetopologyupdater.go:47","timestamp":"2022-06-20T00:16:03.427967Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"error","msg":"Skipping TSC frequency updates on all nodes","pos":"nodetopologyupdater.go:54","reason":"failed to calculate lowest TSC frequency for nodes: no schedulable node exposes a tsc-frequency","timestamp":"2022-06-20T00:16:40.865996Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"TSC Freqency node update status: 0 updated, 0 skipped, 0 errors","pos":"nodetopologyupdater.go:47","timestamp":"2022-06-20T00:16:40.866069Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"error","msg":"Skipping TSC frequency updates on all nodes","pos":"nodetopologyupdater.go:54","reason":"failed to calculate lowest TSC frequency for nodes: no schedulable node exposes a tsc-frequency","timestamp":"2022-06-20T00:17:42.019126Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"TSC Freqency node update status: 0 updated, 0 skipped, 0 errors","pos":"nodetopologyupdater.go:47","timestamp":"2022-06-20T00:17:42.019196Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","kind":"","level":"error","msg":"Updating api version annotations failed","name":"instance","namespace":"default","pos":"vm.go:233","reason":"Operation cannot be fulfilled on virtualmachines.kubevirt.io \"instance\": the object has been modified; please apply your changes to the latest version and try again","timestamp":"2022-06-20T00:18:31.711701Z","uid":"de0082f5-1e03-49bb-abdc-b46ecf5af7d1"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"re-enqueuing VirtualMachine default/instance","pos":"vm.go:199","reason":"Operation cannot be fulfilled on virtualmachines.kubevirt.io \"instance\": the object has been modified; please apply your changes to the latest version and try again","timestamp":"2022-06-20T00:18:31.711766Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","kind":"","level":"info","msg":"Starting VM due to runStrategy: Always","name":"instance","namespace":"default","pos":"vm.go:633","timestamp":"2022-06-20T00:18:31.725304Z","uid":"de0082f5-1e03-49bb-abdc-b46ecf5af7d1"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","kind":"","level":"info","msg":"Started VM by creating the new virtual machine instance instance","name":"instance","namespace":"default","pos":"vm.go:850","timestamp":"2022-06-20T00:18:31.759650Z","uid":"de0082f5-1e03-49bb-abdc-b46ecf5af7d1"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","kind":"","level":"info","msg":"arguments for container-disk \"volumeboot\": --copy-path /var/run/kubevirt-ephemeral-disks/container-disk-data/ea7e77f5-c65e-4a05-8537-bb5d108bb5b0/disk_0","name":"instance","namespace":"default","pos":"container-disk.go:310","timestamp":"2022-06-20T00:18:31.799623Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: E0620 00:18:31.828564 1 util.go:130] Operation cannot be fulfilled on virtualmachines.kubevirt.io "instance": the object has been modified; please apply your changes to the latest version and try again
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"error","msg":"Skipping TSC frequency updates on all nodes","pos":"nodetopologyupdater.go:54","reason":"failed to calculate lowest TSC frequency for nodes: no schedulable node exposes a tsc-frequency","timestamp":"2022-06-20T00:18:37.101611Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"TSC Freqency node update status: 0 updated, 0 skipped, 0 errors","pos":"nodetopologyupdater.go:47","timestamp":"2022-06-20T00:18:37.101695Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"error","msg":"Skipping TSC frequency updates on all nodes","pos":"nodetopologyupdater.go:54","reason":"failed to calculate lowest TSC frequency for nodes: no schedulable node exposes a tsc-frequency","timestamp":"2022-06-20T00:19:08.121293Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"TSC Freqency node update status: 0 updated, 0 skipped, 0 errors","pos":"nodetopologyupdater.go:47","timestamp":"2022-06-20T00:19:08.121621Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"error","msg":"Skipping TSC frequency updates on all nodes","pos":"nodetopologyupdater.go:54","reason":"failed to calculate lowest TSC frequency for nodes: no schedulable node exposes a tsc-frequency","timestamp":"2022-06-20T00:19:43.824493Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"TSC Freqency node update status: 0 updated, 0 skipped, 0 errors","pos":"nodetopologyupdater.go:47","timestamp":"2022-06-20T00:19:43.825128Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"error","msg":"Skipping TSC frequency updates on all nodes","pos":"nodetopologyupdater.go:54","reason":"failed to calculate lowest TSC frequency for nodes: no schedulable node exposes a tsc-frequency","timestamp":"2022-06-20T00:20:35.688166Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"TSC Freqency node update status: 0 updated, 0 skipped, 0 errors","pos":"nodetopologyupdater.go:47","timestamp":"2022-06-20T00:20:35.690942Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"error","msg":"Skipping TSC frequency updates on all nodes","pos":"nodetopologyupdater.go:54","reason":"failed to calculate lowest TSC frequency for nodes: no schedulable node exposes a tsc-frequency","timestamp":"2022-06-20T00:21:11.932132Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"TSC Freqency node update status: 0 updated, 0 skipped, 0 errors","pos":"nodetopologyupdater.go:47","timestamp":"2022-06-20T00:21:11.934976Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"error","msg":"Skipping TSC frequency updates on all nodes","pos":"nodetopologyupdater.go:54","reason":"failed to calculate lowest TSC frequency for nodes: no schedulable node exposes a tsc-frequency","timestamp":"2022-06-20T00:21:51.064673Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"TSC Freqency node update status: 0 updated, 0 skipped, 0 errors","pos":"nodetopologyupdater.go:47","timestamp":"2022-06-20T00:21:51.065242Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"error","msg":"Skipping TSC frequency updates on all nodes","pos":"nodetopologyupdater.go:54","reason":"failed to calculate lowest TSC frequency for nodes: no schedulable node exposes a tsc-frequency","timestamp":"2022-06-20T00:22:31.762303Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"TSC Freqency node update status: 0 updated, 0 skipped, 0 errors","pos":"nodetopologyupdater.go:47","timestamp":"2022-06-20T00:22:31.763809Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"error","msg":"Skipping TSC frequency updates on all nodes","pos":"nodetopologyupdater.go:54","reason":"failed to calculate lowest TSC frequency for nodes: no schedulable node exposes a tsc-frequency","timestamp":"2022-06-20T00:23:36.934325Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"TSC Freqency node update status: 0 updated, 0 skipped, 0 errors","pos":"nodetopologyupdater.go:47","timestamp":"2022-06-20T00:23:36.937621Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"error","msg":"Skipping TSC frequency updates on all nodes","pos":"nodetopologyupdater.go:54","reason":"failed to calculate lowest TSC frequency for nodes: no schedulable node exposes a tsc-frequency","timestamp":"2022-06-20T00:24:18.156999Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"TSC Freqency node update status: 0 updated, 0 skipped, 0 errors","pos":"nodetopologyupdater.go:47","timestamp":"2022-06-20T00:24:18.161363Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: E0620 00:24:36.230249 1 util.go:130] Operation cannot be fulfilled on virtualmachines.kubevirt.io "instance-full": the object has been modified; please apply your changes to the latest version and try again
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","kind":"","level":"info","msg":"Looking for DataVolume Ref","name":"disk-dv-instance-full","namespace":"kube-public","pos":"vm.go:1428","timestamp":"2022-06-20T00:24:36.277628Z","uid":"55dce5ae-78d6-4f44-9185-15a8fa714053"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","kind":"","level":"info","msg":"DataVolume created because disk-dv-instance-full was added.","name":"disk-dv-instance-full","namespace":"kube-public","pos":"vm.go:1435","timestamp":"2022-06-20T00:24:36.287044Z","uid":"55dce5ae-78d6-4f44-9185-15a8fa714053"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","kind":"","level":"error","msg":"Updating the VirtualMachine status failed.","name":"instance-full","namespace":"kube-public","pos":"vm.go:311","reason":"Operation cannot be fulfilled on virtualmachines.kubevirt.io \"instance-full\": the object has been modified; please apply your changes to the latest version and try again","timestamp":"2022-06-20T00:24:36.286962Z","uid":"4d287919-4c7e-4c9a-9097-dbaf6ad8d5c8"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"re-enqueuing VirtualMachine kube-public/instance-full","pos":"vm.go:199","reason":"Operation cannot be fulfilled on virtualmachines.kubevirt.io \"instance-full\": the object has been modified; please apply your changes to the latest version and try again","timestamp":"2022-06-20T00:24:36.287380Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","kind":"","level":"error","msg":"Updating the VirtualMachine status failed.","name":"instance-full","namespace":"kube-public","pos":"vm.go:311","reason":"Operation cannot be fulfilled on virtualmachines.kubevirt.io \"instance-full\": the object has been modified; please apply your changes to the latest version and try again","timestamp":"2022-06-20T00:24:36.308454Z","uid":"4d287919-4c7e-4c9a-9097-dbaf6ad8d5c8"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"re-enqueuing VirtualMachine kube-public/instance-full","pos":"vm.go:199","reason":"Operation cannot be fulfilled on virtualmachines.kubevirt.io \"instance-full\": the object has been modified; please apply your changes to the latest version and try again","timestamp":"2022-06-20T00:24:36.308599Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"warning","msg":"No VolumeSnapshotClass for standard","pos":"snapshot.go:560","timestamp":"2022-06-20T00:24:36.364047Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"warning","msg":"No VolumeSnapshotClass for standard","pos":"snapshot.go:560","timestamp":"2022-06-20T00:24:36.399190Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","kind":"","level":"error","msg":"Updating api version annotations failed","name":"instance-almost-default","namespace":"default","pos":"vm.go:233","reason":"Operation cannot be fulfilled on virtualmachines.kubevirt.io \"instance-almost-default\": the object has been modified; please apply your changes to the latest version and try again","timestamp":"2022-06-20T00:24:37.702026Z","uid":"fa282040-dc37-4fc8-8c79-3950cf6f8d99"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"re-enqueuing VirtualMachine default/instance-almost-default","pos":"vm.go:199","reason":"Operation cannot be fulfilled on virtualmachines.kubevirt.io \"instance-almost-default\": the object has been modified; please apply your changes to the latest version and try again","timestamp":"2022-06-20T00:24:37.702093Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: E0620 00:24:37.737463 1 util.go:130] Operation cannot be fulfilled on virtualmachines.kubevirt.io "instance-almost-default": the object has been modified; please apply your changes to the latest version and try again
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: E0620 00:24:39.061859 1 util.go:130] Operation cannot be fulfilled on virtualmachines.kubevirt.io "instance-running-false": the object has been modified; please apply your changes to the latest version and try again
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","kind":"","level":"error","msg":"Updating the VirtualMachine status failed.","name":"instance-running-false","namespace":"default","pos":"vm.go:311","reason":"Operation cannot be fulfilled on virtualmachines.kubevirt.io \"instance-running-false\": the object has been modified; please apply your changes to the latest version and try again","timestamp":"2022-06-20T00:24:39.085289Z","uid":"2bc62361-abb3-4b92-83ae-57bda699541c"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"info","msg":"re-enqueuing VirtualMachine default/instance-running-false","pos":"vm.go:199","reason":"Operation cannot be fulfilled on virtualmachines.kubevirt.io \"instance-running-false\": the object has been modified; please apply your changes to the latest version and try again","timestamp":"2022-06-20T00:24:39.085346Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"warning","msg":"No VolumeSnapshotClass for standard","pos":"snapshot.go:560","timestamp":"2022-06-20T00:24:43.493470Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: E0620 00:24:43.519475 1 util.go:130] Operation cannot be fulfilled on virtualmachines.kubevirt.io "instance-full": the object has been modified; please apply your changes to the latest version and try again
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"warning","msg":"No VolumeSnapshotClass for standard","pos":"snapshot.go:560","timestamp":"2022-06-20T00:24:43.519663Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"warning","msg":"No VolumeSnapshotClass for standard","pos":"snapshot.go:560","timestamp":"2022-06-20T00:24:43.533560Z"}
kubevirt/virt-controller-749d8d99d4-csgtw[virt-controller]: {"component":"virt-controller","level":"warning","msg":"No VolumeSnapshotClass for standard","pos":"snapshot.go:560","timestamp":"2022-06-20T00:24:47.891113Z"}
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: {"component":"virt-operator","level":"info","msg":"Updating cluster config from KubeVirt to resource version '807'","pos":"configuration.go:320","timestamp":"2022-06-20T00:13:06.812341Z"}
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: {"component":"virt-operator","level":"info","msg":"Updating cluster config from KubeVirt to resource version '1233'","pos":"configuration.go:320","timestamp":"2022-06-20T00:13:30.071486Z"}
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: {"component":"virt-operator","level":"info","msg":"CA update in configmap kube-system/extension-apiserver-authentication detected. Updating from resource version -1 to 40","pos":"ca-manager.go:96","timestamp":"2022-06-20T00:13:40.071759Z"}
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: {"component":"virt-operator","level":"info","msg":"Updating cluster config from KubeVirt to resource version '1293'","pos":"configuration.go:320","timestamp":"2022-06-20T00:13:40.084064Z"}
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: {"component":"virt-operator","level":"info","msg":"Updating cluster config from KubeVirt to resource version '1485'","pos":"configuration.go:320","timestamp":"2022-06-20T00:14:15.287303Z"}
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: {"component":"virt-operator","level":"info","msg":"Updating cluster config from KubeVirt to resource version '1805'","pos":"configuration.go:320","timestamp":"2022-06-20T00:15:20.399842Z"}
kubevirt/virt-operator-564f568975-g2bh6[virt-operator]: {"component":"virt-operator","level":"info","msg":"Updating cluster config from KubeVirt to resource version '1823'","pos":"configuration.go:320","timestamp":"2022-06-20T00:15:25.450806Z"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:25:10.075126Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: W0620 00:13:55.593519 1 client_config.go:617] Neither --kubeconfig nor --master was specified. Using the inClusterConfig. This might not work.
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: W0620 00:13:55.594502 1 client_config.go:617] Neither --kubeconfig nor --master was specified. Using the inClusterConfig. This might not work.
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"STARTING informer extensionsKubeVirtCAConfigMapInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:13:55.620892Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"STARTING informer CRDInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:13:55.620959Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"STARTING informer vmiPresetInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:13:55.620973Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"STARTING informer limitrangeInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:13:55.620986Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"STARTING informer vmRestoreInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:13:55.620995Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"STARTING informer kubeVirtInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:13:55.621004Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"STARTING informer extensionsConfigMapInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:13:55.621013Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"Waiting for cache sync of informer CRDInformer","pos":"virtinformers.go:317","timestamp":"2022-06-20T00:13:55.621023Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"Waiting for cache sync of informer vmiPresetInformer","pos":"virtinformers.go:317","timestamp":"2022-06-20T00:13:55.621033Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"Waiting for cache sync of informer limitrangeInformer","pos":"virtinformers.go:317","timestamp":"2022-06-20T00:13:55.621042Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"Waiting for cache sync of informer vmRestoreInformer","pos":"virtinformers.go:317","timestamp":"2022-06-20T00:13:55.621051Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"Waiting for cache sync of informer kubeVirtInformer","pos":"virtinformers.go:317","timestamp":"2022-06-20T00:13:55.621059Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"Waiting for cache sync of informer extensionsConfigMapInformer","pos":"virtinformers.go:317","timestamp":"2022-06-20T00:13:55.621090Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"Waiting for cache sync of informer extensionsKubeVirtCAConfigMapInformer","pos":"virtinformers.go:317","timestamp":"2022-06-20T00:13:55.621101Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"CDI detected, DataSource integration enabled","pos":"api.go:932","timestamp":"2022-06-20T00:13:55.921928Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"SKIPPING informer vmRestoreInformer","pos":"virtinformers.go:302","timestamp":"2022-06-20T00:13:55.921993Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"STARTING informer dataSourceInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:13:55.922011Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"STARTING informer vmFlavorInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:13:55.922025Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"SKIPPING informer CRDInformer","pos":"virtinformers.go:302","timestamp":"2022-06-20T00:13:55.922040Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"SKIPPING informer vmiPresetInformer","pos":"virtinformers.go:302","timestamp":"2022-06-20T00:13:55.922063Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"SKIPPING informer extensionsKubeVirtCAConfigMapInformer","pos":"virtinformers.go:302","timestamp":"2022-06-20T00:13:55.922076Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"SKIPPING informer limitrangeInformer","pos":"virtinformers.go:302","timestamp":"2022-06-20T00:13:55.922090Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"STARTING informer vmClusterFlavorInformer","pos":"virtinformers.go:305","timestamp":"2022-06-20T00:13:55.922099Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"SKIPPING informer kubeVirtInformer","pos":"virtinformers.go:302","timestamp":"2022-06-20T00:13:55.922117Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"SKIPPING informer extensionsConfigMapInformer","pos":"virtinformers.go:302","timestamp":"2022-06-20T00:13:55.922131Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"Waiting for cache sync of informer vmiPresetInformer","pos":"virtinformers.go:317","timestamp":"2022-06-20T00:13:55.922145Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"Waiting for cache sync of informer vmRestoreInformer","pos":"virtinformers.go:317","timestamp":"2022-06-20T00:13:55.922155Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"Waiting for cache sync of informer dataSourceInformer","pos":"virtinformers.go:317","timestamp":"2022-06-20T00:13:55.922172Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"Waiting for cache sync of informer vmFlavorInformer","pos":"virtinformers.go:317","timestamp":"2022-06-20T00:13:55.922182Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"Waiting for cache sync of informer CRDInformer","pos":"virtinformers.go:317","timestamp":"2022-06-20T00:13:55.922195Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"Waiting for cache sync of informer extensionsConfigMapInformer","pos":"virtinformers.go:317","timestamp":"2022-06-20T00:13:55.922204Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"Waiting for cache sync of informer extensionsKubeVirtCAConfigMapInformer","pos":"virtinformers.go:317","timestamp":"2022-06-20T00:13:55.922219Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"Waiting for cache sync of informer limitrangeInformer","pos":"virtinformers.go:317","timestamp":"2022-06-20T00:13:55.922228Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"Waiting for cache sync of informer vmClusterFlavorInformer","pos":"virtinformers.go:317","timestamp":"2022-06-20T00:13:55.922236Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"Waiting for cache sync of informer kubeVirtInformer","pos":"virtinformers.go:317","timestamp":"2022-06-20T00:13:55.922249Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"Updating cluster config from KubeVirt to resource version '1293'","pos":"configuration.go:320","timestamp":"2022-06-20T00:13:55.922637Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"setting rate limiter for the API to 5 QPS and 10 Burst","pos":"api.go:1002","timestamp":"2022-06-20T00:13:55.922679Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"setting rate limiter for webhooks to 200 QPS and 400 Burst","pos":"api.go:1006","timestamp":"2022-06-20T00:13:55.922699Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"set log verbosity to 2","pos":"api.go:993","timestamp":"2022-06-20T00:13:55.922737Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"set log verbosity to 2","pos":"api.go:993","timestamp":"2022-06-20T00:13:55.922760Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"setting rate limiter for the API to 5 QPS and 10 Burst","pos":"api.go:1002","timestamp":"2022-06-20T00:13:55.922773Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"setting rate limiter for webhooks to 200 QPS and 400 Burst","pos":"api.go:1006","timestamp":"2022-06-20T00:13:55.922787Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"set log verbosity to 2","pos":"api.go:993","timestamp":"2022-06-20T00:13:55.924566Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"setting rate limiter for the API to 5 QPS and 10 Burst","pos":"api.go:1002","timestamp":"2022-06-20T00:13:55.924677Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"setting rate limiter for webhooks to 200 QPS and 400 Burst","pos":"api.go:1006","timestamp":"2022-06-20T00:13:55.924698Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"setting rate limiter for the API to 5 QPS and 10 Burst","pos":"api.go:1002","timestamp":"2022-06-20T00:13:55.924723Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"setting rate limiter for webhooks to 200 QPS and 400 Burst","pos":"api.go:1006","timestamp":"2022-06-20T00:13:55.924734Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"set log verbosity to 2","pos":"api.go:993","timestamp":"2022-06-20T00:13:55.924759Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"set log verbosity to 2","pos":"api.go:993","timestamp":"2022-06-20T00:13:55.924789Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"certificate with common name 'virt-api.kubevirt.pod.cluster.local' retrieved.","pos":"cert-manager.go:198","timestamp":"2022-06-20T00:13:56.022832Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"certificate with common name 'kubevirt.io:system:client:virt-handler' retrieved.","pos":"cert-manager.go:198","timestamp":"2022-06-20T00:13:56.023052Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"CA update in configmap kube-system/extension-apiserver-authentication detected. Updating from resource version -1 to 40","pos":"ca-manager.go:96","timestamp":"2022-06-20T00:14:10.071493Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: W0620 00:14:10.131558 1 client_config.go:617] Neither --kubeconfig nor --master was specified. Using the inClusterConfig. This might not work.
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:10.146592Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:10.402904Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:10.418742Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:10.435147Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:10.457818Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:10.458833Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:10.458879Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:10.459774Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:10.697827Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":67779,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:11.230787Z","url":"/openapi/v2","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:11.566298Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:12.991873Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:13.867018Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:14.949101Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:15.004539Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:15.303822Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"Updating cluster config from KubeVirt to resource version '1485'","pos":"configuration.go:320","timestamp":"2022-06-20T00:14:15.329876Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"setting rate limiter for the API to 5 QPS and 10 Burst","pos":"api.go:1002","timestamp":"2022-06-20T00:14:15.329916Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"setting rate limiter for webhooks to 200 QPS and 400 Burst","pos":"api.go:1006","timestamp":"2022-06-20T00:14:15.329928Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"set log verbosity to 2","pos":"api.go:993","timestamp":"2022-06-20T00:14:15.329959Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:16.167431Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:17.591111Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:17.967495Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:19.890896Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:20.046017Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:20.765937Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:22.191245Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:23.077306Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:23.654562Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:24.493492Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:26.790640Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:27.594336Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:27.615801Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:27.655653Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:27.659248Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:27.665587Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:29.091141Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:30.044264Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:31.390609Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:33.690863Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:35.997922Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:38.290849Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:40.044077Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:40.590845Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:44.979088Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:45.022766Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:50.043960Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:57.506914Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:57.527343Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:57.535208Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:57.536314Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:57.536668Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:14:57.537629Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:00.043409Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:10.049432Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":67779,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:11.309408Z","url":"/openapi/v2","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:15.007782Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:15.035898Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:20.043298Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"Updating cluster config from KubeVirt to resource version '1805'","pos":"configuration.go:320","timestamp":"2022-06-20T00:15:20.402848Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"setting rate limiter for the API to 5 QPS and 10 Burst","pos":"api.go:1002","timestamp":"2022-06-20T00:15:20.405244Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"setting rate limiter for webhooks to 200 QPS and 400 Burst","pos":"api.go:1006","timestamp":"2022-06-20T00:15:20.405298Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"set log verbosity to 2","pos":"api.go:993","timestamp":"2022-06-20T00:15:20.405353Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"Updating cluster config from KubeVirt to resource version '1823'","pos":"configuration.go:320","timestamp":"2022-06-20T00:15:25.442294Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"setting rate limiter for the API to 5 QPS and 10 Burst","pos":"api.go:1002","timestamp":"2022-06-20T00:15:25.442354Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"setting rate limiter for webhooks to 200 QPS and 400 Burst","pos":"api.go:1006","timestamp":"2022-06-20T00:15:25.442369Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"set log verbosity to 2","pos":"api.go:993","timestamp":"2022-06-20T00:15:25.442419Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:27.537157Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:27.551931Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:27.561059Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:27.565821Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:27.574258Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:27.796350Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:30.043768Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:30.094387Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:32.402480Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:34.643405Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:36.996115Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:39.293348Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:40.042901Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:45.042747Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:45.059668Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:50.043158Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:57.552122Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:57.559637Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:57.560707Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:15:57.570681Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:00.043678Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:10.043239Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":67779,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:11.335269Z","url":"/openapi/v2","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:15.100046Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:15.123619Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:20.043897Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:27.508298Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:27.516806Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:27.519141Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:27.527406Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:27.531298Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:30.043912Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:40.043112Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:40.261398Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:42.611076Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:44.860975Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:45.167728Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:45.182545Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:47.163679Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:49.483744Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:50.043606Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:51.811566Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:57.677594Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:57.684150Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:57.698112Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:57.713434Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:57.713783Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:16:57.740965Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:00.043029Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:10.043619Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":67779,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:11.395362Z","url":"/openapi/v2","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:15.220049Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:15.242632Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"http: TLS handshake error from 172.17.0.1:57583: EOF\n","pos":"server.go:3160","timestamp":"2022-06-20T00:17:15.248495Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:20.064287Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:27.537289Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:27.545289Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:27.559133Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:27.565402Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:27.579658Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:27.583316Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:30.047682Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:40.043266Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:45.314637Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"http: TLS handshake error from 172.17.0.1:37270: EOF\n","pos":"server.go:3160","timestamp":"2022-06-20T00:17:45.327202Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:50.043452Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:57.504703Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:57.528178Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:57.552660Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:57.565936Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:57.570656Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:17:57.579920Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:00.042826Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:10.057997Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:15.370685Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:20.058632Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:22.205600Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:23.616452Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:24.877101Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:27.526884Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:27.538823Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:27.530020Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:27.543792Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:30.047700Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:30.338316Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:34.134740Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:36.593159Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:39.079669Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:40.042487Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:41.791315Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:44.584831Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:45.407398Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:45.436768Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"http: TLS handshake error from 172.17.0.1:38676: EOF\n","pos":"server.go:3160","timestamp":"2022-06-20T00:18:45.441915Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:47.095570Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"http: TLS handshake error from 172.17.0.1:52593: read tcp 172.17.0.10:8443-\u003e172.17.0.1:52593: read: connection reset by peer\n","pos":"server.go:3160","timestamp":"2022-06-20T00:18:48.787540Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:50.043550Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:50.658089Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:57.617496Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:57.639788Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:57.672278Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:57.683466Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:57.685915Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:18:57.694328Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:00.044936Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:05.432757Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:07.632451Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:10.035148Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:10.062892Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":67779,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:11.476485Z","url":"/openapi/v2","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:12.337028Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:14.632804Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:15.463774Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:15.494735Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:16.932157Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:20.072649Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:27.617072Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:27.620479Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:27.620901Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:27.623054Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:27.630309Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:27.630720Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:27.631099Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:27.648050Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:27.655835Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:30.043742Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:40.043693Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:45.522312Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:45.539641Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:50.044085Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:57.592610Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:57.593444Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:57.639597Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:57.653107Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:19:57.679908Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:00.046232Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:10.044347Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":67779,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:11.507227Z","url":"/openapi/v2","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:15.590057Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:15.636718Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:18.241014Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:20.046475Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:20.490616Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:22.791742Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:24.940866Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:27.394741Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:27.564766Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:27.632910Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:27.644947Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:27.645401Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:27.645800Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:29.741095Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:30.050581Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:40.064552Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:45.615702Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:45.670751Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:50.044968Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:57.551642Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:57.571653Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:57.576899Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:57.591465Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:57.599303Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:20:57.641173Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:00.046164Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:10.044423Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":67779,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:11.669139Z","url":"/openapi/v2","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:15.664885Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:15.700284Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:20.053936Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:27.503244Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:27.506311Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:27.512698Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:27.529293Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:30.047578Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:30.669957Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:32.348577Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:35.271691Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:37.320242Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:39.869918Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:40.044827Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:42.170376Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:45.699051Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:45.731027Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:50.045905Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:57.664705Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:57.794763Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:57.798904Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:58.045569Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:58.054265Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:58.106420Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:58.114167Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:58.132177Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:58.135158Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:58.138617Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:58.166683Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:58.177756Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:21:58.209181Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:00.047042Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:10.044192Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:20.069873Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:27.599602Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:27.615386Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:27.619210Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:27.631079Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:27.636054Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:30.044539Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:40.044209Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:45.870049Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","level":"info","msg":"http: TLS handshake error from 172.17.0.1:10691: EOF\n","pos":"server.go:3160","timestamp":"2022-06-20T00:22:45.893237Z"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:50.044558Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:57.514374Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:57.524855Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:57.528204Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:57.530700Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:57.542737Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:22:57.596736Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:00.043987Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:10.043506Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:15.950920Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:20.075088Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:27.522136Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:27.525371Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:27.548085Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:27.619702Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:27.643991Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:27.648073Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:27.658991Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:27.815956Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:30.047658Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:40.045877Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:50.043618Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:57.629836Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:57.697066Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:57.700279Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:57.718852Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:23:57.747770Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:00.049140Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:10.042810Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:20.066361Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:27.567129Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:27.569507Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:27.573423Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:27.578284Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:27.578959Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:30.046796Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:40.044553Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:50.043173Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:57.520462Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:57.526206Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:24:57.619643Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:25:00.042965Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-x77j5[virt-api]: {"component":"virt-api","contentLength":370,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:25:10.111809Z","url":"/apis/subresources.kubevirt.io/v1/healthz","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:25:10.513451Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:25:10.550516Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":67779,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:25:11.634850Z","url":"/openapi/v2","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":67779,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:25:11.787220Z","url":"/openapi/v2","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:25:12.794439Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:25:12.843802Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:25:15.093883Z","url":"/apis/subresources.kubevirt.io/v1?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:25:15.144515Z","url":"/apis/subresources.kubevirt.io/v1alpha3?timeout=32s","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:25:16.125313Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:25:16.125531Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:25:16.141252Z","url":"/apis/subresources.kubevirt.io/v1alpha3","username":"-"}
kubevirt/virt-api-77df5c4f87-fj857[virt-api]: {"component":"virt-api","contentLength":2374,"level":"info","method":"GET","pos":"filter.go:46","proto":"HTTP/1.1","remoteAddress":"172.17.0.1","statusCode":200,"timestamp":"2022-06-20T00:25:16.141274Z","url":"/apis/subresources.kubevirt.io/v1","username":"-"}
default/molecule-knmpm[molecule]: py38-ansible_4 create: /opt/molecule_kubevirt/.tox/py38-ansible_4
default/molecule-knmpm[molecule]: py38-ansible_4 installdeps: ansible>=4.0,<5.0, selinux
default/molecule-knmpm[molecule]: py38-ansible_4 develop-inst: /opt/molecule_kubevirt
default/molecule-knmpm[molecule]: py38-ansible_4 installed: ansi2html==1.7.0,ansible==4.10.0,ansible-compat==2.1.0,ansible-core==2.11.12,arrow==1.2.2,attrs==21.4.0,binaryornot==0.4.4,cachetools==5.2.0,Cerberus==1.3.2,certifi==2022.6.15,cffi==1.15.0,cfgv==3.3.1,chardet==4.0.0,charset-normalizer==2.0.12,click==8.1.3,click-help-colors==0.9.1,commonmark==0.9.1,cookiecutter==2.1.1,coverage==6.4.1,cryptography==37.0.2,distlib==0.3.4,distro==1.7.0,enrich==1.2.7,execnet==1.9.0,filelock==3.7.1,google-auth==2.8.0,identify==2.5.1,idna==3.3,importlib-resources==5.8.0,iniconfig==1.1.1,Jinja2==3.1.2,jinja2-time==0.2.0,jsonschema==4.6.0,kubernetes==11.0.0,MarkupSafe==2.1.1,molecule==4.0.0,-e git+https://github.com/jseguillon/molecule-kubevirt@08ae5a4f4bb563582d5b6f95586acacc7e7dc7ee#egg=molecule_kubevirt,more-itertools==8.13.0,nodeenv==1.6.0,oauthlib==3.2.0,openshift==0.11.2,packaging==21.3,pexpect==4.8.0,platformdirs==2.5.2,pluggy==1.0.0,pre-commit==2.19.0,ptyprocess==0.7.0,py==1.11.0,pyasn1==0.4.8,pyasn1-modules==0.2.8,pycparser==2.21,Pygments==2.12.0,pyparsing==3.0.9,pyrsistent==0.18.1,pytest==7.1.2,pytest-cov==3.0.0,pytest-forked==1.4.0,pytest-helpers-namespace==2021.12.29,pytest-html==3.1.1,pytest-metadata==2.0.1,pytest-mock==3.7.0,pytest-plus==0.2,pytest-testinfra==6.8.0,pytest-xdist==2.5.0,python-dateutil==2.8.2,python-slugify==6.1.2,python-string-utils==1.0.0,PyYAML==6.0,requests==2.28.0,requests-oauthlib==1.3.1,resolvelib==0.5.4,rich==12.4.4,rsa==4.8,ruamel.yaml==0.17.21,ruamel.yaml.clib==0.2.6,selinux==0.2.1,six==1.16.0,subprocess-tee==0.3.5,text-unidecode==1.3,toml==0.10.2,tomli==2.0.1,typing_extensions==4.2.0,urllib3==1.26.9,virtualenv==20.14.1,websocket-client==1.3.2,zipp==3.8.0
default/molecule-knmpm[molecule]: py38-ansible_4 run-test-pre: PYTHONHASHSEED='3743274778'
default/molecule-knmpm[molecule]: py38-ansible_4 run-test: commands[0] | pip check
default/molecule-knmpm[molecule]: No broken requirements found.
default/molecule-knmpm[molecule]: py38-ansible_4 run-test: commands[1] | python -m pytest -p no:cov --collect-only
default/molecule-knmpm[molecule]: ============================= test session starts ==============================
default/molecule-knmpm[molecule]: platform linux -- Python 3.8.13, pytest-7.1.2, pluggy-1.0.0
default/molecule-knmpm[molecule]: cachedir: .tox/py38-ansible_4/.pytest_cache
default/molecule-knmpm[molecule]: rootdir: /opt/molecule_kubevirt, configfile: setup.cfg
default/molecule-knmpm[molecule]: plugins: helpers-namespace-2021.12.29, plus-0.2, xdist-2.5.0, mock-3.7.0, forked-1.4.0, metadata-2.0.1, html-3.1.1, testinfra-6.8.0
default/molecule-knmpm[molecule]: [DEPRECATION WARNING]: ANSIBLE_CALLABLE_WHITELIST option, normalizing names to
default/molecule-knmpm[molecule]: new standard, use ANSIBLE_CALLABLE_ENABLED instead. This feature will be
default/molecule-knmpm[molecule]: removed from ansible-core in version 2.15. Deprecation warnings can be disabled
default/molecule-knmpm[molecule]: by setting deprecation_warnings=False in ansible.cfg.
default/molecule-knmpm[molecule]: collected 5 items
default/molecule-knmpm[molecule]:
default/molecule-knmpm[molecule]: <Package test>
default/molecule-knmpm[molecule]: <Module test_driver.py>
default/molecule-knmpm[molecule]: <Function test_driver_is_detected>
default/molecule-knmpm[molecule]: <Module test_init.py>
default/molecule-knmpm[molecule]: <Function test_command_init_and_test_scenario>
default/molecule-knmpm[molecule]: <Module test_scenario_tests.py>
default/molecule-knmpm[molecule]: <Class TestClass>
default/molecule-knmpm[molecule]: <Function test_instance_spec[kube-public-instance-full-instance-full-notmolecule]>
default/molecule-knmpm[molecule]: <Function test_instance_spec[default-instance-almost-default-instance-almost-default-molecule]>
default/molecule-knmpm[molecule]: <Function test_instance_spec[default-instance-running-false--molecule]>
default/molecule-knmpm[molecule]:
default/molecule-knmpm[molecule]: ========================== 5 tests collected in 1.42s ==========================
default/molecule-knmpm[molecule]: py38-ansible_4 run-test: commands[2] | python -m pytest -l
default/molecule-knmpm[molecule]: ============================= test session starts ==============================
default/molecule-knmpm[molecule]: platform linux -- Python 3.8.13, pytest-7.1.2, pluggy-1.0.0
default/molecule-knmpm[molecule]: cachedir: .tox/py38-ansible_4/.pytest_cache
default/molecule-knmpm[molecule]: rootdir: /opt/molecule_kubevirt, configfile: setup.cfg
default/molecule-knmpm[molecule]: plugins: helpers-namespace-2021.12.29, plus-0.2, xdist-2.5.0, mock-3.7.0, forked-1.4.0, metadata-2.0.1, html-3.1.1, testinfra-6.8.0, cov-3.0.0
default/molecule-knmpm[molecule]: [DEPRECATION WARNING]: ANSIBLE_CALLABLE_WHITELIST option, normalizing names to
default/molecule-knmpm[molecule]: new standard, use ANSIBLE_CALLABLE_ENABLED instead. This feature will be
default/molecule-knmpm[molecule]: removed from ansible-core in version 2.15. Deprecation warnings can be disabled
default/molecule-knmpm[molecule]: by setting deprecation_warnings=False in ansible.cfg.
default/molecule-knmpm[molecule]: collected 5 items
default/molecule-knmpm[molecule]:
default/molecule-knmpm[molecule]: molecule_kubevirt/test/test_driver.py . [ 20%]
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"Collected all requested hook sidecar sockets","pos":"manager.go:76","timestamp":"2022-06-20T00:18:45.911855Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"Sorted all collected sidecar sockets per hook point based on their priority and name: map[]","pos":"manager.go:79","timestamp":"2022-06-20T00:18:45.915507Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"Connecting to libvirt daemon: qemu:///system","pos":"libvirt.go:495","timestamp":"2022-06-20T00:18:45.918872Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"Connecting to libvirt daemon failed: virError(Code=38, Domain=7, Message='Failed to connect socket to '/var/run/libvirt/libvirt-sock': No such file or directory')","pos":"libvirt.go:503","timestamp":"2022-06-20T00:18:45.924230Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"libvirt version: 7.6.0, package: 6.el8s (CBS \u003ccbs@centos.org\u003e, 2021-10-29-15:04:36, )","subcomponent":"libvirt","thread":"39","timestamp":"2022-06-20T00:18:45.977000Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"hostname: instance","subcomponent":"libvirt","thread":"39","timestamp":"2022-06-20T00:18:45.977000Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"error","msg":"internal error: Child process (dmidecode -q -t 0,1,2,3,4,11,17) unexpected exit status 1: /dev/mem: No such file or directory","pos":"virCommandWait:2749","subcomponent":"libvirt","thread":"39","timestamp":"2022-06-20T00:18:45.977000Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"Connected to libvirt daemon","pos":"libvirt.go:511","timestamp":"2022-06-20T00:18:46.427934Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"Registered libvirt event notify callback","pos":"client.go:509","timestamp":"2022-06-20T00:18:46.438211Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"Marked as ready","pos":"virt-launcher.go:80","timestamp":"2022-06-20T00:18:46.440886Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"Hardware emulation device '/dev/kvm' not present. Using software emulation.","pos":"converter.go:1194","timestamp":"2022-06-20T00:18:47.637971Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"In-kernel virtio-net device emulation '/dev/vhost-net' not present. Falling back to QEMU userland emulation.","pos":"converter.go:1208","timestamp":"2022-06-20T00:18:47.638021Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","kind":"","level":"info","msg":"Executing PreStartHook on VMI pod environment","name":"instance","namespace":"default","pos":"manager.go:513","timestamp":"2022-06-20T00:18:47.641028Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"error","msg":"could not read secret data from source: /var/run/kubevirt-private/secret/cloudinit/networkdata","pos":"cloud-init.go:290","reason":"open /var/run/kubevirt-private/secret/cloudinit/networkdata: no such file or directory","timestamp":"2022-06-20T00:18:47.641472Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"error","msg":"could not read secret data from source: /var/run/kubevirt-private/secret/cloudinit/networkData","pos":"cloud-init.go:290","reason":"open /var/run/kubevirt-private/secret/cloudinit/networkData: no such file or directory","timestamp":"2022-06-20T00:18:47.641525Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","kind":"","level":"info","msg":"Starting PreCloudInitIso hook","name":"instance","namespace":"default","pos":"manager.go:534","timestamp":"2022-06-20T00:18:47.641543Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"Found nameservers in /etc/resolv.conf: \n`\u0000\n","pos":"network.go:286","timestamp":"2022-06-20T00:18:47.644877Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"Found search domains in /etc/resolv.conf: default.svc.cluster.local svc.cluster.local cluster.local paxcb4zidlwe3mfrpko3itqnmd.bx.internal.cloudapp.net","pos":"network.go:287","timestamp":"2022-06-20T00:18:47.645223Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"Starting SingleClientDHCPServer","pos":"server.go:64","timestamp":"2022-06-20T00:18:47.645427Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"Driver cache mode for /var/run/kubevirt-ephemeral-disks/disk-data/boot/disk.qcow2 set to none","pos":"converter.go:456","timestamp":"2022-06-20T00:18:47.658740Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"Driver cache mode for /var/run/kubevirt-ephemeral-disks/cloud-init-data/default/instance/noCloud.iso set to none","pos":"converter.go:456","timestamp":"2022-06-20T00:18:47.665000Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","kind":"","level":"info","msg":"Domain XML generated. Base64 dump PGRvbWFpbiB0eXBlPSJxZW11IiB4bWxuczpxZW11PSJodHRwOi8vbGlidmlydC5vcmcvc2NoZW1hcy9kb21haW4vcWVtdS8xLjAiPgoJPG5hbWU+ZGVmYXVsdF9pbnN0YW5jZTwvbmFtZT4KCTxtZW1vcnkgdW5pdD0iYiI+MjE0NzQ4MzY0ODwvbWVtb3J5PgoJPG9zPgoJCTx0eXBlIGFyY2g9Ing4Nl82NCIgbWFjaGluZT0icTM1Ij5odm08L3R5cGU+CgkJPHNtYmlvcyBtb2RlPSJzeXNpbmZvIj48L3NtYmlvcz4KCTwvb3M+Cgk8c3lzaW5mbyB0eXBlPSJzbWJpb3MiPgoJCTxzeXN0ZW0+CgkJCTxlbnRyeSBuYW1lPSJ1dWlkIj4wOTI4MmM0MC1kZDE2LTU5NTYtOTIxMi1kZWZjNTFjODM3NGY8L2VudHJ5PgoJCQk8ZW50cnkgbmFtZT0ibWFudWZhY3R1cmVyIj5LdWJlVmlydDwvZW50cnk+CgkJCTxlbnRyeSBuYW1lPSJmYW1pbHkiPkt1YmVWaXJ0PC9lbnRyeT4KCQkJPGVudHJ5IG5hbWU9InByb2R1Y3QiPk5vbmU8L2VudHJ5PgoJCQk8ZW50cnkgbmFtZT0ic2t1Ij48L2VudHJ5PgoJCQk8ZW50cnkgbmFtZT0idmVyc2lvbiI+PC9lbnRyeT4KCQk8L3N5c3RlbT4KCQk8Ymlvcz48L2Jpb3M+CgkJPGJhc2VCb2FyZD48L2Jhc2VCb2FyZD4KCQk8Y2hhc3Npcz48L2NoYXNzaXM+Cgk8L3N5c2luZm8+Cgk8ZGV2aWNlcz4KCQk8aW50ZXJmYWNlIHR5cGU9ImV0aGVybmV0Ij4KCQkJPHNvdXJjZT48L3NvdXJjZT4KCQkJPHRhcmdldCBkZXY9InRhcDAiIG1hbmFnZWQ9Im5vIj48L3RhcmdldD4KCQkJPG1vZGVsIHR5cGU9InZpcnRpby1ub24tdHJhbnNpdGlvbmFsIj48L21vZGVsPgoJCQk8bWFjIGFkZHJlc3M9IjAyOjQyOmFjOjExOjAwOjEwIj48L21hYz4KCQkJPG10dSBzaXplPSIxNTAwIj48L210dT4KCQkJPGFsaWFzIG5hbWU9InVhLWRlZmF1bHQiPjwvYWxpYXM+CgkJCTxyb20gZW5hYmxlZD0ibm8iPjwvcm9tPgoJCTwvaW50ZXJmYWNlPgoJCTxjaGFubmVsIHR5cGU9InVuaXgiPgoJCQk8dGFyZ2V0IG5hbWU9Im9yZy5xZW11Lmd1ZXN0X2FnZW50LjAiIHR5cGU9InZpcnRpbyI+PC90YXJnZXQ+CgkJPC9jaGFubmVsPgoJCTxjb250cm9sbGVyIHR5cGU9InVzYiIgaW5kZXg9IjAiIG1vZGVsPSJub25lIj48L2NvbnRyb2xsZXI+CgkJPGNvbnRyb2xsZXIgdHlwZT0ic2NzaSIgaW5kZXg9IjAiIG1vZGVsPSJ2aXJ0aW8tbm9uLXRyYW5zaXRpb25hbCI+PC9jb250cm9sbGVyPgoJCTxjb250cm9sbGVyIHR5cGU9InZpcnRpby1zZXJpYWwiIGluZGV4PSIwIiBtb2RlbD0idmlydGlvLW5vbi10cmFuc2l0aW9uYWwiPjwvY29udHJvbGxlcj4KCQk8dmlkZW8+CgkJCTxtb2RlbCB0eXBlPSJ2Z2EiIGhlYWRzPSIxIiB2cmFtPSIxNjM4NCI+PC9tb2RlbD4KCQk8L3ZpZGVvPgoJCTxncmFwaGljcyB0eXBlPSJ2bmMiPgoJCQk8bGlzdGVuIHR5cGU9InNvY2tldCIgc29ja2V0PSIvdmFyL3J1bi9rdWJldmlydC1wcml2YXRlL2VhN2U3N2Y1LWM2NWUtNGEwNS04NTM3LWJiNWQxMDhiYjViMC92aXJ0LXZuYyI+PC9saXN0ZW4+CgkJPC9ncmFwaGljcz4KCQk8bWVtYmFsbG9vbiBtb2RlbD0idmlydGlvLW5vbi10cmFuc2l0aW9uYWwiPgoJCQk8c3RhdHMgcGVyaW9kPSIxMCI+PC9zdGF0cz4KCQk8L21lbWJhbGxvb24+CgkJPGRpc2sgZGV2aWNlPSJkaXNrIiB0eXBlPSJmaWxlIiBtb2RlbD0idmlydGlvLW5vbi10cmFuc2l0aW9uYWwiPgoJCQk8c291cmNlIGZpbGU9Ii92YXIvcnVuL2t1YmV2aXJ0LWVwaGVtZXJhbC1kaXNrcy9kaXNrLWRhdGEvYm9vdC9kaXNrLnFjb3cyIj48L3NvdXJjZT4KCQkJPHRhcmdldCBidXM9InZpcnRpbyIgZGV2PSJ2ZGEiPjwvdGFyZ2V0PgoJCQk8ZHJpdmVyIGNhY2hlPSJub25lIiBlcnJvcl9wb2xpY3k9InN0b3AiIG5hbWU9InFlbXUiIHR5cGU9InFjb3cyIiBkaXNjYXJkPSJ1bm1hcCI+PC9kcml2ZXI+CgkJCTxhbGlhcyBuYW1lPSJ1YS1ib290Ij48L2FsaWFzPgoJCQk8YmFja2luZ1N0b3JlIHR5cGU9ImZpbGUiPgoJCQkJPGZvcm1hdCB0eXBlPSJxY293MiI+PC9mb3JtYXQ+CgkJCQk8c291cmNlIGZpbGU9Ii92YXIvcnVuL2t1YmV2aXJ0L2NvbnRhaW5lci1kaXNrcy9kaXNrXzAuaW1nIj48L3NvdXJjZT4KCQkJPC9iYWNraW5nU3RvcmU+CgkJPC9kaXNrPgoJCTxkaXNrIGRldmljZT0iZGlzayIgdHlwZT0iZmlsZSIgbW9kZWw9InZpcnRpby1ub24tdHJhbnNpdGlvbmFsIj4KCQkJPHNvdXJjZSBmaWxlPSIvdmFyL3J1bi9rdWJldmlydC1lcGhlbWVyYWwtZGlza3MvY2xvdWQtaW5pdC1kYXRhL2RlZmF1bHQvaW5zdGFuY2Uvbm9DbG91ZC5pc28iPjwvc291cmNlPgoJCQk8dGFyZ2V0IGJ1cz0idmlydGlvIiBkZXY9InZkYiI+PC90YXJnZXQ+CgkJCTxkcml2ZXIgY2FjaGU9Im5vbmUiIGVycm9yX3BvbGljeT0ic3RvcCIgbmFtZT0icWVtdSIgdHlwZT0icmF3IiBkaXNjYXJkPSJ1bm1hcCI+PC9kcml2ZXI+CgkJCTxhbGlhcyBuYW1lPSJ1YS1jbG91ZGluaXQiPjwvYWxpYXM+CgkJPC9kaXNrPgoJCTxzZXJpYWwgdHlwZT0idW5peCI+CgkJCTx0YXJnZXQgcG9ydD0iMCI+PC90YXJnZXQ+CgkJCTxzb3VyY2UgbW9kZT0iYmluZCIgcGF0aD0iL3Zhci9ydW4va3ViZXZpcnQtcHJpdmF0ZS9lYTdlNzdmNS1jNjVlLTRhMDUtODUzNy1iYjVkMTA4YmI1YjAvdmlydC1zZXJpYWwwIj48L3NvdXJjZT4KCQk8L3NlcmlhbD4KCQk8Y29uc29sZSB0eXBlPSJwdHkiPgoJCQk8dGFyZ2V0IHR5cGU9InNlcmlhbCIgcG9ydD0iMCI+PC90YXJnZXQ+CgkJPC9jb25zb2xlPgoJPC9kZXZpY2VzPgoJPG1ldGFkYXRhPgoJCTxrdWJldmlydCB4bWxucz0iaHR0cDovL2t1YmV2aXJ0LmlvIj4KCQkJPHVpZD5lYTdlNzdmNS1jNjVlLTRhMDUtODUzNy1iYjVkMTA4YmI1YjA8L3VpZD4KCQkJPGdyYWNlcGVyaW9kPgoJCQkJPGRlbGV0aW9uR3JhY2VQZXJpb2RTZWNvbmRzPjA8L2RlbGV0aW9uR3JhY2VQZXJpb2RTZWNvbmRzPgoJCQk8L2dyYWNlcGVyaW9kPgoJCTwva3ViZXZpcnQ+Cgk8L21ldGFkYXRhPgoJPGZlYXR1cmVzPgoJCTxhY3BpPjwvYWNwaT4KCTwvZmVhdHVyZXM+Cgk8Y3B1IG1vZGU9Imhvc3QtbW9kZWwiPgoJCTx0b3BvbG9neSBzb2NrZXRzPSIxIiBjb3Jlcz0iMSIgdGhyZWFkcz0iMSI+PC90b3BvbG9neT4KCTwvY3B1PgoJPHZjcHUgcGxhY2VtZW50PSJzdGF0aWMiPjE8L3ZjcHU+Cgk8aW90aHJlYWRzPjE8L2lvdGhyZWFkcz4KPC9kb21haW4+","name":"instance","namespace":"default","pos":"libvirt_helper.go:129","timestamp":"2022-06-20T00:18:47.670433Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"Reaped pid 56 with status 0","pos":"virt-launcher.go:554","timestamp":"2022-06-20T00:18:47.741775Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"Reaped pid 58 with status 9","pos":"virt-launcher.go:554","timestamp":"2022-06-20T00:18:47.897997Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"DomainLifecycle event 0 with reason 0 received","pos":"client.go:435","timestamp":"2022-06-20T00:18:47.924552Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","kind":"","level":"info","msg":"Domain defined.","name":"instance","namespace":"default","pos":"manager.go:860","timestamp":"2022-06-20T00:18:47.924394Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"kubevirt domain status: Shutoff(5):Unknown(0)","pos":"client.go:288","timestamp":"2022-06-20T00:18:47.933709Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"Successfully connected to domain notify socket at /var/run/kubevirt/domain-notify-pipe.sock","pos":"client.go:167","timestamp":"2022-06-20T00:18:47.940508Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"Monitoring loop: rate 1s start timeout 5m1s","pos":"monitor.go:178","timestamp":"2022-06-20T00:18:47.963545Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"Domain name event: default_instance","pos":"client.go:413","timestamp":"2022-06-20T00:18:47.969088Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"generated nocloud iso file /var/run/kubevirt-ephemeral-disks/cloud-init-data/default/instance/noCloud.iso","pos":"cloud-init.go:655","timestamp":"2022-06-20T00:18:48.079337Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"error","msg":"At least one cgroup controller is required: No such device or address","pos":"virCgroupDetectControllers:455","subcomponent":"libvirt","thread":"25","timestamp":"2022-06-20T00:18:48.109000Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"GuestAgentLifecycle event state 2 with reason 1 received","pos":"client.go:492","timestamp":"2022-06-20T00:18:48.211358Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"kubevirt domain status: Paused(3):StartingUp(11)","pos":"client.go:288","timestamp":"2022-06-20T00:18:48.215524Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"Domain name event: default_instance","pos":"client.go:413","timestamp":"2022-06-20T00:18:48.226275Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"DomainLifecycle event 4 with reason 0 received","pos":"client.go:435","timestamp":"2022-06-20T00:18:48.681143Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"DomainLifecycle event 2 with reason 0 received","pos":"client.go:435","timestamp":"2022-06-20T00:18:48.685453Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","kind":"","level":"info","msg":"Domain started.","name":"instance","namespace":"default","pos":"manager.go:888","timestamp":"2022-06-20T00:18:48.688573Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","kind":"","level":"info","msg":"Synced vmi","name":"instance","namespace":"default","pos":"server.go:190","timestamp":"2022-06-20T00:18:48.693312Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"kubevirt domain status: Running(1):Unknown(1)","pos":"client.go:288","timestamp":"2022-06-20T00:18:48.696253Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"Domain name event: default_instance","pos":"client.go:413","timestamp":"2022-06-20T00:18:48.709012Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"kubevirt domain status: Running(1):Unknown(1)","pos":"client.go:288","timestamp":"2022-06-20T00:18:48.712722Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"Domain name event: default_instance","pos":"client.go:413","timestamp":"2022-06-20T00:18:48.715574Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"Hardware emulation device '/dev/kvm' not present. Using software emulation.","pos":"converter.go:1194","timestamp":"2022-06-20T00:18:48.737087Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"In-kernel virtio-net device emulation '/dev/vhost-net' not present. Falling back to QEMU userland emulation.","pos":"converter.go:1208","timestamp":"2022-06-20T00:18:48.737236Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","kind":"","level":"info","msg":"Synced vmi","name":"instance","namespace":"default","pos":"server.go:190","timestamp":"2022-06-20T00:18:48.739512Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"Hardware emulation device '/dev/kvm' not present. Using software emulation.","pos":"converter.go:1194","timestamp":"2022-06-20T00:18:48.846229Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"In-kernel virtio-net device emulation '/dev/vhost-net' not present. Falling back to QEMU userland emulation.","pos":"converter.go:1208","timestamp":"2022-06-20T00:18:48.846909Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","kind":"","level":"info","msg":"Synced vmi","name":"instance","namespace":"default","pos":"server.go:190","timestamp":"2022-06-20T00:18:48.851173Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"Hardware emulation device '/dev/kvm' not present. Using software emulation.","pos":"converter.go:1194","timestamp":"2022-06-20T00:18:48.880962Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"In-kernel virtio-net device emulation '/dev/vhost-net' not present. Falling back to QEMU userland emulation.","pos":"converter.go:1208","timestamp":"2022-06-20T00:18:48.883678Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","kind":"","level":"info","msg":"Synced vmi","name":"instance","namespace":"default","pos":"server.go:190","timestamp":"2022-06-20T00:18:48.888717Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"2022-06-20 00:18:48.103+0000: starting up libvirt version: 7.6.0, package: 6.el8s (CBS \u003ccbs@centos.org\u003e, 2021-10-29-15:04:36, ), qemu version: 6.0.0qemu-kvm-6.0.0-33.el8s, kernel: 5.13.0-1029-azure, hostname: instance","subcomponent":"qemu","timestamp":"2022-06-20T00:18:48.930165Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"LC_ALL=C \\PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin \\HOME=/var/lib/libvirt/qemu/domain-1-default_instance \\XDG_DATA_HOME=/var/lib/libvirt/qemu/domain-1-default_instance/.local/share \\XDG_CACHE_HOME=/var/lib/libvirt/qemu/domain-1-default_instance/.cache \\XDG_CONFIG_HOME=/var/lib/libvirt/qemu/domain-1-default_instance/.config \\/usr/libexec/qemu-kvm \\-name guest=default_instance,debug-threads=on \\-S \\-object '{\"qom-type\":\"secret\",\"id\":\"masterKey0\",\"format\":\"raw\",\"file\":\"/var/lib/libvirt/qemu/domain-1-default_instance/master-key.aes\"}' \\-machine pc-q35-rhel8.5.0,accel=tcg,usb=off,dump-guest-core=off,memory-backend=pc.ram \\-cpu EPYC,acpi=on,ss=on,monitor=on,hypervisor=on,erms=on,mpx=on,pcommit=on,clwb=on,pku=on,la57=on,3dnowext=on,3dnow=on,npt=on,vme=off,fma=off,avx=off,f16c=off,avx2=off,rdseed=off,sha-ni=off,xsavec=off,fxsr-opt=off,misalignsse=off,3dnowprefetch=off,osvw=off,topoext=off,nrip-save=off \\-m 2048 \\-object '{\"qom-type\":\"memory-backend-ram\",\"id\":\"pc.ram\",\"size\":2147483648}' \\-overcommit mem-lock=off \\-smp 1,sockets=1,dies=1,cores=1,threads=1 \\-object '{\"qom-type\":\"iothread\",\"id\":\"iothread1\"}' \\-uuid 09282c40-dd16-5956-9212-defc51c8374f \\-smbios type=1,manufacturer=KubeVirt,product=None,uuid=09282c40-dd16-5956-9212-defc51c8374f,family=KubeVirt \\-no-user-config \\-nodefaults \\-chardev socket,id=charmonitor,fd=19,server=on,wait=off \\-mon chardev=charmonitor,id=monitor,mode=control \\-rtc base=utc \\-no-shutdown \\-boot strict=on \\-device pcie-root-port,port=0x10,chassis=1,id=pci.1,bus=pcie.0,multifunction=on,addr=0x2 \\-device pcie-root-port,port=0x11,chassis=2,id=pci.2,bus=pcie.0,addr=0x2.0x1 \\-device pcie-root-port,port=0x12,chassis=3,id=pci.3,bus=pcie.0,addr=0x2.0x2 \\-device pcie-root-port,port=0x13,chassis=4,id=pci.4,bus=pcie.0,addr=0x2.0x3 \\-device pcie-root-port,port=0x14,chassis=5,id=pci.5,bus=pcie.0,addr=0x2.0x4 \\-device pcie-root-port,port=0x15,chassis=6,id=pci.6,bus=pcie.0,addr=0x2.0x5 \\-device pcie-root-port,port=0x16,chassis=7,id=pci.7,bus=pcie.0,addr=0x2.0x6 \\-device virtio-scsi-pci-non-transitional,id=scsi0,bus=pci.2,addr=0x0 \\-device virtio-serial-pci-non-transitional,id=virtio-serial0,bus=pci.3,addr=0x0 \\-blockdev '{\"driver\":\"file\",\"filename\":\"/var/run/kubevirt/container-disks/disk_0.img\",\"node-name\":\"libvirt-3-storage\",\"cache\":{\"direct\":true,\"no-flush\":false},\"auto-read-only\":true,\"discard\":\"unmap\"}' \\-blockdev '{\"node-name\":\"libvirt-3-format\",\"read-only\":true,\"discard\":\"unmap\",\"cache\":{\"direct\":true,\"no-flush\":false},\"driver\":\"qcow2\",\"file\":\"libvirt-3-storage\"}' \\-blockdev '{\"driver\":\"file\",\"filename\":\"/var/run/kubevirt-ephemeral-disks/disk-data/boot/disk.qcow2\",\"node-name\":\"libvirt-2-storage\",\"cache\":{\"direct\":true,\"no-flush\":false},\"auto-read-only\":true,\"discard\":\"unmap\"}' \\-blockdev '{\"node-name\":\"libvirt-2-format\",\"read-only\":false,\"discard\":\"unmap\",\"cache\":{\"direct\":true,\"no-flush\":false},\"driver\":\"qcow2\",\"file\":\"libvirt-2-storage\",\"backing\":\"libvirt-3-format\"}' \\-device virtio-blk-pci-non-transitional,bus=pci.4,addr=0x0,drive=libvirt-2-format,id=ua-boot,bootindex=1,write-cache=on,werror=stop,rerror=stop \\-blockdev '{\"driver\":\"file\",\"filename\":\"/var/run/kubevirt-ephemeral-disks/cloud-init-data/default/instance/noCloud.iso\",\"node-name\":\"libvirt-1-storage\",\"cache\":{\"direct\":true,\"no-flush\":false},\"auto-read-only\":true,\"discard\":\"unmap\"}' \\-blockdev '{\"node-name\":\"libvirt-1-format\",\"read-only\":false,\"discard\":\"unmap\",\"cache\":{\"direct\":true,\"no-flush\":false},\"driver\":\"raw\",\"file\":\"libvirt-1-storage\"}' \\-device virtio-blk-pci-non-transitional,bus=pci.5,addr=0x0,drive=libvirt-1-format,id=ua-cloudinit,write-cache=on,werror=stop,rerror=stop \\-netdev tap,fd=21,id=hostua-default \\-device virtio-net-pci-non-transitional,host_mtu=1500,netdev=hostua-default,id=ua-default,mac=02:42:ac:11:00:10,bus=pci.1,addr=0x0,romfile= \\-chardev socket,id=charserial0,fd=22,server=on,wait=off \\-device isa-serial,chardev=charserial0,id=serial0 \\-chardev socket,id=charchannel0,fd=23,server=on,wait=off \\-device virtserialport,bus=virtio-serial0.0,nr=1,chardev=charchannel0,id=channel0,name=org.qemu.guest_agent.0 \\-audiodev id=audio1,driver=none \\-vnc vnc=unix:/var/run/kubevirt-private/ea7e77f5-c65e-4a05-8537-bb5d108bb5b0/virt-vnc,audiodev=audio1 \\-device VGA,id=video0,vgamem_mb=16,bus=pcie.0,addr=0x1 \\-device virtio-balloon-pci-non-transitional,id=balloon0,bus=pci.6,addr=0x0 \\-sandbox on,obsolete=deny,elevateprivileges=deny,spawn=deny,resourcecontrol=deny \\-msg timestamp=on","subcomponent":"qemu","timestamp":"2022-06-20T00:18:48.933685Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"Found PID for 09282c40-dd16-5956-9212-defc51c8374f: 68","pos":"monitor.go:139","timestamp":"2022-06-20T00:18:48.971076Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"Hardware emulation device '/dev/kvm' not present. Using software emulation.","pos":"converter.go:1194","timestamp":"2022-06-20T00:19:46.997307Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"In-kernel virtio-net device emulation '/dev/vhost-net' not present. Falling back to QEMU userland emulation.","pos":"converter.go:1208","timestamp":"2022-06-20T00:19:46.997386Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","kind":"","level":"info","msg":"Synced vmi","name":"instance","namespace":"default","pos":"server.go:190","timestamp":"2022-06-20T00:19:47.001274Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"Received signal terminated","pos":"virt-launcher.go:495","timestamp":"2022-06-20T00:24:17.885088Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"Reaped pid 68 with status 0","pos":"virt-launcher.go:554","timestamp":"2022-06-20T00:24:18.012709Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"DomainLifecycle event 6 with reason 2 received","pos":"client.go:435","timestamp":"2022-06-20T00:24:18.029745Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","kind":"","level":"info","msg":"Domain stopped.","name":"instance","namespace":"default","pos":"manager.go:1443","timestamp":"2022-06-20T00:24:18.035843Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","kind":"","level":"info","msg":"Signaled vmi kill","name":"instance","namespace":"default","pos":"server.go:293","timestamp":"2022-06-20T00:24:18.036378Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"DomainLifecycle event 5 with reason 1 received","pos":"client.go:435","timestamp":"2022-06-20T00:24:18.046375Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","kind":"VirtualMachineInstance","level":"info","msg":"Domain XML generated. Base64 dump <domain type="qemu">
	<name>default_instance</name>
	<uuid>09282c40-dd16-5956-9212-defc51c8374f</uuid>
	<memory unit="KiB">2097152</memory>
	<os>
		<type arch="x86_64" machine="pc-q35-rhel8.5.0">hvm</type>
		<smbios mode="sysinfo"></smbios>
		<boot dev="hd"></boot>
	</os>
	<sysinfo type="smbios">
		<system>
			<entry name="manufacturer">KubeVirt</entry>
			<entry name="product">None</entry>
			<entry name="uuid">09282c40-dd16-5956-9212-defc51c8374f</entry>
			<entry name="family">KubeVirt</entry>
		</system>
		<bios></bios>
		<baseBoard></baseBoard>
		<chassis></chassis>
	</sysinfo>
	<devices>
		<emulator>/usr/libexec/qemu-kvm</emulator>
		<interface type="ethernet">
			<address type="pci" domain="0x0000" bus="0x01" slot="0x00" function="0x0"></address>
			<source></source>
			<target dev="tap0" managed="no"></target>
			<model type="virtio-non-transitional"></model>
			<mac address="02:42:ac:11:00:10"></mac>
			<mtu size="1500"></mtu>
			<alias name="ua-default"></alias>
			<rom enabled="no"></rom>
		</interface>
		<channel type="unix">
			<source mode="bind" path="/var/lib/libvirt/qemu/channel/target/domain-1-default_instance/org.qemu.guest_agent.0"></source>
			<target name="org.qemu.guest_agent.0" type="virtio" state="disconnected"></target>
		</channel>
		<controller type="usb" index="0" model="none">
			<alias name="usb"></alias>
		</controller>
		<controller type="scsi" index="0" model="virtio-non-transitional">
			<alias name="scsi0"></alias>
			<address type="pci" domain="0x0000" bus="0x02" slot="0x00" function="0x0"></address>
		</controller>
		<controller type="virtio-serial" index="0" model="virtio-non-transitional">
			<alias name="virtio-serial0"></alias>
			<address type="pci" domain="0x0000" bus="0x03" slot="0x00" function="0x0"></address>
		</controller>
		<controller type="sata" index="0">
			<alias name="ide"></alias>
			<address type="pci" domain="0x0000" bus="0x00" slot="0x1f" function="0x2"></address>
		</controller>
		<controller type="pci" index="0" model="pcie-root">
			<alias name="pcie.0"></alias>
		</controller>
		<controller type="pci" index="1" model="pcie-root-port">
			<alias name="pci.1"></alias>
			<address type="pci" domain="0x0000" bus="0x00" slot="0x02" function="0x0"></address>
		</controller>
		<controller type="pci" index="2" model="pcie-root-port">
			<alias name="pci.2"></alias>
			<address type="pci" domain="0x0000" bus="0x00" slot="0x02" function="0x1"></address>
		</controller>
		<controller type="pci" index="3" model="pcie-root-port">
			<alias name="pci.3"></alias>
			<address type="pci" domain="0x0000" bus="0x00" slot="0x02" function="0x2"></address>
		</controller>
		<controller type="pci" index="4" model="pcie-root-port">
			<alias name="pci.4"></alias>
			<address type="pci" domain="0x0000" bus="0x00" slot="0x02" function="0x3"></address>
		</controller>
		<controller type="pci" index="5" model="pcie-root-port">
			<alias name="pci.5"></alias>
			<address type="pci" domain="0x0000" bus="0x00" slot="0x02" function="0x4"></address>
		</controller>
		<controller type="pci" index="6" model="pcie-root-port">
			<alias name="pci.6"></alias>
			<address type="pci" domain="0x0000" bus="0x00" slot="0x02" function="0x5"></address>
		</controller>
		<controller type="pci" index="7" model="pcie-root-port">
			<alias name="pci.7"></alias>
			<address type="pci" domain="0x0000" bus="0x00" slot="0x02" function="0x6"></address>
		</controller>
		<video>
			<model type="vga" heads="1" vram="16384"></model>
		</video>
		<graphics type="vnc">
			<listen type="socket" socket="/var/run/kubevirt-private/ea7e77f5-c65e-4a05-8537-bb5d108bb5b0/virt-vnc"></listen>
		</graphics>
		<memballoon model="virtio-non-transitional">
			<stats period="10"></stats>
			<address type="pci" domain="0x0000" bus="0x06" slot="0x00" function="0x0"></address>
		</memballoon>
		<disk device="disk" type="file" model="virtio-non-transitional">
			<source file="/var/run/kubevirt-ephemeral-disks/disk-data/boot/disk.qcow2"></source>
			<target bus="virtio" dev="vda"></target>
			<driver cache="none" error_policy="stop" name="qemu" type="qcow2" discard="unmap"></driver>
			<alias name="ua-boot"></alias>
			<backingStore type="file">
				<format type="qcow2"></format>
				<source file="/var/run/kubevirt/container-disks/disk_0.img"></source>
			</backingStore>
			<address type="pci" domain="0x0000" bus="0x04" slot="0x00" function="0x0"></address>
		</disk>
		<disk device="disk" type="file" model="virtio-non-transitional">
			<source file="/var/run/kubevirt-ephemeral-disks/cloud-init-data/default/instance/noCloud.iso"></source>
			<target bus="virtio" dev="vdb"></target>
			<driver cache="none" error_policy="stop" name="qemu" type="raw" discard="unmap"></driver>
			<alias name="ua-cloudinit"></alias>
			<backingStore></backingStore>
			<address type="pci" domain="0x0000" bus="0x05" slot="0x00" function="0x0"></address>
		</disk>
		<input type="mouse" bus="ps2">
			<alias name="input0"></alias>
		</input>
		<input type="keyboard" bus="ps2">
			<alias name="input1"></alias>
		</input>
		<serial type="unix">
			<target port="0"></target>
			<source mode="bind" path="/var/run/kubevirt-private/ea7e77f5-c65e-4a05-8537-bb5d108bb5b0/virt-serial0"></source>
			<alias name="serial0"></alias>
		</serial>
		<console type="unix">
			<target type="serial" port="0"></target>
			<source mode="bind" path="/var/run/kubevirt-private/ea7e77f5-c65e-4a05-8537-bb5d108bb5b0/virt-serial0"></source>
			<alias name="serial0"></alias>
		</console>
	</devices>
	<clock offset="utc"></clock>
	<resource>
		<partition>/machine</partition>
	</resource>
	<metadata>
		<kubevirt xmlns="http://kubevirt.io">
			<uid>ea7e77f5-c65e-4a05-8537-bb5d108bb5b0</uid>
			<graceperiod>
				<deletionGracePeriodSeconds>0</deletionGracePeriodSeconds>
				<markedForGracefulShutdown>true</markedForGracefulShutdown>
			</graceperiod>
		</kubevirt>
	</metadata>
	<features>
		<acpi></acpi>
	</features>
	<cpu mode="custom">
		<model>EPYC</model>
		<feature name="acpi" policy="require"></feature>
		<feature name="ss" policy="require"></feature>
		<feature name="monitor" policy="require"></feature>
		<feature name="hypervisor" policy="require"></feature>
		<feature name="erms" policy="require"></feature>
		<feature name="mpx" policy="require"></feature>
		<feature name="pcommit" policy="require"></feature>
		<feature name="clwb" policy="require"></feature>
		<feature name="pku" policy="require"></feature>
		<feature name="la57" policy="require"></feature>
		<feature name="3dnowext" policy="require"></feature>
		<feature name="3dnow" policy="require"></feature>
		<feature name="npt" policy="require"></feature>
		<feature name="vme" policy="disable"></feature>
		<feature name="fma" policy="disable"></feature>
		<feature name="avx" policy="disable"></feature>
		<feature name="f16c" policy="disable"></feature>
		<feature name="avx2" policy="disable"></feature>
		<feature name="rdseed" policy="disable"></feature>
		<feature name="sha-ni" policy="disable"></feature>
		<feature name="xsavec" policy="disable"></feature>
		<feature name="fxsr_opt" policy="disable"></feature>
		<feature name="misalignsse" policy="disable"></feature>
		<feature name="3dnowprefetch" policy="disable"></feature>
		<feature name="osvw" policy="disable"></feature>
		<feature name="topoext" policy="disable"></feature>
		<feature name="nrip-save" policy="disable"></feature>
		<topology sockets="1" cores="1" threads="1"></topology>
	</cpu>
	<vcpu placement="static">1</vcpu>
	<iothreads>1</iothreads>
</domain>","name":"instance","namespace":"default","pos":"libvirt_helper.go:129","timestamp":"2022-06-20T00:24:18.051048Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","kind":"","level":"info","msg":"Domain not running, paused or shut down, nothing to do.","name":"instance","namespace":"default","pos":"manager.go:1447","timestamp":"2022-06-20T00:24:18.052253Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","kind":"","level":"info","msg":"Signaled vmi kill","name":"instance","namespace":"default","pos":"server.go:293","timestamp":"2022-06-20T00:24:18.052592Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"kubevirt domain status: Shutoff(5):Destroyed(2)","pos":"client.go:288","timestamp":"2022-06-20T00:24:18.053037Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"Domain name event: default_instance","pos":"client.go:413","timestamp":"2022-06-20T00:24:18.066974Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"DomainLifecycle event 0 with reason 1 received","pos":"client.go:435","timestamp":"2022-06-20T00:24:18.071268Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","kind":"VirtualMachineInstance","level":"info","msg":"Successfully signaled graceful shutdown","name":"instance","namespace":"default","pos":"virt-launcher.go:464","timestamp":"2022-06-20T00:24:18.071784Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"Process 09282c40-dd16-5956-9212-defc51c8374f and pid 68 is gone!","pos":"monitor.go:148","timestamp":"2022-06-20T00:24:18.080208Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"Waiting on final notifications to be sent to virt-handler.","pos":"virt-launcher.go:277","timestamp":"2022-06-20T00:24:18.080262Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"kubevirt domain status: Shutoff(5):Destroyed(2)","pos":"client.go:288","timestamp":"2022-06-20T00:24:18.074512Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"Domain name event: default_instance","pos":"client.go:413","timestamp":"2022-06-20T00:24:18.085896Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","kind":"","level":"info","msg":"Domain undefined.","name":"instance","namespace":"default","pos":"manager.go:1470","timestamp":"2022-06-20T00:24:18.089516Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","kind":"","level":"info","msg":"Signaled vmi deletion","name":"instance","namespace":"default","pos":"server.go:329","timestamp":"2022-06-20T00:24:18.089661Z","uid":"ea7e77f5-c65e-4a05-8537-bb5d108bb5b0"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"DomainLifecycle event 1 with reason 0 received","pos":"client.go:435","timestamp":"2022-06-20T00:24:18.089610Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"error","msg":"failed to get domain metadata","pos":"libvirt_helper.go:167","reason":"virError(Code=42, Domain=10, Message='Domain not found: no domain with matching uuid '09282c40-dd16-5956-9212-defc51c8374f' (default_instance)')","timestamp":"2022-06-20T00:24:18.091006Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"kubevirt domain status: Shutoff(5):Destroyed(2)","pos":"client.go:288","timestamp":"2022-06-20T00:24:18.091558Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"Domain name event: default_instance","pos":"client.go:413","timestamp":"2022-06-20T00:24:18.097501Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"Domain name event: ","pos":"client.go:413","timestamp":"2022-06-20T00:24:18.099477Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"Final Delete notification sent","pos":"virt-launcher.go:292","timestamp":"2022-06-20T00:24:18.099507Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"stopping cmd server","pos":"server.go:580","timestamp":"2022-06-20T00:24:18.099976Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"error","msg":"timeout on stopping the cmd server, continuing anyway.","pos":"server.go:591","timestamp":"2022-06-20T00:24:19.100517Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"Exiting...","pos":"virt-launcher.go:524","timestamp":"2022-06-20T00:24:19.100614Z"}
default/virt-launcher-instance-c8n66[compute]: {"component":"virt-launcher","level":"info","msg":"Reaped pid 14 with status 0","pos":"virt-launcher.go:554","timestamp":"2022-06-20T00:24:19.106035Z"}
default/molecule-knmpm[molecule]: molecule_kubevirt/test/test_init.py . [ 40%]
default/molecule-knmpm[molecule]: molecule_kubevirt/test/test_scenario_tests.py ... [100%]
default/molecule-knmpm[molecule]:
default/molecule-knmpm[molecule]: ============================= slowest 10 durations =============================
default/molecule-knmpm[molecule]: 373.98s call molecule_kubevirt/test/test_init.py::test_command_init_and_test_scenario
default/molecule-knmpm[molecule]: 23.19s setup molecule_kubevirt/test/test_scenario_tests.py::TestClass::test_instance_spec[kube-public-instance-full-instance-full-notmolecule]
default/molecule-knmpm[molecule]: 21.54s teardown molecule_kubevirt/test/test_scenario_tests.py::TestClass::test_instance_spec[default-instance-running-false--molecule]
default/molecule-knmpm[molecule]: 0.21s call molecule_kubevirt/test/test_scenario_tests.py::TestClass::test_instance_spec[default-instance-almost-default-instance-almost-default-molecule]
default/molecule-knmpm[molecule]: 0.20s call molecule_kubevirt/test/test_scenario_tests.py::TestClass::test_instance_spec[kube-public-instance-full-instance-full-notmolecule]
default/molecule-knmpm[molecule]: 0.10s call molecule_kubevirt/test/test_scenario_tests.py::TestClass::test_instance_spec[default-instance-running-false--molecule]
default/molecule-knmpm[molecule]: 0.02s call molecule_kubevirt/test/test_driver.py::test_driver_is_detected
default/molecule-knmpm[molecule]:
default/molecule-knmpm[molecule]: (3 durations < 0.005s hidden. Use -vv to show these durations.)
default/molecule-knmpm[molecule]: ======================== 5 passed in 420.67s (0:07:00) =========================
default/molecule-knmpm[molecule]: ___________________________________ summary ____________________________________
default/molecule-knmpm[molecule]: py38-ansible_4: commands succeeded
default/molecule-knmpm[molecule]: congratulations :)
default/molecule-knmpm[molecule]: configmap/molecule-result created
default/molecule-knmpm[molecule]: configmap "molecule-job-running" deleted
[ 0.000000] Linux version 5.13.0-1029-azure (buildd@lcy02-amd64-051) (gcc (Ubuntu 9.4.0-1ubuntu1~20.04.1) 9.4.0, GNU ld (GNU Binutils for Ubuntu) 2.34) #34~20.04.1-Ubuntu SMP Thu Jun 9 12:37:07 UTC 2022 (Ubuntu 5.13.0-1029.34~20.04.1-azure 5.13.19)
[ 0.000000] Command line: BOOT_IMAGE=/boot/vmlinuz-5.13.0-1029-azure root=PARTUUID=2b1f5b8e-4041-4065-af1c-792f94a6d205 ro console=tty1 console=ttyS0 earlyprintk=ttyS0 panic=-1
[ 0.000000] KERNEL supported cpus:
[ 0.000000] Intel GenuineIntel
[ 0.000000] AMD AuthenticAMD
[ 0.000000] Hygon HygonGenuine
[ 0.000000] Centaur CentaurHauls
[ 0.000000] zhaoxin Shanghai
[ 0.000000] x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers'
[ 0.000000] x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers'
[ 0.000000] x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers'
[ 0.000000] x86/fpu: Supporting XSAVE feature 0x008: 'MPX bounds registers'
[ 0.000000] x86/fpu: Supporting XSAVE feature 0x010: 'MPX CSR'
[ 0.000000] x86/fpu: Supporting XSAVE feature 0x020: 'AVX-512 opmask'
[ 0.000000] x86/fpu: Supporting XSAVE feature 0x040: 'AVX-512 Hi256'
[ 0.000000] x86/fpu: Supporting XSAVE feature 0x080: 'AVX-512 ZMM_Hi256'
[ 0.000000] x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256
[ 0.000000] x86/fpu: xstate_offset[3]: 832, xstate_sizes[3]: 64
[ 0.000000] x86/fpu: xstate_offset[4]: 896, xstate_sizes[4]: 64
[ 0.000000] x86/fpu: xstate_offset[5]: 960, xstate_sizes[5]: 64
[ 0.000000] x86/fpu: xstate_offset[6]: 1024, xstate_sizes[6]: 512
[ 0.000000] x86/fpu: xstate_offset[7]: 1536, xstate_sizes[7]: 1024
[ 0.000000] x86/fpu: Enabled xstate features 0xff, context size is 2560 bytes, using 'compacted' format.
[ 0.000000] BIOS-provided physical RAM map:
[ 0.000000] BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable
[ 0.000000] BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved
[ 0.000000] BIOS-e820: [mem 0x00000000000e0000-0x00000000000fffff] reserved
[ 0.000000] BIOS-e820: [mem 0x0000000000100000-0x000000003ffeffff] usable
[ 0.000000] BIOS-e820: [mem 0x000000003fff0000-0x000000003fffefff] ACPI data
[ 0.000000] BIOS-e820: [mem 0x000000003ffff000-0x000000003fffffff] ACPI NVS
[ 0.000000] BIOS-e820: [mem 0x0000000100000000-0x000000027fffffff] usable
[ 0.000000] printk: bootconsole [earlyser0] enabled
[ 0.000000] NX (Execute Disable) protection: active
[ 0.000000] SMBIOS 2.3 present.
[ 0.000000] DMI: Microsoft Corporation Virtual Machine/Virtual Machine, BIOS 090008 12/07/2018
[ 0.000000] Hypervisor detected: Microsoft Hyper-V
[ 0.000000] Hyper-V: privilege flags low 0x2e7f, high 0x3880b0, hints 0x60c2c, misc 0xed7b2
[ 0.000000] Hyper-V Host Build:18362-10.0-3-0.3446
[ 0.000000] Hyper-V: LAPIC Timer Frequency: 0xc3500
[ 0.000000] Hyper-V: Using hypercall for remote TLB flush
[ 0.000000] clocksource: hyperv_clocksource_tsc_page: mask: 0xffffffffffffffff max_cycles: 0x24e6a1710, max_idle_ns: 440795202120 ns
[ 0.000003] tsc: Marking TSC unstable due to running on Hyper-V
[ 0.003947] tsc: Detected 2095.196 MHz processor
[ 0.007461] e820: update [mem 0x00000000-0x00000fff] usable ==> reserved
[ 0.007465] e820: remove [mem 0x000a0000-0x000fffff] usable
[ 0.007471] last_pfn = 0x280000 max_arch_pfn = 0x400000000
[ 0.011169] x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT
[ 0.015674] e820: update [mem 0x40000000-0xffffffff] usable ==> reserved
[ 0.017712] last_pfn = 0x3fff0 max_arch_pfn = 0x400000000
[ 0.037293] found SMP MP-table at [mem 0x000ff780-0x000ff78f]
[ 0.041879] Using GB pages for direct mapping
[ 0.045225] ACPI: Early table checksum verification disabled
[ 0.048903] ACPI: RSDP 0x00000000000F5C00 000014 (v00 ACPIAM)
[ 0.053107] ACPI: RSDT 0x000000003FFF0000 000040 (v01 VRTUAL MICROSFT 12001807 MSFT 00000097)
[ 0.059059] ACPI: FACP 0x000000003FFF0200 000081 (v02 VRTUAL MICROSFT 12001807 MSFT 00000097)
[ 0.064931] ACPI: DSDT 0x000000003FFF1D24 003CD5 (v01 MSFTVM MSFTVM02 00000002 INTL 02002026)
[ 0.070524] ACPI: FACS 0x000000003FFFF000 000040
[ 0.073665] ACPI: WAET 0x000000003FFF1A80 000028 (v01 VRTUAL MICROSFT 12001807 MSFT 00000097)
[ 0.079293] ACPI: SLIC 0x000000003FFF1AC0 000176 (v01 VRTUAL MICROSFT 12001807 MSFT 00000097)
[ 0.084927] ACPI: OEM0 0x000000003FFF1CC0 000064 (v01 VRTUAL MICROSFT 12001807 MSFT 00000097)
[ 0.090581] ACPI: SRAT 0x000000003FFF0800 000140 (v02 VRTUAL MICROSFT 00000001 MSFT 00000001)
[ 0.096149] ACPI: APIC 0x000000003FFF0300 000062 (v01 VRTUAL MICROSFT 12001807 MSFT 00000097)
[ 0.101541] ACPI: OEMB 0x000000003FFFF040 000064 (v01 VRTUAL MICROSFT 12001807 MSFT 00000097)
[ 0.107261] ACPI: Reserving FACP table memory at [mem 0x3fff0200-0x3fff0280]
[ 0.111937] ACPI: Reserving DSDT table memory at [mem 0x3fff1d24-0x3fff59f8]
[ 0.116717] ACPI: Reserving FACS table memory at [mem 0x3ffff000-0x3ffff03f]
[ 0.122401] ACPI: Reserving WAET table memory at [mem 0x3fff1a80-0x3fff1aa7]
[ 0.127642] ACPI: Reserving SLIC table memory at [mem 0x3fff1ac0-0x3fff1c35]
[ 0.132692] ACPI: Reserving OEM0 table memory at [mem 0x3fff1cc0-0x3fff1d23]
[ 0.137512] ACPI: Reserving SRAT table memory at [mem 0x3fff0800-0x3fff093f]
[ 0.142124] ACPI: Reserving APIC table memory at [mem 0x3fff0300-0x3fff0361]
[ 0.146929] ACPI: Reserving OEMB table memory at [mem 0x3ffff040-0x3ffff0a3]
[ 0.151563] ACPI: Local APIC address 0xfee00000
[ 0.151648] SRAT: PXM 0 -> APIC 0x00 -> Node 0
[ 0.153929] SRAT: PXM 0 -> APIC 0x01 -> Node 0
[ 0.156022] ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x3fffffff] hotplug
[ 0.159039] ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000-0x27fffffff] hotplug
[ 0.162709] ACPI: SRAT: Node 0 PXM 0 [mem 0x280200000-0xfdfffffff] hotplug
[ 0.167239] ACPI: SRAT: Node 0 PXM 0 [mem 0x1000000000-0xffffffffff] hotplug
[ 0.171701] ACPI: SRAT: Node 0 PXM 0 [mem 0x10000200000-0x1ffffffffff] hotplug
[ 0.176391] ACPI: SRAT: Node 0 PXM 0 [mem 0x20000200000-0x3ffffffffff] hotplug
[ 0.181214] NUMA: Node 0 [mem 0x00000000-0x3fffffff] + [mem 0x100000000-0x27fffffff] -> [mem 0x00000000-0x27fffffff]
[ 0.188131] NODE_DATA(0) allocated [mem 0x27ffd6000-0x27fffffff]
[ 0.192416] Zone ranges:
[ 0.194120] DMA [mem 0x0000000000001000-0x0000000000ffffff]
[ 0.198051] DMA32 [mem 0x0000000001000000-0x00000000ffffffff]
[ 0.202033] Normal [mem 0x0000000100000000-0x000000027fffffff]
[ 0.206140] Device empty
[ 0.208000] Movable zone start for each node
[ 0.211357] Early memory node ranges
[ 0.213920] node 0: [mem 0x0000000000001000-0x000000000009efff]
[ 0.217904] node 0: [mem 0x0000000000100000-0x000000003ffeffff]
[ 0.221816] node 0: [mem 0x0000000100000000-0x000000027fffffff]
[ 0.225853] Initmem setup node 0 [mem 0x0000000000001000-0x000000027fffffff]
[ 0.230517] On node 0 totalpages: 1834894
[ 0.230519] DMA zone: 64 pages used for memmap
[ 0.230520] DMA zone: 158 pages reserved
[ 0.230521] DMA zone: 3998 pages, LIFO batch:0
[ 0.230522] DMA32 zone: 4032 pages used for memmap
[ 0.230523] DMA32 zone: 258032 pages, LIFO batch:63
[ 0.230524] Normal zone: 24576 pages used for memmap
[ 0.230525] Normal zone: 1572864 pages, LIFO batch:63
[ 0.230528] On node 0, zone DMA: 1 pages in unavailable ranges
[ 0.230557] On node 0, zone DMA: 97 pages in unavailable ranges
[ 0.246881] On node 0, zone Normal: 16 pages in unavailable ranges
[ 0.255245] ACPI: PM-Timer IO Port: 0x408
[ 0.262257] ACPI: Local APIC address 0xfee00000
[ 0.262266] ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1])
[ 0.266656] IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23
[ 0.270967] ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl)
[ 0.274506] ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level)
[ 0.278668] ACPI: IRQ0 used by override.
[ 0.278669] ACPI: IRQ9 used by override.
[ 0.278672] Using ACPI (MADT) for SMP configuration information
[ 0.282268] smpboot: Allowing 2 CPUs, 0 hotplug CPUs
[ 0.285528] PM: hibernation: Registered nosave memory: [mem 0x00000000-0x00000fff]
[ 0.290404] PM: hibernation: Registered nosave memory: [mem 0x0009f000-0x0009ffff]
[ 0.295457] PM: hibernation: Registered nosave memory: [mem 0x000a0000-0x000dffff]
[ 0.300336] PM: hibernation: Registered nosave memory: [mem 0x000e0000-0x000fffff]
[ 0.305240] PM: hibernation: Registered nosave memory: [mem 0x3fff0000-0x3fffefff]
[ 0.310017] PM: hibernation: Registered nosave memory: [mem 0x3ffff000-0x3fffffff]
[ 0.315020] PM: hibernation: Registered nosave memory: [mem 0x40000000-0xffffffff]
[ 0.319897] [mem 0x40000000-0xffffffff] available for PCI devices
[ 0.323855] Booting paravirtualized kernel on Hyper-V
[ 0.327232] clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 7645519600211568 ns
[ 0.334359] setup_percpu: NR_CPUS:8192 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1
[ 0.340195] percpu: Embedded 63 pages/cpu s221184 r8192 d28672 u1048576
[ 0.345437] pcpu-alloc: s221184 r8192 d28672 u1048576 alloc=1*2097152
[ 0.345440] pcpu-alloc: [0] 0 1
[ 0.345457] Hyper-V: PV spinlocks enabled
[ 0.348191] PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear)
[ 0.353030] Built 1 zonelists, mobility grouping on. Total pages: 1806064
[ 0.358293] Policy zone: Normal
[ 0.360395] Kernel command line: BOOT_IMAGE=/boot/vmlinuz-5.13.0-1029-azure root=PARTUUID=2b1f5b8e-4041-4065-af1c-792f94a6d205 ro console=tty1 console=ttyS0 earlyprintk=ttyS0 panic=-1
[ 0.372818] Dentry cache hash table entries: 1048576 (order: 11, 8388608 bytes, linear)
[ 0.379554] Inode-cache hash table entries: 524288 (order: 10, 4194304 bytes, linear)
[ 0.384857] mem auto-init: stack:off, heap alloc:on, heap free:off
[ 0.419629] Memory: 7105684K/7339576K available (14346K kernel code, 3432K rwdata, 9780K rodata, 2608K init, 6104K bss, 233632K reserved, 0K cma-reserved)
[ 0.431854] random: get_random_u64 called from __kmem_cache_create+0x2d/0x440 with crng_init=0
[ 0.431953] SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1
[ 0.442442] Kernel/User page tables isolation: enabled
[ 0.446133] ftrace: allocating 46678 entries in 183 pages
[ 0.467444] ftrace: allocated 183 pages with 6 groups
[ 0.471091] rcu: Hierarchical RCU implementation.
[ 0.474330] rcu: RCU restricting CPUs from NR_CPUS=8192 to nr_cpu_ids=2.
[ 0.479241] Rude variant of Tasks RCU enabled.
[ 0.482380] Tracing variant of Tasks RCU enabled.
[ 0.485788] rcu: RCU calculated value of scheduler-enlistment delay is 25 jiffies.
[ 0.491001] rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2
[ 0.500164] NR_IRQS: 524544, nr_irqs: 440, preallocated irqs: 16
[ 0.505278] random: crng done (trusting CPU's manufacturer)
[ 0.514815] Console: colour VGA+ 80x25
[ 0.594069] printk: console [tty1] enabled
[ 0.597555] printk: console [ttyS0] enabled
[ 0.603581] printk: bootconsole [earlyser0] disabled
[ 0.610807] ACPI: Core revision 20210331
[ 0.614069] APIC: Switch to symmetric I/O mode setup
[ 0.618311] Hyper-V: Using IPI hypercalls
[ 0.622119] Hyper-V: Using enlightened APIC (xapic mode)
[ 0.634002] ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1
[ 0.643239] Calibrating delay loop (skipped), value calculated using timer frequency.. 4190.39 BogoMIPS (lpj=8380784)
[ 0.647237] pid_max: default: 32768 minimum: 301
[ 0.651258] LSM: Security Framework initializing
[ 0.654986] Yama: becoming mindful.
[ 0.655236] AppArmor: AppArmor initialized
[ 0.655272] Mount-cache hash table entries: 16384 (order: 5, 131072 bytes, linear)
[ 0.659236] Mountpoint-cache hash table entries: 16384 (order: 5, 131072 bytes, linear)
[ 0.669065] Last level iTLB entries: 4KB 64, 2MB 8, 4MB 8
[ 0.671237] Last level dTLB entries: 4KB 64, 2MB 0, 4MB 0, 1GB 4
[ 0.675236] Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization
[ 0.675237] Spectre V2 : Mitigation: Retpolines
[ 0.678825] Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch
[ 0.679236] Speculative Store Bypass: Vulnerable
[ 0.682742] TAA: Mitigation: Clear CPU buffers
[ 0.683236] MDS: Mitigation: Clear CPU buffers
[ 0.689835] Freeing SMP alternatives memory: 40K
[ 0.691484] smpboot: CPU0: Intel(R) Xeon(R) Platinum 8171M CPU @ 2.60GHz (family: 0x6, model: 0x55, stepping: 0x4)
[ 0.695403] Performance Events: unsupported p6 CPU model 85 no PMU driver, software events only.
[ 0.699272] rcu: Hierarchical SRCU implementation.
[ 0.703385] NMI watchdog: Perf NMI watchdog permanently disabled
[ 0.707291] smp: Bringing up secondary CPUs ...
[ 0.711140] x86: Booting SMP configuration:
[ 0.711240] .... node #0, CPUs: #1
[ 0.711786] smp: Brought up 1 node, 2 CPUs
[ 0.718887] smpboot: Max logical packages: 1
[ 0.719240] smpboot: Total of 2 processors activated (8380.78 BogoMIPS)
[ 0.723671] devtmpfs: initialized
[ 0.727236] x86/mm: Memory block size: 128MB
[ 0.731825] PM: Registering ACPI NVS region [mem 0x3ffff000-0x3fffffff] (4096 bytes)
[ 0.739286] clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 7645041785100000 ns
[ 0.751255] futex hash table entries: 512 (order: 3, 32768 bytes, linear)
[ 0.755406] pinctrl core: initialized pinctrl subsystem
[ 0.760151] PM: RTC time: 00:09:29, date: 2022-06-20
[ 0.767428] NET: Registered protocol family 16
[ 0.771400] DMA: preallocated 1024 KiB GFP_KERNEL pool for atomic allocations
[ 0.775472] DMA: preallocated 1024 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations
[ 0.783434] DMA: preallocated 1024 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations
[ 0.787251] audit: initializing netlink subsys (disabled)
[ 0.791295] audit: type=2000 audit(1655683769.152:1): state=initialized audit_enabled=0 res=1
[ 0.791471] thermal_sys: Registered thermal governor 'fair_share'
[ 0.795242] thermal_sys: Registered thermal governor 'bang_bang'
[ 0.799239] thermal_sys: Registered thermal governor 'step_wise'
[ 0.803238] thermal_sys: Registered thermal governor 'user_space'
[ 0.807239] thermal_sys: Registered thermal governor 'power_allocator'
[ 0.811247] EISA bus registered
[ 0.817951] cpuidle: using governor ladder
[ 0.819241] cpuidle: using governor menu
[ 0.822949] ACPI: bus type PCI registered
[ 0.823259] acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5
[ 0.827774] PCI: Using configuration type 1 for base access
[ 0.832657] Kprobes globally optimized
[ 0.835299] HugeTLB registered 1.00 GiB page size, pre-allocated 0 pages
[ 0.839247] HugeTLB registered 2.00 MiB page size, pre-allocated 0 pages
[ 0.847322] ACPI: Added _OSI(Module Device)
[ 0.851241] ACPI: Added _OSI(Processor Device)
[ 0.855044] ACPI: Added _OSI(3.0 _SCP Extensions)
[ 0.859243] ACPI: Added _OSI(Processor Aggregator Device)
[ 0.863240] ACPI: Added _OSI(Linux-Dell-Video)
[ 0.866962] ACPI: Added _OSI(Linux-Lenovo-NV-HDMI-Audio)
[ 0.871240] ACPI: Added _OSI(Linux-HPI-Hybrid-Graphics)
[ 0.876515] ACPI: 1 ACPI AML tables successfully acquired and loaded
[ 0.881390] ACPI: Interpreter enabled
[ 0.887250] ACPI: (supports S0 S5)
[ 0.890348] ACPI: Using IOAPIC for interrupt routing
[ 0.891261] PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug
[ 0.899471] ACPI: Enabled 1 GPEs in block 00 to 0F
[ 0.922192] ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff])
[ 0.927244] acpi PNP0A03:00: _OSC: OS supports [ASPM ClockPM Segments MSI HPX-Type3]
[ 0.935248] acpi PNP0A03:00: fail to add MMCONFIG information, can't access extended PCI configuration space under this bridge.
[ 0.943407] PCI host bridge to bus 0000:00
[ 0.946747] pci_bus 0000:00: root bus resource [bus 00-ff]
[ 0.951241] pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window]
[ 0.959241] pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window]
[ 0.963248] pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window]
[ 0.971277] pci_bus 0000:00: root bus resource [mem 0x40000000-0xfffbffff window]
[ 0.975239] pci_bus 0000:00: root bus resource [mem 0xfe0000000-0xfffffffff window]
[ 0.983488] pci 0000:00:00.0: [8086:7192] type 00 class 0x060000
[ 0.991170] pci 0000:00:07.0: [8086:7110] type 00 class 0x060100
[ 0.998980] pci 0000:00:07.1: [8086:7111] type 00 class 0x010180
[ 1.005837] pci 0000:00:07.1: reg 0x20: [io 0xffa0-0xffaf]
[ 1.012233] pci 0000:00:07.1: legacy IDE quirk: reg 0x10: [io 0x01f0-0x01f7]
[ 1.015239] pci 0000:00:07.1: legacy IDE quirk: reg 0x14: [io 0x03f6]
[ 1.023239] pci 0000:00:07.1: legacy IDE quirk: reg 0x18: [io 0x0170-0x0177]
[ 1.027239] pci 0000:00:07.1: legacy IDE quirk: reg 0x1c: [io 0x0376]
[ 1.031925] * Found PM-Timer Bug on the chipset. Due to workarounds for a bug,
* this clock source is slow. Consider trying other clock sources
[ 1.043241] pci 0000:00:07.3: acpi_pm_check_blacklist+0x0/0x20 took 11718 usecs
[ 1.047239] pci 0000:00:07.3: [8086:7113] type 00 class 0x068000
[ 1.058480] pci 0000:00:07.3: quirk: [io 0x0400-0x043f] claimed by PIIX4 ACPI
[ 1.064526] pci 0000:00:08.0: [1414:5353] type 00 class 0x030000
[ 1.067930] pci 0000:00:08.0: reg 0x10: [mem 0xf8000000-0xfbffffff]
[ 1.077288] pci 0000:00:08.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff]
[ 1.098813] ACPI: PCI: Interrupt link LNKA configured for IRQ 11
[ 1.103481] ACPI: PCI: Interrupt link LNKB configured for IRQ 0
[ 1.111241] ACPI: PCI: Interrupt link LNKB disabled
[ 1.115250] ACPI: PCI: Interrupt link LNKC configured for IRQ 0
[ 1.119239] ACPI: PCI: Interrupt link LNKC disabled
[ 1.123319] ACPI: PCI: Interrupt link LNKD configured for IRQ 0
[ 1.127239] ACPI: PCI: Interrupt link LNKD disabled
[ 1.131379] iommu: Default domain type: Translated
[ 1.135397] SCSI subsystem initialized
[ 1.138665] libata version 3.00 loaded.
[ 1.138665] pci 0000:00:08.0: vgaarb: setting as boot VGA device
[ 1.139236] pci 0000:00:08.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none
[ 1.147253] pci 0000:00:08.0: vgaarb: bridge control possible
[ 1.151240] vgaarb: loaded
[ 1.153784] ACPI: bus type USB registered
[ 1.159256] usbcore: registered new interface driver usbfs
[ 1.163244] usbcore: registered new interface driver hub
[ 1.167248] usbcore: registered new device driver usb
[ 1.171254] pps_core: LinuxPPS API ver. 1 registered
[ 1.175239] pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti <giometti@linux.it>
[ 1.183241] PTP clock support registered
[ 1.187290] EDAC MC: Ver: 3.0.0
[ 1.191934] hv_vmbus: Vmbus version:4.0
[ 1.195290] hv_vmbus: Unknown GUID: c376c1c3-d276-48d2-90a9-c04748072c60
[ 1.199738] NetLabel: Initializing
[ 1.202361] NetLabel: domain hash size = 128
[ 1.203238] NetLabel: protocols = UNLABELED CIPSOv4 CALIPSO
[ 1.207253] NetLabel: unlabeled traffic allowed by default
[ 1.211274] PCI: Using ACPI for IRQ routing
[ 1.214597] PCI: pci_cache_line_size set to 64 bytes
[ 1.215103] e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff]
[ 1.215105] e820: reserve RAM buffer [mem 0x3fff0000-0x3fffffff]
[ 1.215327] clocksource: Switched to clocksource hyperv_clocksource_tsc_page
[ 1.233113] VFS: Disk quotas dquot_6.6.0
[ 1.236625] VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes)
[ 1.242163] AppArmor: AppArmor Filesystem Enabled
[ 1.246012] pnp: PnP ACPI init
[ 1.248792] pnp 00:00: Plug and Play ACPI device, IDs PNP0b00 (active)
[ 1.248840] pnp 00:01: Plug and Play ACPI device, IDs PNP0303 PNP030b (active)
[ 1.248881] pnp 00:02: Plug and Play ACPI device, IDs PNP0f03 PNP0f13 (active)
[ 1.249610] pnp 00:03: [dma 0 disabled]
[ 1.249634] pnp 00:03: Plug and Play ACPI device, IDs PNP0501 (active)
[ 1.250411] pnp 00:04: [dma 0 disabled]
[ 1.250432] pnp 00:04: Plug and Play ACPI device, IDs PNP0501 (active)
[ 1.251085] pnp 00:05: [dma 2]
[ 1.251131] pnp 00:05: Plug and Play ACPI device, IDs PNP0700 (active)
[ 1.251168] system 00:06: [io 0x01e0-0x01ef] has been reserved
[ 1.255876] system 00:06: [io 0x0160-0x016f] has been reserved
[ 1.261066] system 00:06: [io 0x0278-0x027f] has been reserved
[ 1.265610] system 00:06: [io 0x0378-0x037f] has been reserved
[ 1.270172] system 00:06: [io 0x0678-0x067f] has been reserved
[ 1.274668] system 00:06: [io 0x0778-0x077f] has been reserved
[ 1.279293] system 00:06: [io 0x04d0-0x04d1] has been reserved
[ 1.284494] system 00:06: Plug and Play ACPI device, IDs PNP0c02 (active)
[ 1.284635] system 00:07: [io 0x0400-0x043f] has been reserved
[ 1.289569] system 00:07: [io 0x0370-0x0371] has been reserved
[ 1.294219] system 00:07: [io 0x0440-0x044f] has been reserved
[ 1.299006] system 00:07: [mem 0xfec00000-0xfec00fff] could not be reserved
[ 1.305015] system 00:07: [mem 0xfee00000-0xfee00fff] has been reserved
[ 1.310557] system 00:07: Plug and Play ACPI device, IDs PNP0c02 (active)
[ 1.310651] system 00:08: [mem 0x00000000-0x0009ffff] could not be reserved
[ 1.315768] system 00:08: [mem 0x000c0000-0x000dffff] could not be reserved
[ 1.321686] system 00:08: [mem 0x000e0000-0x000fffff] could not be reserved
[ 1.328973] system 00:08: [mem 0x00100000-0x3fffffff] could not be reserved
[ 1.334464] system 00:08: [mem 0xfffc0000-0xffffffff] has been reserved
[ 1.339778] system 00:08: Plug and Play ACPI device, IDs PNP0c01 (active)
[ 1.340045] pnp: PnP ACPI: found 9 devices
[ 1.350659] clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns
[ 1.357522] NET: Registered protocol family 2
[ 1.361501] IP idents hash table entries: 131072 (order: 8, 1048576 bytes, linear)
[ 1.368607] tcp_listen_portaddr_hash hash table entries: 4096 (order: 4, 65536 bytes, linear)
[ 1.375122] TCP established hash table entries: 65536 (order: 7, 524288 bytes, linear)
[ 1.381290] TCP bind hash table entries: 65536 (order: 8, 1048576 bytes, linear)
[ 1.386786] TCP: Hash tables configured (established 65536 bind 65536)
[ 1.392093] MPTCP token hash table entries: 8192 (order: 5, 196608 bytes, linear)
[ 1.397618] UDP hash table entries: 4096 (order: 5, 131072 bytes, linear)
[ 1.402978] UDP-Lite hash table entries: 4096 (order: 5, 131072 bytes, linear)
[ 1.408458] NET: Registered protocol family 1
[ 1.411954] NET: Registered protocol family 44
[ 1.415639] pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window]
[ 1.420688] pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window]
[ 1.425489] pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window]
[ 1.430895] pci_bus 0000:00: resource 7 [mem 0x40000000-0xfffbffff window]
[ 1.435974] pci_bus 0000:00: resource 8 [mem 0xfe0000000-0xfffffffff window]
[ 1.441564] pci 0000:00:00.0: Limiting direct PCI/PCI transfers
[ 1.446435] PCI: CLS 0 bytes, default 64
[ 1.449756] PCI-DMA: Using software bounce buffering for IO (SWIOTLB)
[ 1.454686] software IO TLB: mapped [mem 0x000000003bff0000-0x000000003fff0000] (64MB)
[ 1.461244] Initialise system trusted keyrings
[ 1.467715] Key type blacklist registered
[ 1.472088] workingset: timestamp_bits=36 max_order=21 bucket_order=0
[ 1.479500] zbud: loaded
[ 1.483084] squashfs: version 4.0 (2009/01/31) Phillip Lougher
[ 1.489495] fuse: init (API version 7.34)
[ 1.493809] integrity: Platform Keyring initialized
[ 1.508299] Key type asymmetric registered
[ 1.511850] Asymmetric key parser 'x509' registered
[ 1.515651] Block layer SCSI generic (bsg) driver version 0.4 loaded (major 244)
[ 1.521342] io scheduler mq-deadline registered
[ 1.526057] shpchp: Standard Hot Plug PCI Controller Driver version: 0.4
[ 1.532006] hv_vmbus: registering driver hv_pci
[ 1.536365] input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input0
[ 1.543606] ACPI: button: Power Button [PWRF]
[ 1.548913] Serial: 8250/16550 driver, 32 ports, IRQ sharing enabled
[ 1.585747] 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A
[ 1.621384] 00:04: ttyS1 at I/O 0x2f8 (irq = 3, base_baud = 115200) is a 16550A
[ 1.629535] Linux agpgart interface v0.103
[ 1.759346] loop: module loaded
[ 1.763522] hv_vmbus: registering driver hv_storvsc
[ 1.768769] ata_piix 0000:00:07.1: version 2.13
[ 1.769339] ata_piix 0000:00:07.1: Hyper-V Virtual Machine detected, ATA device ignore set
[ 1.775353] scsi host0: storvsc_host_t
[ 1.777881] scsi host3: storvsc_host_t
[ 1.781835] scsi host4: ata_piix
[ 1.785505] scsi host2: storvsc_host_t
[ 1.788420] scsi 0:0:0:0: Direct-Access Msft Virtual Disk 1.0 PQ: 0 ANSI: 5
[ 1.792139] scsi host1: storvsc_host_t
[ 1.798871] scsi host5: ata_piix
[ 1.802323] scsi: waiting for bus probes to complete ...
[ 1.805352] ata1: PATA max UDMA/33 cmd 0x1f0 ctl 0x3f6 bmdma 0xffa0 irq 14
[ 1.815387] ata2: PATA max UDMA/33 cmd 0x170 ctl 0x376 bmdma 0xffa8 irq 15
[ 1.815587] sd 0:0:0:0: Attached scsi generic sg0 type 0
[ 1.820882] tun: Universal TUN/TAP device driver, 1.6
[ 1.830389] PPP generic driver version 2.4.2
[ 1.831493] sd 0:0:0:0: [sda] 180355072 512-byte logical blocks: (92.3 GB/86.0 GiB)
[ 1.834492] VFIO - User Level meta-driver version: 0.3
[ 1.840523] sd 0:0:0:0: [sda] 4096-byte physical blocks
[ 1.844928] i8042: PNP: PS/2 Controller [PNP0303:PS2K,PNP0f03:PS2M] at 0x60,0x64 irq 1,12
[ 1.849547] sd 0:0:0:0: [sda] Write Protect is off
[ 1.859768] sd 0:0:0:0: [sda] Mode Sense: 0f 00 10 00
[ 1.859930] sd 0:0:0:0: [sda] Write cache: disabled, read cache: enabled, supports DPO and FUA
[ 1.861908] serio: i8042 KBD port at 0x60,0x64 irq 1
[ 1.871116] serio: i8042 AUX port at 0x60,0x64 irq 12
[ 1.876766] scsi 1:0:1:0: Direct-Access Msft Virtual Disk 1.0 PQ: 0 ANSI: 5
[ 1.879269] mousedev: PS/2 mouse device common for all mice
[ 1.885180] sda: sda1 sda14 sda15
[ 1.889618] rtc_cmos 00:00: RTC can wake from S4
[ 1.897488] sd 1:0:1:0: Attached scsi generic sg1 type 0
[ 1.901049] rtc_cmos 00:00: registered as rtc0
[ 1.906971] rtc_cmos 00:00: setting system clock to 2022-06-20T00:09:30 UTC (1655683770)
[ 1.906981] sd 1:0:1:0: [sdb] 29360128 512-byte logical blocks: (15.0 GB/14.0 GiB)
[ 1.913258] rtc_cmos 00:00: alarms up to one month, 114 bytes nvram
[ 1.924677] device-mapper: uevent: version 1.0.3
[ 1.924702] sd 1:0:1:0: [sdb] Write Protect is off
[ 1.930269] input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input1
[ 1.933307] sd 1:0:1:0: [sdb] Mode Sense: 0f 00 10 00
[ 1.940560] device-mapper: ioctl: 4.45.0-ioctl (2021-03-22) initialised: dm-devel@redhat.com
[ 1.943526] sd 0:0:0:0: [sda] Attached SCSI disk
[ 1.947552] platform eisa.0: Probing EISA bus 0
[ 1.955489] platform eisa.0: EISA: Cannot allocate resource for mainboard
[ 1.955497] sd 1:0:1:0: [sdb] Write cache: disabled, read cache: enabled, supports DPO and FUA
[ 1.960667] platform eisa.0: Cannot allocate resource for EISA slot 1
[ 1.960669] platform eisa.0: Cannot allocate resource for EISA slot 2
[ 1.960669] platform eisa.0: Cannot allocate resource for EISA slot 3
[ 1.960670] platform eisa.0: Cannot allocate resource for EISA slot 4
[ 1.960671] platform eisa.0: Cannot allocate resource for EISA slot 5
[ 1.960672] platform eisa.0: Cannot allocate resource for EISA slot 6
[ 1.960673] platform eisa.0: Cannot allocate resource for EISA slot 7
[ 1.990987] ata1.01: host indicates ignore ATA devices, ignored
[ 1.993908] platform eisa.0: Cannot allocate resource for EISA slot 8
[ 1.993910] platform eisa.0: EISA: Detected 0 cards
[ 1.993913] intel_pstate: CPU model not supported
[ 1.993996] drop_monitor: Initializing network drop monitor service
[ 2.001369] ata1.00: host indicates ignore ATA devices, ignored
[ 2.005757] NET: Registered protocol family 10
[ 2.030051] Segment Routing with IPv6
[ 2.030169] sdb: sdb1
[ 2.034025] NET: Registered protocol family 17
[ 2.041143] Key type dns_resolver registered
[ 2.045072] No MBM correction factor available
[ 2.048937] IPI shorthand broadcast: enabled
[ 2.052429] sched_clock: Marking stable (1918415800, 133997700)->(2135540900, -83127400)
[ 2.059117] registered taskstats version 1
[ 2.062796] Loading compiled-in X.509 certificates
[ 2.067580] Loaded X.509 cert 'Build time autogenerated kernel key: 07f5640c3d7bf043074dc27d0b5799302e473486'
[ 2.075411] Loaded X.509 cert 'Canonical Ltd. Live Patch Signing: 14df34d1a87cf37625abec039ef2bf521249b969'
[ 2.083182] Loaded X.509 cert 'Canonical Ltd. Kernel Module Signing: 88f752e560a1e0737e31163a466ad7b70a850c19'
[ 2.090710] blacklist: Loading compiled-in revocation X.509 certificates
[ 2.096163] Loaded X.509 cert 'Canonical Ltd. Secure Boot Signing: 61482aa2830d0ab2ad5af10b7250da9033ddcef0'
[ 2.115323] zswap: loaded using pool lzo/zbud
[ 2.119178] Key type ._fscrypt registered
[ 2.122476] Key type .fscrypt registered
[ 2.127703] Key type fscrypt-provisioning registered
[ 2.131587] sd 1:0:1:0: [sdb] Attached SCSI disk
[ 2.135937] Key type encrypted registered
[ 2.139275] AppArmor: AppArmor sha1 policy hashing enabled
[ 2.143402] ima: No TPM chip found, activating TPM-bypass!
[ 2.147673] Loading compiled-in module X.509 certificates
[ 2.152851] Loaded X.509 cert 'Build time autogenerated kernel key: 07f5640c3d7bf043074dc27d0b5799302e473486'
[ 2.159620] ima: Allocated hash algorithm: sha1
[ 2.163299] ima: No architecture policies found
[ 2.166847] evm: Initialising EVM extended attributes:
[ 2.170616] evm: security.selinux
[ 2.173375] evm: security.SMACK64
[ 2.176151] evm: security.SMACK64EXEC
[ 2.179079] evm: security.SMACK64TRANSMUTE
[ 2.184144] evm: security.SMACK64MMAP
[ 2.188524] evm: security.apparmor
[ 2.192711] evm: security.ima
[ 2.195873] evm: security.capability
[ 2.199639] evm: HMAC attrs: 0x1
[ 2.204404] PM: Magic number: 10:116:153
[ 2.210374] RAS: Correctable Errors collector initialized.
[ 2.217450] md: Waiting for all devices to be available before autodetect
[ 2.224175] md: If you don't use raid, use raid=noautodetect
[ 2.230137] md: Autodetecting RAID arrays.
[ 2.234718] md: autorun ...
[ 2.237950] md: ... autorun DONE.
[ 2.245406] EXT4-fs (sda1): mounted filesystem with ordered data mode. Opts: (null). Quota mode: none.
[ 2.254982] VFS: Mounted root (ext4 filesystem) readonly on device 8:1.
[ 2.261710] devtmpfs: mounted
[ 2.266005] Freeing unused decrypted memory: 2036K
[ 2.271993] Freeing unused kernel image (initmem) memory: 2608K
[ 2.283320] Write protecting the kernel read-only data: 26624k
[ 2.290314] Freeing unused kernel image (text/rodata gap) memory: 2036K
[ 2.297019] Freeing unused kernel image (rodata/data gap) memory: 460K
[ 2.367601] x86/mm: Checked W+X mappings: passed, no W+X pages found.
[ 2.372927] x86/mm: Checking user space page tables
[ 2.434835] x86/mm: Checked W+X mappings: passed, no W+X pages found.
[ 2.440223] Run /sbin/init as init process
[ 2.443204] with arguments:
[ 2.443205] /sbin/init
[ 2.443206] with environment:
[ 2.443207] HOME=/
[ 2.443208] TERM=linux
[ 2.443209] BOOT_IMAGE=/boot/vmlinuz-5.13.0-1029-azure
[ 2.549627] systemd[1]: Inserted module 'autofs4'
[ 2.568658] systemd[1]: systemd 245.4-4ubuntu3.17 running in system mode. (+PAM +AUDIT +SELINUX +IMA +APPARMOR +SMACK +SYSVINIT +UTMP +LIBCRYPTSETUP +GCRYPT +GNUTLS +ACL +XZ +LZ4 +SECCOMP +BLKID +ELFUTILS +KMOD +IDN2 -IDN +PCRE2 default-hierarchy=hybrid)
[ 2.580375] systemd[1]: Detected virtualization microsoft.
[ 2.583537] systemd[1]: Detected architecture x86-64.
[ 2.604092] systemd[1]: Set hostname to <fv-az72-309>.
[ 2.777938] systemd[1]: /etc/systemd/system/runner-provisioner.service:3: Invalid URL, ignoring: None
[ 2.786328] systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network-online.target ignored (device units cannot be delayed).
[ 2.798000] systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network.target ignored (device units cannot be delayed).
[ 2.855931] systemd[1]: Unnecessary job for /sys/devices/virtual/misc/vmbus!hv_fcopy was removed.
[ 2.862804] systemd[1]: Unnecessary job for /sys/devices/virtual/misc/vmbus!hv_vss was removed.
[ 2.870773] systemd[1]: Created slice Slice for Azure VM Agent and Extensions.
[ 2.887880] systemd[1]: Created slice system-modprobe.slice.
[ 2.896196] systemd[1]: Created slice system-serial\x2dgetty.slice.
[ 2.905037] systemd[1]: Created slice system-systemd\x2dfsck.slice.
[ 2.914151] systemd[1]: Created slice User and Session Slice.
[ 2.922529] systemd[1]: Started Forward Password Requests to Wall Directory Watch.
[ 2.933020] systemd[1]: Set up automount Arbitrary Executable File Formats File System Automount Point.
[ 2.944377] systemd[1]: Reached target User and Group Name Lookups.
[ 2.954032] systemd[1]: Reached target Slices.
[ 2.960848] systemd[1]: Reached target Swap.
[ 2.966910] systemd[1]: Reached target System Time Set.
[ 2.974011] systemd[1]: Listening on Device-mapper event daemon FIFOs.
[ 2.984313] systemd[1]: Listening on LVM2 poll daemon socket.
[ 2.992189] systemd[1]: Listening on multipathd control socket.
[ 3.000530] systemd[1]: Listening on Syslog Socket.
[ 3.007510] systemd[1]: Listening on fsck to fsckd communication Socket.
[ 3.016525] systemd[1]: Listening on initctl Compatibility Named Pipe.
[ 3.026784] systemd[1]: Listening on Journal Audit Socket.
[ 3.034444] systemd[1]: Listening on Journal Socket (/dev/log).
[ 3.042507] systemd[1]: Listening on Journal Socket.
[ 3.049607] systemd[1]: Listening on Network Service Netlink Socket.
[ 3.058458] systemd[1]: Listening on udev Control Socket.
[ 3.068440] systemd[1]: Listening on udev Kernel Socket.
[ 3.076832] systemd[1]: Mounting Huge Pages File System...
[ 3.085236] systemd[1]: Mounting POSIX Message Queue File System...
[ 3.096003] systemd[1]: Mounting Kernel Debug File System...
[ 3.104508] systemd[1]: Mounting Kernel Trace File System...
[ 3.113953] systemd[1]: Starting Journal Service...
[ 3.122599] systemd[1]: Starting Set the console keyboard layout...
[ 3.132799] systemd[1]: Starting Create list of static device nodes for the current kernel...
[ 3.145710] systemd[1]: Starting Monitoring of LVM2 mirrors, snapshots etc. using dmeventd or progress polling...
[ 3.164138] systemd[1]: Starting Load Kernel Module drm...
[ 3.176384] systemd[1]: Condition check resulted in OpenVSwitch configuration for cleanup being skipped.
[ 3.185582] systemd[1]: Condition check resulted in Set Up Additional Binary Formats being skipped.
[ 3.193555] systemd[1]: Starting File System Check on Root Device...
[ 3.208194] systemd[1]: Starting Load Kernel Modules...
[ 3.220951] systemd[1]: Starting udev Coldplug all Devices...
[ 3.230545] systemd[1]: Starting Uncomplicated firewall...
[ 3.240173] systemd[1]: Starting Setup network rules for WALinuxAgent...
[ 3.251005] systemd[1]: Mounted Huge Pages File System.
[ 3.258335] systemd[1]: Mounted POSIX Message Queue File System.
[ 3.263139] IPMI message handler: version 39.2
[ 3.269778] systemd[1]: Mounted Kernel Debug File System.
[ 3.277611] systemd[1]: Started Journal Service.
[ 3.285221] ipmi device interface
[ 3.390859] EXT4-fs (sda1): re-mounted. Opts: discard. Quota mode: none.
[ 3.415983] systemd-journald[189]: Received client request to flush runtime journal.
[ 3.774545] hv_vmbus: registering driver hyperv_fb
[ 3.775538] hyperv_fb: Synthvid Version major 3, minor 5
[ 3.775586] hyperv_fb: Screen resolution: 1024x768, Color depth: 32, Frame buffer size: 8388608
[ 3.775590] hyperv_fb: Unable to allocate enough contiguous physical memory on Gen 1 VM. Using MMIO instead.
[ 3.778864] Console: switching to colour frame buffer device 128x48
[ 3.791655] hid: raw HID events driver (C) Jiri Kosina
[ 3.794288] hv_vmbus: registering driver hyperv_keyboard
[ 3.794936] input: AT Translated Set 2 keyboard as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0A03:00/device:07/VMBUS:01/d34b2567-b9b6-42b9-8778-0a4ec0b955bf/serio2/input/input3
[ 3.795652] hv_vmbus: registering driver hid_hyperv
[ 3.796114] input: Microsoft Vmbus HID-compliant Mouse as /devices/0006:045E:0621.0001/input/input4
[ 3.796162] hid 0006:045E:0621.0001: input: VIRTUAL HID v0.01 Mouse [Microsoft Vmbus HID-compliant Mouse] on
[ 3.800680] hv_vmbus: registering driver hv_balloon
[ 3.801323] hv_balloon: Using Dynamic Memory protocol version 2.0
[ 3.802112] hv_utils: Registering HyperV Utility Driver
[ 3.802114] hv_vmbus: registering driver hv_utils
[ 3.802498] hv_utils: Heartbeat IC version 3.0
[ 3.802894] hv_utils: TimeSync IC version 4.0
[ 3.803014] hv_utils: Shutdown IC version 3.2
[ 3.835525] hv_vmbus: registering driver hv_netvsc
[ 3.908860] cryptd: max_cpu_qlen set to 1000
[ 3.918334] AVX2 version of gcm_enc/dec engaged.
[ 3.918392] AES CTR mode by8 optimization enabled
[ 4.049638] bpfilter: Loaded bpfilter_umh pid 297
[ 4.051439] Started bpfilter
[ 4.130138] hv_utils: KVP IC version 4.0
[ 4.477911] alua: device handler registered
[ 4.478997] emc: device handler registered
[ 4.480520] rdac: device handler registered
[ 4.534218] loop0: detected capacity change from 0 to 126824
[ 4.547211] loop1: detected capacity change from 0 to 96160
[ 4.560510] loop2: detected capacity change from 0 to 138880
[ 4.663418] audit: type=1400 audit(1655683776.240:2): apparmor="STATUS" operation="profile_load" profile="unconfined" name="lsb_release" pid=438 comm="apparmor_parser"
[ 4.670890] audit: type=1400 audit(1655683776.248:3): apparmor="STATUS" operation="profile_load" profile="unconfined" name="/usr/lib/NetworkManager/nm-dhcp-client.action" pid=437 comm="apparmor_parser"
[ 4.670895] audit: type=1400 audit(1655683776.248:4): apparmor="STATUS" operation="profile_load" profile="unconfined" name="/usr/lib/NetworkManager/nm-dhcp-helper" pid=437 comm="apparmor_parser"
[ 4.670897] audit: type=1400 audit(1655683776.248:5): apparmor="STATUS" operation="profile_load" profile="unconfined" name="/usr/lib/connman/scripts/dhclient-script" pid=437 comm="apparmor_parser"
[ 4.670899] audit: type=1400 audit(1655683776.248:6): apparmor="STATUS" operation="profile_load" profile="unconfined" name="/{,usr/}sbin/dhclient" pid=437 comm="apparmor_parser"
[ 4.673295] audit: type=1400 audit(1655683776.252:7): apparmor="STATUS" operation="profile_load" profile="unconfined" name="/usr/sbin/haveged" pid=441 comm="apparmor_parser"
[ 4.681775] audit: type=1400 audit(1655683776.260:8): apparmor="STATUS" operation="profile_load" profile="unconfined" name="/usr/sbin/mysqld" pid=444 comm="apparmor_parser"
[ 4.687573] audit: type=1400 audit(1655683776.264:9): apparmor="STATUS" operation="profile_load" profile="unconfined" name="/usr/bin/man" pid=445 comm="apparmor_parser"
[ 4.687582] audit: type=1400 audit(1655683776.264:10): apparmor="STATUS" operation="profile_load" profile="unconfined" name="man_filter" pid=445 comm="apparmor_parser"
[ 5.528109] IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready
[ 8.388243] sdb: sdb1
[ 9.290071] EXT4-fs (sdb1): mounted filesystem with ordered data mode. Opts: (null). Quota mode: none.
[ 10.751284] Adding 4194300k swap on /mnt/swapfile. Priority:-2 extents:9 across:4505596k FS
[ 11.884750] aufs 5.x-rcN-20210809
[ 13.098160] kauditd_printk_skb: 22 callbacks suppressed
[ 13.098165] audit: type=1400 audit(1655683784.674:33): apparmor="STATUS" operation="profile_load" profile="unconfined" name="docker-default" pid=1019 comm="apparmor_parser"
[ 13.605076] bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this.
[ 13.610458] Bridge firewalling registered
[ 13.847391] Initializing XFRM netlink socket
[ 15.370396] loop3: detected capacity change from 0 to 8
[ 51.904594] hv_balloon: Max. dynamic memory size: 7168 MB
[ 122.843239] docker0: port 1(vethf1b0b28) entered blocking state
[ 122.843243] docker0: port 1(vethf1b0b28) entered disabled state
[ 122.843949] device vethf1b0b28 entered promiscuous mode
[ 122.853633] docker0: port 1(vethf1b0b28) entered blocking state
[ 122.853637] docker0: port 1(vethf1b0b28) entered forwarding state
[ 122.853683] docker0: port 1(vethf1b0b28) entered disabled state
[ 123.050679] eth0: renamed from veth63d8682
[ 123.070371] IPv6: ADDRCONF(NETDEV_CHANGE): vethf1b0b28: link becomes ready
[ 123.070415] docker0: port 1(vethf1b0b28) entered blocking state
[ 123.070417] docker0: port 1(vethf1b0b28) entered forwarding state
[ 123.070449] IPv6: ADDRCONF(NETDEV_CHANGE): docker0: link becomes ready
[ 123.145475] docker0: port 1(vethf1b0b28) entered disabled state
[ 123.146381] veth63d8682: renamed from eth0
[ 123.185593] docker0: port 1(vethf1b0b28) entered disabled state
[ 123.186189] device vethf1b0b28 left promiscuous mode
[ 123.186194] docker0: port 1(vethf1b0b28) entered disabled state
[ 123.313707] docker0: port 1(vethfb268c1) entered blocking state
[ 123.313712] docker0: port 1(vethfb268c1) entered disabled state
[ 123.313796] device vethfb268c1 entered promiscuous mode
[ 123.313978] docker0: port 1(vethfb268c1) entered blocking state
[ 123.314003] docker0: port 1(vethfb268c1) entered forwarding state
[ 123.314098] docker0: port 1(vethfb268c1) entered disabled state
[ 123.515861] eth0: renamed from vethecca8aa
[ 123.534376] IPv6: ADDRCONF(NETDEV_CHANGE): vethfb268c1: link becomes ready
[ 123.534420] docker0: port 1(vethfb268c1) entered blocking state
[ 123.534423] docker0: port 1(vethfb268c1) entered forwarding state
[ 127.321950] docker0: port 1(vethfb268c1) entered disabled state
[ 127.323417] vethecca8aa: renamed from eth0
[ 127.381466] docker0: port 1(vethfb268c1) entered disabled state
[ 127.382218] device vethfb268c1 left promiscuous mode
[ 127.382222] docker0: port 1(vethfb268c1) entered disabled state
[ 130.977631] br-d1e9d479f443: port 1(vethc0cce7e) entered blocking state
[ 130.977636] br-d1e9d479f443: port 1(vethc0cce7e) entered disabled state
[ 130.979736] device vethc0cce7e entered promiscuous mode
[ 131.314967] eth0: renamed from veth37c7cf5
[ 131.333674] IPv6: ADDRCONF(NETDEV_CHANGE): vethc0cce7e: link becomes ready
[ 131.333714] br-d1e9d479f443: port 1(vethc0cce7e) entered blocking state
[ 131.333718] br-d1e9d479f443: port 1(vethc0cce7e) entered forwarding state
[ 131.333751] IPv6: ADDRCONF(NETDEV_CHANGE): br-d1e9d479f443: link becomes ready
[ 131.942844] systemd-journald[180]: Received client request to flush runtime journal.
[ 162.921398] IPVS: Registered protocols (TCP, UDP, SCTP, AH, ESP)
[ 162.921414] IPVS: Connection hash table configured (size=4096, memory=64Kbytes)
[ 162.921504] IPVS: ipvs loaded.
[ 162.926542] IPVS: [rr] scheduler registered.
[ 162.930988] IPVS: [wrr] scheduler registered.
[ 162.934508] IPVS: [sh] scheduler registered.
[ 163.823643] docker0: port 1(vethf3c0380) entered blocking state
[ 163.823648] docker0: port 1(vethf3c0380) entered disabled state
[ 163.823689] device vethf3c0380 entered promiscuous mode
[ 164.071128] eth0: renamed from veth76d993e
[ 164.087408] docker0: port 1(vethf3c0380) entered blocking state
[ 164.087413] docker0: port 1(vethf3c0380) entered forwarding state
[ 173.054451] audit: type=1400 audit(1655683944.647:34): apparmor="STATUS" operation="profile_load" profile="unconfined" name="virt-aa-helper" pid=11337 comm="apparmor_parser"
[ 173.129330] audit: type=1400 audit(1655683944.719:35): apparmor="STATUS" operation="profile_load" profile="unconfined" name="libvirtd" pid=11343 comm="apparmor_parser"
[ 173.130841] audit: type=1400 audit(1655683944.723:36): apparmor="STATUS" operation="profile_load" profile="unconfined" name="libvirtd//qemu_bridge_helper" pid=11343 comm="apparmor_parser"
[ 175.918861] virbr0: port 1(virbr0-nic) entered blocking state
[ 175.918866] virbr0: port 1(virbr0-nic) entered disabled state
[ 175.918954] device virbr0-nic entered promiscuous mode
[ 176.287042] virbr0: port 1(virbr0-nic) entered blocking state
[ 176.287047] virbr0: port 1(virbr0-nic) entered listening state
[ 176.353283] virbr0: port 1(virbr0-nic) entered disabled state
[ 187.974699] audit: type=1400 audit(1655683959.568:37): apparmor="STATUS" operation="profile_replace" info="same as current profile, skipping" profile="unconfined" name="/usr/lib/NetworkManager/nm-dhcp-client.action" pid=12686 comm="apparmor_parser"
[ 187.974706] audit: type=1400 audit(1655683959.568:38): apparmor="STATUS" operation="profile_replace" info="same as current profile, skipping" profile="unconfined" name="/usr/lib/NetworkManager/nm-dhcp-helper" pid=12686 comm="apparmor_parser"
[ 187.974709] audit: type=1400 audit(1655683959.568:39): apparmor="STATUS" operation="profile_replace" info="same as current profile, skipping" profile="unconfined" name="/usr/lib/connman/scripts/dhclient-script" pid=12686 comm="apparmor_parser"
[ 187.974711] audit: type=1400 audit(1655683959.568:40): apparmor="STATUS" operation="profile_replace" info="same as current profile, skipping" profile="unconfined" name="/{,usr/}sbin/dhclient" pid=12686 comm="apparmor_parser"
[ 187.984624] audit: type=1400 audit(1655683959.576:41): apparmor="STATUS" operation="profile_replace" info="same as current profile, skipping" profile="unconfined" name="lsb_release" pid=12689 comm="apparmor_parser"
[ 188.003895] audit: type=1400 audit(1655683959.596:42): apparmor="STATUS" operation="profile_replace" info="same as current profile, skipping" profile="unconfined" name="/usr/sbin/haveged" pid=12698 comm="apparmor_parser"
[ 188.017221] audit: type=1400 audit(1655683959.612:43): apparmor="STATUS" operation="profile_replace" info="same as current profile, skipping" profile="unconfined" name="/usr/sbin/mysqld" pid=12701 comm="apparmor_parser"
[ 188.036963] audit: type=1400 audit(1655683959.632:44): apparmor="STATUS" operation="profile_replace" info="same as current profile, skipping" profile="unconfined" name="virt-aa-helper" pid=12706 comm="apparmor_parser"
[ 188.041828] audit: type=1400 audit(1655683959.636:45): apparmor="STATUS" operation="profile_replace" info="same as current profile, skipping" profile="unconfined" name="/usr/bin/man" pid=12713 comm="apparmor_parser"
[ 188.041832] audit: type=1400 audit(1655683959.636:46): apparmor="STATUS" operation="profile_replace" info="same as current profile, skipping" profile="unconfined" name="man_filter" pid=12713 comm="apparmor_parser"
[ 189.269764] docker0: port 2(vethf6689f9) entered blocking state
[ 189.269768] docker0: port 2(vethf6689f9) entered disabled state
[ 189.269823] device vethf6689f9 entered promiscuous mode
[ 189.388186] docker0: port 3(vethfeb4b6f) entered blocking state
[ 189.388191] docker0: port 3(vethfeb4b6f) entered disabled state
[ 189.388251] device vethfeb4b6f entered promiscuous mode
[ 189.388344] docker0: port 3(vethfeb4b6f) entered blocking state
[ 189.388346] docker0: port 3(vethfeb4b6f) entered forwarding state
[ 189.629134] eth0: renamed from vethc766298
[ 189.649308] docker0: port 3(vethfeb4b6f) entered disabled state
[ 189.649341] docker0: port 2(vethf6689f9) entered blocking state
[ 189.649343] docker0: port 2(vethf6689f9) entered forwarding state
[ 189.681035] eth0: renamed from vethe33eb6d
[ 189.697651] docker0: port 3(vethfeb4b6f) entered blocking state
[ 189.697656] docker0: port 3(vethfeb4b6f) entered forwarding state
[ 195.824288] docker0: port 4(vethc7ff58d) entered blocking state
[ 195.824292] docker0: port 4(vethc7ff58d) entered disabled state
[ 195.824365] device vethc7ff58d entered promiscuous mode
[ 196.104666] eth0: renamed from veth9d45b28
[ 196.124660] docker0: port 4(vethc7ff58d) entered blocking state
[ 196.124665] docker0: port 4(vethc7ff58d) entered forwarding state
[ 200.822457] docker0: port 5(vethdc7f949) entered blocking state
[ 200.822462] docker0: port 5(vethdc7f949) entered disabled state
[ 200.822743] device vethdc7f949 entered promiscuous mode
[ 202.012140] eth0: renamed from veth212df7a
[ 202.024114] docker0: port 5(vethdc7f949) entered blocking state
[ 202.024119] docker0: port 5(vethdc7f949) entered forwarding state
[ 215.628628] docker0: port 6(vethe779ac0) entered blocking state
[ 215.628633] docker0: port 6(vethe779ac0) entered disabled state
[ 215.628694] device vethe779ac0 entered promiscuous mode
[ 215.871345] eth0: renamed from veth24127ed
[ 215.890951] docker0: port 6(vethe779ac0) entered blocking state
[ 215.890955] docker0: port 6(vethe779ac0) entered forwarding state
[ 222.610724] wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information.
[ 222.610728] wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld <Jason@zx2c4.com>. All Rights Reserved.
[ 223.025912] docker0: port 6(vethe779ac0) entered disabled state
[ 223.025979] veth24127ed: renamed from eth0
[ 223.064463] docker0: port 6(vethe779ac0) entered disabled state
[ 223.064999] device vethe779ac0 left promiscuous mode
[ 223.065003] docker0: port 6(vethe779ac0) entered disabled state
[ 224.763707] ipip: IPv4 and MPLS over IPv4 tunneling driver
[ 227.448013] docker0: port 6(vethb75fa03) entered blocking state
[ 227.448019] docker0: port 6(vethb75fa03) entered disabled state
[ 227.448077] device vethb75fa03 entered promiscuous mode
[ 227.878045] eth0: renamed from vethc1a763e
[ 227.894005] docker0: port 6(vethb75fa03) entered blocking state
[ 227.894010] docker0: port 6(vethb75fa03) entered forwarding state
[ 230.428500] docker0: port 7(veth360360f) entered blocking state
[ 230.428505] docker0: port 7(veth360360f) entered disabled state
[ 230.428548] device veth360360f entered promiscuous mode
[ 231.230069] eth0: renamed from veth002f52a
[ 231.249762] docker0: port 7(veth360360f) entered blocking state
[ 231.249776] docker0: port 7(veth360360f) entered forwarding state
[ 239.109008] docker0: port 8(veth551f867) entered blocking state
[ 239.109013] docker0: port 8(veth551f867) entered disabled state
[ 239.109069] device veth551f867 entered promiscuous mode
[ 239.139110] docker0: port 9(vethcd694ff) entered blocking state
[ 239.139114] docker0: port 9(vethcd694ff) entered disabled state
[ 239.139162] device vethcd694ff entered promiscuous mode
[ 239.139266] docker0: port 9(vethcd694ff) entered blocking state
[ 239.139268] docker0: port 9(vethcd694ff) entered forwarding state
[ 240.108830] docker0: port 9(vethcd694ff) entered disabled state
[ 240.701186] eth0: renamed from veth573781b
[ 240.726081] docker0: port 9(vethcd694ff) entered blocking state
[ 240.726089] docker0: port 9(vethcd694ff) entered forwarding state
[ 241.056848] eth0: renamed from veth9960fe9
[ 241.073199] docker0: port 8(veth551f867) entered blocking state
[ 241.073205] docker0: port 8(veth551f867) entered forwarding state
[ 250.702648] docker0: port 10(veth395917c) entered blocking state
[ 250.702658] docker0: port 10(veth395917c) entered disabled state
[ 250.702830] device veth395917c entered promiscuous mode
[ 251.432454] eth0: renamed from vethf3a465a
[ 251.440359] docker0: port 10(veth395917c) entered blocking state
[ 251.440364] docker0: port 10(veth395917c) entered forwarding state
[ 275.850895] docker0: port 11(veth8ce295f) entered blocking state
[ 275.850900] docker0: port 11(veth8ce295f) entered disabled state
[ 275.850951] device veth8ce295f entered promiscuous mode
[ 276.090310] eth0: renamed from veth0aa0305
[ 276.106200] docker0: port 11(veth8ce295f) entered blocking state
[ 276.106206] docker0: port 11(veth8ce295f) entered forwarding state
[ 280.206137] docker0: port 11(veth8ce295f) entered disabled state
[ 280.206263] veth0aa0305: renamed from eth0
[ 280.240524] docker0: port 11(veth8ce295f) entered disabled state
[ 280.241174] device veth8ce295f left promiscuous mode
[ 280.241178] docker0: port 11(veth8ce295f) entered disabled state
[ 282.364781] docker0: port 11(veth6e9387e) entered blocking state
[ 282.364786] docker0: port 11(veth6e9387e) entered disabled state
[ 282.364901] device veth6e9387e entered promiscuous mode
[ 282.605794] eth0: renamed from veth52d3a39
[ 282.621692] docker0: port 11(veth6e9387e) entered blocking state
[ 282.621697] docker0: port 11(veth6e9387e) entered forwarding state
[ 283.136309] docker0: port 11(veth6e9387e) entered disabled state
[ 283.136874] veth52d3a39: renamed from eth0
[ 283.167483] docker0: port 11(veth6e9387e) entered disabled state
[ 283.168004] device veth6e9387e left promiscuous mode
[ 283.168008] docker0: port 11(veth6e9387e) entered disabled state
[ 283.969138] docker0: port 11(veth2163595) entered blocking state
[ 283.969142] docker0: port 11(veth2163595) entered disabled state
[ 283.969216] device veth2163595 entered promiscuous mode
[ 283.969924] docker0: port 11(veth2163595) entered blocking state
[ 283.969927] docker0: port 11(veth2163595) entered forwarding state
[ 284.077179] docker0: port 12(vethdbba819) entered blocking state
[ 284.077184] docker0: port 12(vethdbba819) entered disabled state
[ 284.077505] device vethdbba819 entered promiscuous mode
[ 284.078146] docker0: port 12(vethdbba819) entered blocking state
[ 284.078150] docker0: port 12(vethdbba819) entered forwarding state
[ 284.101116] docker0: port 13(veth90d3104) entered blocking state
[ 284.101120] docker0: port 13(veth90d3104) entered disabled state
[ 284.101644] device veth90d3104 entered promiscuous mode
[ 284.102245] docker0: port 13(veth90d3104) entered blocking state
[ 284.102248] docker0: port 13(veth90d3104) entered forwarding state
[ 284.397307] docker0: port 11(veth2163595) entered disabled state
[ 284.397405] docker0: port 12(vethdbba819) entered disabled state
[ 284.397450] docker0: port 13(veth90d3104) entered disabled state
[ 284.414239] eth0: renamed from vethac29a90
[ 284.442930] docker0: port 11(veth2163595) entered blocking state
[ 284.442934] docker0: port 11(veth2163595) entered forwarding state
[ 284.697655] eth0: renamed from vethe9eb585
[ 284.717491] docker0: port 12(vethdbba819) entered blocking state
[ 284.717496] docker0: port 12(vethdbba819) entered forwarding state
[ 284.734121] eth0: renamed from veth0690f1b
[ 284.750154] docker0: port 13(veth90d3104) entered blocking state
[ 284.750158] docker0: port 13(veth90d3104) entered forwarding state
[ 346.412689] docker0: port 14(veth31f2917) entered blocking state
[ 346.412694] docker0: port 14(veth31f2917) entered disabled state
[ 346.413054] device veth31f2917 entered promiscuous mode
[ 346.644782] eth0: renamed from veth9a7409d
[ 346.656799] docker0: port 14(veth31f2917) entered blocking state
[ 346.656806] docker0: port 14(veth31f2917) entered forwarding state
[ 540.604561] docker0: port 15(veth4420858) entered blocking state
[ 540.604565] docker0: port 15(veth4420858) entered disabled state
[ 540.604748] device veth4420858 entered promiscuous mode
[ 541.037465] eth0: renamed from veth0c65c92
[ 541.057314] docker0: port 15(veth4420858) entered blocking state
[ 541.057320] docker0: port 15(veth4420858) entered forwarding state
[ 555.856654] docker0: port 15(veth4420858) entered disabled state
[ 555.857255] eth0-nic: renamed from eth0
[ 555.907532] k6t-eth0: port 1(eth0-nic) entered blocking state
[ 555.907536] k6t-eth0: port 1(eth0-nic) entered disabled state
[ 555.907600] device eth0-nic entered promiscuous mode
[ 556.001164] k6t-eth0: port 2(tap0) entered blocking state
[ 556.001170] k6t-eth0: port 2(tap0) entered disabled state
[ 556.001252] device tap0 entered promiscuous mode
[ 556.001348] k6t-eth0: port 2(tap0) entered blocking state
[ 556.001351] k6t-eth0: port 2(tap0) entered forwarding state
[ 556.001527] k6t-eth0: port 1(eth0-nic) entered blocking state
[ 556.001530] k6t-eth0: port 1(eth0-nic) entered forwarding state
[ 556.001557] docker0: port 15(veth4420858) entered blocking state
[ 556.001559] docker0: port 15(veth4420858) entered forwarding state
[ 886.184925] k6t-eth0: port 2(tap0) entered disabled state
[ 887.911689] veth0c65c92: renamed from eth0
[ 888.021265] device tap0 left promiscuous mode
[ 888.021291] k6t-eth0: port 2(tap0) entered disabled state
[ 888.029352] device eth0-nic left promiscuous mode
[ 888.029367] k6t-eth0: port 1(eth0-nic) entered disabled state
[ 888.045574] docker0: port 15(veth4420858) entered disabled state
[ 888.077529] device veth4420858 left promiscuous mode
[ 888.077535] docker0: port 15(veth4420858) entered disabled state
[ 916.413810] docker0: port 15(vetha966948) entered blocking state
[ 916.413815] docker0: port 15(vetha966948) entered disabled state
[ 916.413873] device vetha966948 entered promiscuous mode
[ 916.715286] eth0: renamed from vethdaacc5d
[ 916.731275] docker0: port 15(vetha966948) entered blocking state
[ 916.731280] docker0: port 15(vetha966948) entered forwarding state
[ 927.323294] docker0: port 15(vetha966948) entered disabled state
[ 927.338227] vethdaacc5d: renamed from eth0
[ 927.372610] docker0: port 15(vetha966948) entered disabled state
[ 927.373203] device vetha966948 left promiscuous mode
[ 927.373207] docker0: port 15(vetha966948) entered disabled state
[ 937.520625] veth9a7409d: renamed from eth0
[ 937.533585] docker0: port 14(veth31f2917) entered disabled state
[ 937.554770] docker0: port 14(veth31f2917) entered disabled state
[ 937.555761] device veth31f2917 left promiscuous mode
[ 937.555766] docker0: port 14(veth31f2917) entered disabled state
This file has been truncated, but you can view the full file.
-- Logs begin at Thu 2022-06-16 07:26:56 UTC, end at Mon 2022-06-20 00:25:18 UTC. --
Jun 16 07:26:56 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:56 fv-az72-309 provisioner[1638]: Factory script /opt/post-generation/cleanup-logs.sh has finished
Jun 16 07:26:56 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:56 fv-az72-309 provisioner[1638]: Invoke factory script /opt/post-generation/environment-variables.sh
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: Factory script /opt/post-generation/environment-variables.sh has finished
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: Vacuuming done, freed 0B of archived journals from /var/log/journal.
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: Vacuuming done, freed 0B of archived journals from /run/log/journal.
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: Deleted archived journal /var/log/journal/3d8d945fc71147a483fa20cb6792de9d/system@00000000000000000000000000000000-000000000000256a-0005e16fa65af853.journal (8.0M).
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: Deleted archived journal /var/log/journal/3d8d945fc71147a483fa20cb6792de9d/user-1000@00000000000000000000000000000000-000000000000257a-0005e16fa6698e8b.journal (8.0M).
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: Deleted archived journal /var/log/journal/3d8d945fc71147a483fa20cb6792de9d/system@00000000000000000000000000000000-0000000000002a24-0005e18b85a906d3.journal (8.0M).
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: Vacuuming done, freed 24.0M of archived journals from /var/log/journal/3d8d945fc71147a483fa20cb6792de9d.
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: Downloading abuse tools to /opt/runner/provisioner/etc
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: Archive: /opt/runner/provisioner/etc/tools.zip
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: inflating: /opt/runner/provisioner/etc/jobkeepalive
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: inflating: /opt/runner/provisioner/etc/provjobd
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: Finished installing abuse tools
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: --2022-06-16 07:26:55-- https://abusetoolscdn.blob.core.windows.net/binaries/v0.40-linux
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: Resolving abusetoolscdn.blob.core.windows.net (abusetoolscdn.blob.core.windows.net)... 52.239.170.164, 20.150.90.4, 52.239.171.4
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: Connecting to abusetoolscdn.blob.core.windows.net (abusetoolscdn.blob.core.windows.net)|52.239.170.164|:443... connected.
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: HTTP request sent, awaiting response... 200 OK
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: Length: 3104000 (3.0M) [application/octet-stream]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: Saving to: ‘/opt/runner/provisioner/etc/tools.zip’
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 0K .......... .......... .......... .......... .......... 1% 37.4M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 50K .......... .......... .......... .......... .......... 3% 73.1M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 100K .......... .......... .......... .......... .......... 4% 33.4M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 150K .......... .......... .......... .......... .......... 6% 55.6M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 200K .......... .......... .......... .......... .......... 8% 264M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 250K .......... .......... .......... .......... .......... 9% 51.8M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 300K .......... .......... .......... .......... .......... 11% 265M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 350K .......... .......... .......... .......... .......... 13% 45.4M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 400K .......... .......... .......... .......... .......... 14% 264M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 450K .......... .......... .......... .......... .......... 16% 47.0M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 500K .......... .......... .......... .......... .......... 18% 256M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 550K .......... .......... .......... .......... .......... 19% 235M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 600K .......... .......... .......... .......... .......... 21% 74.3M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 650K .......... .......... .......... .......... .......... 23% 258M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 700K .......... .......... .......... .......... .......... 24% 266M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 750K .......... .......... .......... .......... .......... 26% 219M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 800K .......... .......... .......... .......... .......... 28% 93.3M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 850K .......... .......... .......... .......... .......... 29% 233M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 900K .......... .......... .......... .......... .......... 31% 106M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 950K .......... .......... .......... .......... .......... 32% 232M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 1000K .......... .......... .......... .......... .......... 34% 250M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 1050K .......... .......... .......... .......... .......... 36% 233M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 1100K .......... .......... .......... .......... .......... 37% 268M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 1150K .......... .......... .......... .......... .......... 39% 218M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 1200K .......... .......... .......... .......... .......... 41% 53.2M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 1250K .......... .......... .......... .......... .......... 42% 118M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 1300K .......... .......... .......... .......... .......... 44% 122M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 1350K .......... .......... .......... .......... .......... 46% 240M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 1400K .......... .......... .......... .......... .......... 47% 273M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 1450K .......... .......... .......... .......... .......... 49% 135M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 1500K .......... .......... .......... .......... .......... 51% 253M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 1550K .......... .......... .......... .......... .......... 52% 227M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 1600K .......... .......... .......... .......... .......... 54% 154M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 1650K .......... .......... .......... .......... .......... 56% 252M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 1700K .......... .......... .......... .......... .......... 57% 246M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 1750K .......... .......... .......... .......... .......... 59% 245M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 1800K .......... .......... .......... .......... .......... 61% 140M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 1850K .......... .......... .......... .......... .......... 62% 274M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 1900K .......... .......... .......... .......... .......... 64% 259M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 1950K .......... .......... .......... .......... .......... 65% 131M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 2000K .......... .......... .......... .......... .......... 67% 276M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 2050K .......... .......... .......... .......... .......... 69% 264M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 2100K .......... .......... .......... .......... .......... 70% 156M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 2150K .......... .......... .......... .......... .......... 72% 243M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 2200K .......... .......... .......... .......... .......... 74% 248M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 2250K .......... .......... .......... .......... .......... 75% 133M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 2300K .......... .......... .......... .......... .......... 77% 222M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 2350K .......... .......... .......... .......... .......... 79% 226M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 2400K .......... .......... .......... .......... .......... 80% 163M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 2450K .......... .......... .......... .......... .......... 82% 256M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 2500K .......... .......... .......... .......... .......... 84% 234M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 2550K .......... .......... .......... .......... .......... 85% 235M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 2600K .......... .......... .......... .......... .......... 87% 164M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 2650K .......... .......... .......... .......... .......... 89% 277M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 2700K .......... .......... .......... .......... .......... 90% 270M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 2750K .......... .......... .......... .......... .......... 92% 144M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 2800K .......... .......... .......... .......... .......... 94% 209M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 2850K .......... .......... .......... .......... .......... 95% 276M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 2900K .......... .......... .......... .......... .......... 97% 283M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 2950K .......... .......... .......... .......... .......... 98% 249M 0s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 3000K .......... .......... .......... . 100% 313M=0.02s
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: 2022-06-16 07:26:55 (139 MB/s) - ‘/opt/runner/provisioner/etc/tools.zip’ saved [3104000/3104000]
Jun 16 07:26:57 fv-az72-309 provisioner[1638]: [07:26:56] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:58 fv-az72-309 provisioner[1638]: [07:26:58] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:58 fv-az72-309 provisioner[1638]: INSTALL_OS_TOOL empty; skipping...
Jun 16 07:26:58 fv-az72-309 provisioner[1638]: [07:26:58] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:58 fv-az72-309 provisioner[1638]: RUNNER_TOOL_CACHE=/opt/hostedtoolcache
Jun 16 07:26:58 fv-az72-309 provisioner[1638]: [07:26:58] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:58 fv-az72-309 provisioner[1638]: RUNNER_TOOL_CACHE set to match AGENT_TOOLSDIRECTORY: /opt/hostedtoolcache
Jun 16 07:26:59 fv-az72-309 provisioner[1638]: [07:26:59] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:59 fv-az72-309 provisioner[1638]: + ImageName=Ubuntu20
Jun 16 07:26:59 fv-az72-309 provisioner[1638]: [07:26:59] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:59 fv-az72-309 provisioner[1638]: + [[ Ubuntu20 = *Ubuntu* ]]
Jun 16 07:26:59 fv-az72-309 provisioner[1638]: [07:26:59] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:59 fv-az72-309 provisioner[1638]: ++ grep ResourceDisk.EnableSwap=y
Jun 16 07:26:59 fv-az72-309 provisioner[1638]: [07:26:59] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:59 fv-az72-309 provisioner[1638]: ++ cat /etc/waagent.conf
Jun 16 07:26:59 fv-az72-309 provisioner[1638]: [07:26:59] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:59 fv-az72-309 provisioner[1638]: + isMntSwap=ResourceDisk.EnableSwap=y
Jun 16 07:26:59 fv-az72-309 provisioner[1638]: [07:26:59] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:59 fv-az72-309 provisioner[1638]: + '[' -z ResourceDisk.EnableSwap=y ']'
Jun 16 07:26:59 fv-az72-309 provisioner[1638]: [07:26:59] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:26:59 fv-az72-309 provisioner[1638]: + apt-get install -y libcgroup1 cgroup-tools
Jun 16 07:27:05 fv-az72-309 provisioner[1638]: [07:27:05] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:05 fv-az72-309 provisioner[1638]: WARNING! Using --password via the CLI is insecure. Use --password-stdin.
Jun 16 07:27:05 fv-az72-309 provisioner[1638]: [07:27:05] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:05 fv-az72-309 provisioner[1638]: WARNING! Your password will be stored unencrypted in /home/runner/.docker/config.json.
Jun 16 07:27:05 fv-az72-309 provisioner[1638]: [07:27:05] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:05 fv-az72-309 provisioner[1638]: Configure a credential helper to remove this warning. See
Jun 16 07:27:05 fv-az72-309 provisioner[1638]: [07:27:05] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:05 fv-az72-309 provisioner[1638]: https://docs.docker.com/engine/reference/commandline/login/#credentials-store
Jun 16 07:27:05 fv-az72-309 provisioner[1638]: [07:27:05] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:05 fv-az72-309 provisioner[1638]: [07:27:05] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:05 fv-az72-309 provisioner[1638]: Login Succeeded
Jun 16 07:27:05 fv-az72-309 sudo[1681]: pam_unix(sudo:session): session closed for user runner
Jun 16 07:27:07 fv-az72-309 provisioner[1638]: [07:27:07] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:07 fv-az72-309 provisioner[1638]: Reading package lists...
Jun 16 07:27:07 fv-az72-309 provisioner[1638]: [07:27:07] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:07 fv-az72-309 provisioner[1638]: Building dependency tree...
Jun 16 07:27:07 fv-az72-309 provisioner[1638]: [07:27:07] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:07 fv-az72-309 provisioner[1638]: Reading state information...
Jun 16 07:27:08 fv-az72-309 provisioner[1638]: [07:27:08] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:08 fv-az72-309 provisioner[1638]: The following NEW packages will be installed:
Jun 16 07:27:08 fv-az72-309 provisioner[1638]: [07:27:08] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:08 fv-az72-309 provisioner[1638]: cgroup-tools libcgroup1
Jun 16 07:27:08 fv-az72-309 provisioner[1638]: [07:27:08] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:08 fv-az72-309 provisioner[1638]: 0 upgraded, 2 newly installed, 0 to remove and 13 not upgraded.
Jun 16 07:27:08 fv-az72-309 provisioner[1638]: [07:27:08] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:08 fv-az72-309 provisioner[1638]: Need to get 109 kB of archives.
Jun 16 07:27:08 fv-az72-309 provisioner[1638]: [07:27:08] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:08 fv-az72-309 provisioner[1638]: After this operation, 472 kB of additional disk space will be used.
Jun 16 07:27:08 fv-az72-309 provisioner[1638]: [07:27:08] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:08 fv-az72-309 provisioner[1638]: Get:1 http://azure.archive.ubuntu.com/ubuntu focal/universe amd64 libcgroup1 amd64 0.41-10 [42.9 kB]
Jun 16 07:27:08 fv-az72-309 provisioner[1638]: [07:27:08] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:08 fv-az72-309 provisioner[1638]: Get:2 http://azure.archive.ubuntu.com/ubuntu focal/universe amd64 cgroup-tools amd64 0.41-10 [66.2 kB]
Jun 16 07:27:08 fv-az72-309 provisioner[1638]: [07:27:08] info: Provisioner.Framework.JobRunner[0]
Jun 16 07:27:08 fv-az72-309 provisioner[1638]: Starting job Machine Info Monitor
Jun 16 07:27:08 fv-az72-309 provisioner[1638]: [07:27:08] info: Microsoft.AzureDevOps.Provisioner.Framework.Monitoring.MachineInfoMonitorJob[7000]
Jun 16 07:27:08 fv-az72-309 provisioner[1638]: {"Processor":"0","VendorId":"GenuineIntel","CpuFamily":"6","Model":"63","ModelName":"Intel(R) Xeon(R) CPU E5-2673 v3 @ 2.40GHz","Stepping":"2","CpuMHz":"2397.223","CacheSize":"30720 KB","CpuCores":"2"}
Jun 16 07:27:08 fv-az72-309 provisioner[1638]: [07:27:08] info: Provisioner.Framework.JobRunner[0]
Jun 16 07:27:08 fv-az72-309 provisioner[1638]: Finished job Machine Info Monitor
Jun 16 07:27:08 fv-az72-309 provisioner[1638]: [07:27:08] info: Provisioner.Framework.JobRunner[0]
Jun 16 07:27:08 fv-az72-309 provisioner[1638]: Starting job Machine Health Monitor
Jun 16 07:27:08 fv-az72-309 provisioner[1638]: [07:27:08] fail: Microsoft.AzureDevOps.Provisioner.Framework.Monitoring.MachineHealthMonitorJob[1007]
Jun 16 07:27:08 fv-az72-309 provisioner[1638]: Exception during the attempt to publish machine metrics to Mms.
Jun 16 07:27:08 fv-az72-309 provisioner[1638]: System.ArgumentException: The collection must contain at least one element. (Parameter 'bytes')
Jun 16 07:27:08 fv-az72-309 provisioner[1638]: at GitHub.Services.Common.ArgumentUtility.CheckEnumerableForNullOrEmpty(IEnumerable enumerable, String enumerableName, String expectedServiceArea) in /home/vsts/work/1/s/mms.client/Common/Utility/ArgumentUtility.cs:line 194
Jun 16 07:27:08 fv-az72-309 provisioner[1638]: at GitHub.Services.Common.ArgumentUtility.CheckEnumerableForNullOrEmpty(IEnumerable enumerable, String enumerableName) in /home/vsts/work/1/s/mms.client/Common/Utility/ArgumentUtility.cs:line 178
Jun 16 07:27:08 fv-az72-309 provisioner[1638]: at GitHub.Services.Common.PrimitiveExtensions.ToBase64StringNoPadding(Byte[] bytes) in /home/vsts/work/1/s/mms.client/Common/Utility/PrimitiveExtensions.cs:line 53
Jun 16 07:27:08 fv-az72-309 provisioner[1638]: at MachineManagement.Provisioning.MachineManagementClient.PublishMetricsAsync(String poolName, String instanceName, MachineMetric[] metrics, Byte[] postRegistrationAccessToken, CancellationToken cancellationToken) in /home/vsts/work/1/s/provisioner.framework/MachineManagementClient.cs:line 46
Jun 16 07:27:08 fv-az72-309 provisioner[1638]: at Microsoft.AzureDevOps.Provisioner.Framework.Monitoring.MachineHealthMonitorJob.PublishSampleMetrics(CancellationToken cancellationToken) in /home/vsts/work/1/s/provisioner.framework/Monitoring/Jobs/MachineHealthMonitorJob.cs:line 75
Jun 16 07:27:08 fv-az72-309 provisioner[1638]: [07:27:08] info: Microsoft.AzureDevOps.Provisioner.Framework.Monitoring.MachineHealthMonitorJob[0]
Jun 16 07:27:08 fv-az72-309 provisioner[1638]: Machine is healthy.
Jun 16 07:27:08 fv-az72-309 provisioner[1638]: [07:27:08] info: Provisioner.Framework.JobRunner[0]
Jun 16 07:27:08 fv-az72-309 provisioner[1638]: Finished job Machine Health Monitor
Jun 16 07:27:08 fv-az72-309 provisioner[1638]: [07:27:08] info: Provisioner.Framework.JobRunner[0]
Jun 16 07:27:08 fv-az72-309 provisioner[1638]: Job Machine Health Monitor is scheduled to run in 300 seconds
Jun 16 07:27:08 fv-az72-309 provisioner[1638]: [07:27:08] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:08 fv-az72-309 provisioner[1638]: Fetched 109 kB in 0s (1299 kB/s)
Jun 16 07:27:09 fv-az72-309 provisioner[1638]: [07:27:09] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:09 fv-az72-309 provisioner[1638]: Selecting previously unselected package libcgroup1:amd64.
Jun 16 07:27:09 fv-az72-309 provisioner[1638]: [07:27:09] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:09 fv-az72-309 provisioner[1638]: (Reading database ...
Jun 16 07:27:09 fv-az72-309 provisioner[1638]: [07:27:09] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:09 fv-az72-309 provisioner[1638]: (Reading database ... 5%
Jun 16 07:27:09 fv-az72-309 provisioner[1638]: [07:27:09] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:09 fv-az72-309 provisioner[1638]: (Reading database ... 10%
Jun 16 07:27:09 fv-az72-309 provisioner[1638]: [07:27:09] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:09 fv-az72-309 provisioner[1638]: (Reading database ... 15%
Jun 16 07:27:09 fv-az72-309 provisioner[1638]: [07:27:09] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:09 fv-az72-309 provisioner[1638]: (Reading database ... 20%
Jun 16 07:27:09 fv-az72-309 provisioner[1638]: [07:27:09] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:09 fv-az72-309 provisioner[1638]: (Reading database ... 25%
Jun 16 07:27:09 fv-az72-309 provisioner[1638]: [07:27:09] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:09 fv-az72-309 provisioner[1638]: (Reading database ... 30%
Jun 16 07:27:09 fv-az72-309 provisioner[1638]: [07:27:09] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:09 fv-az72-309 provisioner[1638]: (Reading database ... 35%
Jun 16 07:27:09 fv-az72-309 provisioner[1638]: [07:27:09] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:09 fv-az72-309 provisioner[1638]: (Reading database ... 40%
Jun 16 07:27:09 fv-az72-309 provisioner[1638]: [07:27:09] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:09 fv-az72-309 provisioner[1638]: (Reading database ... 45%
Jun 16 07:27:09 fv-az72-309 provisioner[1638]: [07:27:09] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:09 fv-az72-309 provisioner[1638]: (Reading database ... 50%
Jun 16 07:27:09 fv-az72-309 provisioner[1638]: [07:27:09] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:09 fv-az72-309 provisioner[1638]: (Reading database ... 55%
Jun 16 07:27:09 fv-az72-309 provisioner[1638]: [07:27:09] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:09 fv-az72-309 provisioner[1638]: (Reading database ... 60%
Jun 16 07:27:10 fv-az72-309 provisioner[1638]: [07:27:10] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:10 fv-az72-309 provisioner[1638]: (Reading database ... 65%
Jun 16 07:27:10 fv-az72-309 provisioner[1638]: [07:27:10] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:10 fv-az72-309 provisioner[1638]: (Reading database ... 70%
Jun 16 07:27:11 fv-az72-309 provisioner[1638]: [07:27:11] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:11 fv-az72-309 provisioner[1638]: (Reading database ... 75%
Jun 16 07:27:11 fv-az72-309 provisioner[1638]: [07:27:11] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:11 fv-az72-309 provisioner[1638]: (Reading database ... 80%
Jun 16 07:27:12 fv-az72-309 provisioner[1638]: [07:27:12] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:12 fv-az72-309 provisioner[1638]: (Reading database ... 85%
Jun 16 07:27:12 fv-az72-309 provisioner[1638]: [07:27:12] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:12 fv-az72-309 provisioner[1638]: (Reading database ... 90%
Jun 16 07:27:13 fv-az72-309 provisioner[1638]: [07:27:13] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:13 fv-az72-309 provisioner[1638]: (Reading database ... 95%
Jun 16 07:27:13 fv-az72-309 provisioner[1638]: [07:27:13] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:13 fv-az72-309 provisioner[1638]: (Reading database ... 100%
Jun 16 07:27:13 fv-az72-309 provisioner[1638]: [07:27:13] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:13 fv-az72-309 provisioner[1638]: (Reading database ... 231549 files and directories currently installed.)
Jun 16 07:27:13 fv-az72-309 provisioner[1638]: [07:27:13] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:13 fv-az72-309 provisioner[1638]: Preparing to unpack .../libcgroup1_0.41-10_amd64.deb ...
Jun 16 07:27:13 fv-az72-309 provisioner[1638]: [07:27:13] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:13 fv-az72-309 provisioner[1638]: Unpacking libcgroup1:amd64 (0.41-10) ...
Jun 16 07:27:13 fv-az72-309 provisioner[1638]: [07:27:13] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:13 fv-az72-309 provisioner[1638]: Selecting previously unselected package cgroup-tools.
Jun 16 07:27:13 fv-az72-309 provisioner[1638]: [07:27:13] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:13 fv-az72-309 provisioner[1638]: Preparing to unpack .../cgroup-tools_0.41-10_amd64.deb ...
Jun 16 07:27:13 fv-az72-309 provisioner[1638]: [07:27:13] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:13 fv-az72-309 provisioner[1638]: Unpacking cgroup-tools (0.41-10) ...
Jun 16 07:27:13 fv-az72-309 provisioner[1638]: [07:27:13] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:13 fv-az72-309 provisioner[1638]: Setting up libcgroup1:amd64 (0.41-10) ...
Jun 16 07:27:14 fv-az72-309 provisioner[1638]: [07:27:14] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:14 fv-az72-309 provisioner[1638]: Setting up cgroup-tools (0.41-10) ...
Jun 16 07:27:14 fv-az72-309 provisioner[1638]: [07:27:14] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:14 fv-az72-309 provisioner[1638]: Processing triggers for libc-bin (2.31-0ubuntu9.9) ...
Jun 16 07:27:16 fv-az72-309 provisioner[1638]: [07:27:16] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:16 fv-az72-309 provisioner[1638]: Processing triggers for man-db (2.9.1-1) ...
Jun 16 07:27:18 fv-az72-309 dbus-daemon[661]: [system] Activating via systemd: service name='org.freedesktop.PackageKit' unit='packagekit.service' requested by ':1.14' (uid=0 pid=2069 comm="/usr/bin/gdbus call --system --dest org.freedeskto" label="unconfined")
Jun 16 07:27:18 fv-az72-309 systemd[1]: Starting PackageKit Daemon...
Jun 16 07:27:18 fv-az72-309 PackageKit[2072]: daemon start
Jun 16 07:27:18 fv-az72-309 dbus-daemon[661]: [system] Successfully activated service 'org.freedesktop.PackageKit'
Jun 16 07:27:18 fv-az72-309 systemd[1]: Started PackageKit Daemon.
Jun 16 07:27:20 fv-az72-309 provisioner[1638]: [07:27:20] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:20 fv-az72-309 provisioner[1638]: + apt-get install -y libcgroup1 cgroup-tools
Jun 16 07:27:20 fv-az72-309 provisioner[1638]: [07:27:20] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:20 fv-az72-309 provisioner[1638]: Reading package lists...
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: [07:27:21] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: Building dependency tree...
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: [07:27:21] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: Reading state information...
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: [07:27:21] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: cgroup-tools is already the newest version (0.41-10).
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: [07:27:21] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: libcgroup1 is already the newest version (0.41-10).
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: [07:27:21] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: 0 upgraded, 0 newly installed, 0 to remove and 13 not upgraded.
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: [07:27:21] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: ++ command -v docker
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: [07:27:21] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: + '[' '!' -x /usr/bin/docker ']'
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: [07:27:21] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: ++ grep MemTotal /proc/meminfo
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: [07:27:21] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: ++ awk '{print $2}'
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: [07:27:21] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: + mem_total_in_bytes=7283838976
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: [07:27:21] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: + mem_total_minus1g_in_bytes=6210097152
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: [07:27:21] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: ++ grep SwapTotal /proc/meminfo
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: [07:27:21] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: ++ awk '{print $2}'
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: [07:27:21] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: + swap_total_in_bytes=4294963200
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: [07:27:21] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: + total_in_bytes=11578802176
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: [07:27:21] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: + total_minus2g_in_bytes=9431318528
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: [07:27:21] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: + actions_runner_cgroup='group actions_runner { memory { } }'
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: [07:27:21] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: + actions_job_cgroup='group actions_job { memory { memory.limit_in_bytes = 6210097152; memory.memsw.limit_in_bytes = 9431318528; } }'
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: [07:27:21] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: + echo 'group actions_runner { memory { } }'
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: [07:27:21] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: + echo 'group actions_job { memory { memory.limit_in_bytes = 6210097152; memory.memsw.limit_in_bytes = 9431318528; } }'
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: [07:27:21] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: + echo 'root:provisioner memory actions_runner'
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: [07:27:21] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: + echo 'runner:Runner.Listener memory actions_runner'
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: [07:27:21] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: + echo 'runner:Runner.Worker memory actions_runner'
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: [07:27:21] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: + echo 'runner memory actions_job'
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: [07:27:21] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: + '[' '!' -f /etc/docker/daemon.json ']'
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: [07:27:21] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: + echo '{ "exec-opts": ["native.cgroupdriver=cgroupfs"], "cgroup-parent": "/actions_job" }'
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: [07:27:21] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: + echo 'GRUB_CMDLINE_LINUX="cgroup_enable=memory swapaccount=1"'
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: [07:27:21] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:21 fv-az72-309 provisioner[1638]: + update-grub
Jun 16 07:27:27 fv-az72-309 provisioner[1638]: [07:27:27] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:27 fv-az72-309 provisioner[1638]: Sourcing file `/etc/default/grub'
Jun 16 07:27:27 fv-az72-309 provisioner[1638]: [07:27:27] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:27 fv-az72-309 provisioner[1638]: Sourcing file `/etc/default/grub.d/40-force-partuuid.cfg'
Jun 16 07:27:27 fv-az72-309 provisioner[1638]: [07:27:27] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:27 fv-az72-309 provisioner[1638]: Sourcing file `/etc/default/grub.d/40-runner.cfg'
Jun 16 07:27:27 fv-az72-309 provisioner[1638]: [07:27:27] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:27 fv-az72-309 provisioner[1638]: Sourcing file `/etc/default/grub.d/50-cloudimg-settings.cfg'
Jun 16 07:27:27 fv-az72-309 provisioner[1638]: [07:27:27] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:27 fv-az72-309 provisioner[1638]: Sourcing file `/etc/default/grub.d/init-select.cfg'
Jun 16 07:27:27 fv-az72-309 provisioner[1638]: [07:27:27] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:27 fv-az72-309 provisioner[1638]: Generating grub configuration file ...
Jun 16 07:27:27 fv-az72-309 provisioner[1638]: [07:27:27] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:27 fv-az72-309 provisioner[1638]: GRUB_FORCE_PARTUUID is set, will attempt initrdless boot
Jun 16 07:27:27 fv-az72-309 provisioner[1638]: [07:27:27] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:27 fv-az72-309 provisioner[1638]: Found linux image: /boot/vmlinuz-5.13.0-1029-azure
Jun 16 07:27:27 fv-az72-309 provisioner[1638]: [07:27:27] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:27 fv-az72-309 provisioner[1638]: Found initrd image: /boot/initrd.img-5.13.0-1029-azure
Jun 16 07:27:28 fv-az72-309 kernel: SGI XFS with ACLs, security attributes, realtime, quota, no debug enabled
Jun 16 07:27:28 fv-az72-309 kernel: raid6: avx2x4 gen() 424996 MB/s
Jun 16 07:27:28 fv-az72-309 kernel: raid6: avx2x4 xor() 135157 MB/s
Jun 16 07:27:28 fv-az72-309 kernel: raid6: avx2x2 gen() 392111 MB/s
Jun 16 07:27:28 fv-az72-309 kernel: raid6: avx2x2 xor() 239587 MB/s
Jun 16 07:27:28 fv-az72-309 kernel: raid6: avx2x1 gen() 373907 MB/s
Jun 16 07:27:28 fv-az72-309 kernel: raid6: avx2x1 xor() 219432 MB/s
Jun 16 07:27:28 fv-az72-309 kernel: raid6: sse2x4 gen() 215555 MB/s
Jun 16 07:27:28 fv-az72-309 kernel: raid6: sse2x4 xor() 128859 MB/s
Jun 16 07:27:28 fv-az72-309 kernel: raid6: sse2x2 gen() 228099 MB/s
Jun 16 07:27:29 fv-az72-309 kernel: raid6: sse2x2 xor() 131971 MB/s
Jun 16 07:27:29 fv-az72-309 kernel: raid6: sse2x1 gen() 167466 MB/s
Jun 16 07:27:29 fv-az72-309 kernel: raid6: sse2x1 xor() 88530 MB/s
Jun 16 07:27:29 fv-az72-309 kernel: raid6: using algorithm avx2x4 gen() 18784 MB/s
Jun 16 07:27:29 fv-az72-309 kernel: raid6: .... xor() 135157 MB/s, rmw enabled
Jun 16 07:27:29 fv-az72-309 kernel: raid6: using avx2x2 recovery algorithm
Jun 16 07:27:29 fv-az72-309 kernel: xor: automatically using best checksumming function avx
Jun 16 07:27:29 fv-az72-309 kernel: Btrfs loaded, crc32c=crc32c-intel, zoned=yes
Jun 16 07:27:29 fv-az72-309 provisioner[1638]: [07:27:29] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:29 fv-az72-309 provisioner[1638]: File descriptor 8 (socket:[32650]) leaked on lvs invocation. Parent PID 2736: /bin/sh
Jun 16 07:27:29 fv-az72-309 os-prober[2755]: debug: running /usr/lib/os-probes/mounted/05efi on mounted /dev/sda1
Jun 16 07:27:29 fv-az72-309 05efi[2757]: debug: Not on UEFI platform
Jun 16 07:27:29 fv-az72-309 os-prober[2758]: debug: running /usr/lib/os-probes/mounted/10freedos on mounted /dev/sda1
Jun 16 07:27:29 fv-az72-309 10freedos[2760]: debug: /dev/sda1 is not a FAT partition: exiting
Jun 16 07:27:29 fv-az72-309 os-prober[2761]: debug: running /usr/lib/os-probes/mounted/10qnx on mounted /dev/sda1
Jun 16 07:27:29 fv-az72-309 10qnx[2763]: debug: /dev/sda1 is not a QNX4 partition: exiting
Jun 16 07:27:29 fv-az72-309 os-prober[2764]: debug: running /usr/lib/os-probes/mounted/20macosx on mounted /dev/sda1
Jun 16 07:27:29 fv-az72-309 macosx-prober[2766]: debug: /dev/sda1 is not an HFS+ partition: exiting
Jun 16 07:27:29 fv-az72-309 os-prober[2767]: debug: running /usr/lib/os-probes/mounted/20microsoft on mounted /dev/sda1
Jun 16 07:27:29 fv-az72-309 20microsoft[2769]: debug: /dev/sda1 is not a MS partition: exiting
Jun 16 07:27:29 fv-az72-309 os-prober[2770]: debug: running /usr/lib/os-probes/mounted/30utility on mounted /dev/sda1
Jun 16 07:27:29 fv-az72-309 30utility[2772]: debug: /dev/sda1 is not a FAT partition: exiting
Jun 16 07:27:29 fv-az72-309 os-prober[2773]: debug: running /usr/lib/os-probes/mounted/40lsb on mounted /dev/sda1
Jun 16 07:27:29 fv-az72-309 os-prober[2775]: debug: running /usr/lib/os-probes/mounted/70hurd on mounted /dev/sda1
Jun 16 07:27:29 fv-az72-309 os-prober[2777]: debug: running /usr/lib/os-probes/mounted/80minix on mounted /dev/sda1
Jun 16 07:27:29 fv-az72-309 os-prober[2779]: debug: running /usr/lib/os-probes/mounted/83haiku on mounted /dev/sda1
Jun 16 07:27:29 fv-az72-309 83haiku[2781]: debug: /dev/sda1 is not a BeFS partition: exiting
Jun 16 07:27:29 fv-az72-309 os-prober[2782]: debug: running /usr/lib/os-probes/mounted/90linux-distro on mounted /dev/sda1
Jun 16 07:27:29 fv-az72-309 os-prober[2786]: debug: running /usr/lib/os-probes/mounted/90solaris on mounted /dev/sda1
Jun 16 07:27:29 fv-az72-309 os-prober[2793]: debug: running /usr/lib/os-probes/50mounted-tests on /dev/sdb1
Jun 16 07:27:29 fv-az72-309 50mounted-tests[2813]: debug: mounted using GRUB ext2 filesystem driver
Jun 16 07:27:29 fv-az72-309 50mounted-tests[2814]: debug: running subtest /usr/lib/os-probes/mounted/05efi
Jun 16 07:27:29 fv-az72-309 05efi[2816]: debug: Not on UEFI platform
Jun 16 07:27:29 fv-az72-309 50mounted-tests[2817]: debug: running subtest /usr/lib/os-probes/mounted/10freedos
Jun 16 07:27:29 fv-az72-309 10freedos[2819]: debug: /dev/sdb1 is not a FAT partition: exiting
Jun 16 07:27:29 fv-az72-309 50mounted-tests[2820]: debug: running subtest /usr/lib/os-probes/mounted/10qnx
Jun 16 07:27:29 fv-az72-309 10qnx[2822]: debug: /dev/sdb1 is not a QNX4 partition: exiting
Jun 16 07:27:29 fv-az72-309 50mounted-tests[2823]: debug: running subtest /usr/lib/os-probes/mounted/20macosx
Jun 16 07:27:29 fv-az72-309 macosx-prober[2825]: debug: /dev/sdb1 is not an HFS+ partition: exiting
Jun 16 07:27:29 fv-az72-309 50mounted-tests[2826]: debug: running subtest /usr/lib/os-probes/mounted/20microsoft
Jun 16 07:27:29 fv-az72-309 20microsoft[2828]: debug: /dev/sdb1 is not a MS partition: exiting
Jun 16 07:27:29 fv-az72-309 50mounted-tests[2829]: debug: running subtest /usr/lib/os-probes/mounted/30utility
Jun 16 07:27:29 fv-az72-309 30utility[2831]: debug: /dev/sdb1 is not a FAT partition: exiting
Jun 16 07:27:29 fv-az72-309 50mounted-tests[2832]: debug: running subtest /usr/lib/os-probes/mounted/40lsb
Jun 16 07:27:29 fv-az72-309 40lsb[2858]: result: /dev/sdb1:Ubuntu 20.04.4 LTS (20.04):Ubuntu:linux
Jun 16 07:27:29 fv-az72-309 50mounted-tests[2859]: debug: os found by subtest /usr/lib/os-probes/mounted/40lsb
Jun 16 07:27:29 fv-az72-309 systemd[1]: var-lib-os\x2dprober-mount.mount: Succeeded.
Jun 16 07:27:29 fv-az72-309 os-prober[2862]: debug: os detected by /usr/lib/os-probes/50mounted-tests
Jun 16 07:27:29 fv-az72-309 os-prober[2868]: debug: running /usr/lib/os-probes/50mounted-tests on /dev/sdb14
Jun 16 07:27:29 fv-az72-309 50mounted-tests[2876]: debug: /dev/sdb14 type not recognised; skipping
Jun 16 07:27:29 fv-az72-309 os-prober[2877]: debug: os detected by /usr/lib/os-probes/50mounted-tests
Jun 16 07:27:29 fv-az72-309 os-prober[2894]: debug: running /usr/lib/os-probes/mounted/05efi on mounted /dev/sdb15
Jun 16 07:27:29 fv-az72-309 05efi[2896]: debug: Not on UEFI platform
Jun 16 07:27:29 fv-az72-309 os-prober[2897]: debug: running /usr/lib/os-probes/mounted/10freedos on mounted /dev/sdb15
Jun 16 07:27:29 fv-az72-309 10freedos[2899]: debug: /dev/sdb15 is a FAT32 partition
Jun 16 07:27:29 fv-az72-309 os-prober[2902]: debug: running /usr/lib/os-probes/mounted/10qnx on mounted /dev/sdb15
Jun 16 07:27:29 fv-az72-309 10qnx[2904]: debug: /dev/sdb15 is not a QNX4 partition: exiting
Jun 16 07:27:29 fv-az72-309 os-prober[2905]: debug: running /usr/lib/os-probes/mounted/20macosx on mounted /dev/sdb15
Jun 16 07:27:29 fv-az72-309 macosx-prober[2907]: debug: /dev/sdb15 is not an HFS+ partition: exiting
Jun 16 07:27:29 fv-az72-309 os-prober[2908]: debug: running /usr/lib/os-probes/mounted/20microsoft on mounted /dev/sdb15
Jun 16 07:27:29 fv-az72-309 20microsoft[2910]: debug: /dev/sdb15 is a FAT32 partition
Jun 16 07:27:29 fv-az72-309 os-prober[2919]: debug: running /usr/lib/os-probes/mounted/30utility on mounted /dev/sdb15
Jun 16 07:27:29 fv-az72-309 30utility[2921]: debug: /dev/sdb15 is a FAT32 partition
Jun 16 07:27:29 fv-az72-309 os-prober[2926]: debug: running /usr/lib/os-probes/mounted/40lsb on mounted /dev/sdb15
Jun 16 07:27:29 fv-az72-309 os-prober[2928]: debug: running /usr/lib/os-probes/mounted/70hurd on mounted /dev/sdb15
Jun 16 07:27:29 fv-az72-309 os-prober[2930]: debug: running /usr/lib/os-probes/mounted/80minix on mounted /dev/sdb15
Jun 16 07:27:29 fv-az72-309 os-prober[2932]: debug: running /usr/lib/os-probes/mounted/83haiku on mounted /dev/sdb15
Jun 16 07:27:29 fv-az72-309 83haiku[2934]: debug: /dev/sdb15 is not a BeFS partition: exiting
Jun 16 07:27:29 fv-az72-309 os-prober[2935]: debug: running /usr/lib/os-probes/mounted/90linux-distro on mounted /dev/sdb15
Jun 16 07:27:29 fv-az72-309 os-prober[2939]: debug: running /usr/lib/os-probes/mounted/90solaris on mounted /dev/sdb15
Jun 16 07:27:30 fv-az72-309 provisioner[1638]: [07:27:30] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:30 fv-az72-309 provisioner[1638]: Found Ubuntu 20.04.4 LTS (20.04) on /dev/sdb1
Jun 16 07:27:30 fv-az72-309 linux-boot-prober[2999]: debug: running /usr/lib/linux-boot-probes/50mounted-tests
Jun 16 07:27:30 fv-az72-309 50mounted-tests[3022]: debug: running /usr/lib/linux-boot-probes/mounted/40grub /dev/sdb1 /dev/sdb1 /var/lib/os-prober/mount ext2
Jun 16 07:27:30 fv-az72-309 50mounted-tests[3024]: debug: running /usr/lib/linux-boot-probes/mounted/40grub2 /dev/sdb1 /dev/sdb1 /var/lib/os-prober/mount ext2
Jun 16 07:27:30 fv-az72-309 40grub2[3027]: debug: parsing grub.cfg
Jun 16 07:27:30 fv-az72-309 40grub2[3028]: debug: parsing: #
Jun 16 07:27:30 fv-az72-309 40grub2[3029]: debug: parsing: # DO NOT EDIT THIS FILE
Jun 16 07:27:30 fv-az72-309 40grub2[3030]: debug: parsing: #
Jun 16 07:27:30 fv-az72-309 40grub2[3031]: debug: parsing: # It is automatically generated by grub-mkconfig using templates
Jun 16 07:27:30 fv-az72-309 40grub2[3032]: debug: parsing: # from /etc/grub.d and settings from /etc/default/grub
Jun 16 07:27:30 fv-az72-309 40grub2[3033]: debug: parsing: #
Jun 16 07:27:30 fv-az72-309 40grub2[3034]: debug: parsing:
Jun 16 07:27:30 fv-az72-309 40grub2[3035]: debug: parsing: ### BEGIN /etc/grub.d/00_header ###
Jun 16 07:27:30 fv-az72-309 40grub2[3036]: debug: parsing: if [ -s $prefix/grubenv ]; then
Jun 16 07:27:30 fv-az72-309 40grub2[3037]: debug: parsing: set have_grubenv=true
Jun 16 07:27:30 fv-az72-309 40grub2[3038]: debug: parsing: load_env
Jun 16 07:27:30 fv-az72-309 40grub2[3039]: debug: parsing: fi
Jun 16 07:27:30 fv-az72-309 40grub2[3040]: debug: parsing: if [ "${initrdfail}" = 2 ]; then
Jun 16 07:27:30 fv-az72-309 40grub2[3041]: debug: parsing: set initrdfail=
Jun 16 07:27:30 fv-az72-309 40grub2[3042]: debug: parsing: elif [ "${initrdfail}" = 1 ]; then
Jun 16 07:27:30 fv-az72-309 40grub2[3043]: debug: parsing: set next_entry="${prev_entry}"
Jun 16 07:27:30 fv-az72-309 40grub2[3044]: debug: parsing: set prev_entry=
Jun 16 07:27:30 fv-az72-309 40grub2[3045]: debug: parsing: save_env prev_entry
Jun 16 07:27:30 fv-az72-309 40grub2[3046]: debug: parsing: if [ "${next_entry}" ]; then
Jun 16 07:27:30 fv-az72-309 40grub2[3047]: debug: parsing: set initrdfail=2
Jun 16 07:27:30 fv-az72-309 40grub2[3048]: debug: parsing: fi
Jun 16 07:27:30 fv-az72-309 40grub2[3049]: debug: parsing: fi
Jun 16 07:27:30 fv-az72-309 40grub2[3050]: debug: parsing: if [ "${next_entry}" ] ; then
Jun 16 07:27:30 fv-az72-309 40grub2[3051]: debug: parsing: set default="${next_entry}"
Jun 16 07:27:30 fv-az72-309 40grub2[3052]: debug: parsing: set next_entry=
Jun 16 07:27:30 fv-az72-309 40grub2[3053]: debug: parsing: save_env next_entry
Jun 16 07:27:30 fv-az72-309 40grub2[3054]: debug: parsing: set boot_once=true
Jun 16 07:27:30 fv-az72-309 40grub2[3055]: debug: parsing: else
Jun 16 07:27:30 fv-az72-309 40grub2[3056]: debug: parsing: set default="0"
Jun 16 07:27:30 fv-az72-309 40grub2[3057]: debug: parsing: fi
Jun 16 07:27:30 fv-az72-309 40grub2[3058]: debug: parsing:
Jun 16 07:27:30 fv-az72-309 40grub2[3059]: debug: parsing: if [ x"${feature_menuentry_id}" = xy ]; then
Jun 16 07:27:30 fv-az72-309 40grub2[3060]: debug: parsing: menuentry_id_option="--id"
Jun 16 07:27:30 fv-az72-309 40grub2[3061]: debug: parsing: else
Jun 16 07:27:30 fv-az72-309 40grub2[3062]: debug: parsing: menuentry_id_option=""
Jun 16 07:27:30 fv-az72-309 40grub2[3063]: debug: parsing: fi
Jun 16 07:27:30 fv-az72-309 40grub2[3064]: debug: parsing:
Jun 16 07:27:30 fv-az72-309 40grub2[3065]: debug: parsing: export menuentry_id_option
Jun 16 07:27:30 fv-az72-309 40grub2[3066]: debug: parsing:
Jun 16 07:27:30 fv-az72-309 40grub2[3067]: debug: parsing: if [ "${prev_saved_entry}" ]; then
Jun 16 07:27:30 fv-az72-309 40grub2[3068]: debug: parsing: set saved_entry="${prev_saved_entry}"
Jun 16 07:27:30 fv-az72-309 40grub2[3069]: debug: parsing: save_env saved_entry
Jun 16 07:27:30 fv-az72-309 40grub2[3070]: debug: parsing: set prev_saved_entry=
Jun 16 07:27:30 fv-az72-309 40grub2[3071]: debug: parsing: save_env prev_saved_entry
Jun 16 07:27:30 fv-az72-309 40grub2[3072]: debug: parsing: set boot_once=true
Jun 16 07:27:30 fv-az72-309 40grub2[3073]: debug: parsing: fi
Jun 16 07:27:30 fv-az72-309 40grub2[3074]: debug: parsing:
Jun 16 07:27:30 fv-az72-309 40grub2[3075]: debug: parsing: function savedefault {
Jun 16 07:27:30 fv-az72-309 40grub2[3076]: debug: parsing: if [ -z "${boot_once}" ]; then
Jun 16 07:27:30 fv-az72-309 40grub2[3077]: debug: parsing: saved_entry="${chosen}"
Jun 16 07:27:30 fv-az72-309 40grub2[3078]: debug: parsing: save_env saved_entry
Jun 16 07:27:30 fv-az72-309 40grub2[3079]: debug: parsing: fi
Jun 16 07:27:30 fv-az72-309 40grub2[3080]: debug: parsing: }
Jun 16 07:27:30 fv-az72-309 40grub2[3081]: debug: parsing: function initrdfail {
Jun 16 07:27:30 fv-az72-309 40grub2[3082]: debug: parsing: if [ -n "${have_grubenv}" ]; then if [ -n "${partuuid}" ]; then
Jun 16 07:27:30 fv-az72-309 40grub2[3083]: debug: parsing: if [ -z "${initrdfail}" ]; then
Jun 16 07:27:30 fv-az72-309 40grub2[3084]: debug: parsing: set initrdfail=1
Jun 16 07:27:30 fv-az72-309 40grub2[3085]: debug: parsing: if [ -n "${boot_once}" ]; then
Jun 16 07:27:30 fv-az72-309 40grub2[3086]: debug: parsing: set prev_entry="${default}"
Jun 16 07:27:30 fv-az72-309 40grub2[3087]: debug: parsing: save_env prev_entry
Jun 16 07:27:30 fv-az72-309 40grub2[3088]: debug: parsing: fi
Jun 16 07:27:30 fv-az72-309 40grub2[3089]: debug: parsing: fi
Jun 16 07:27:30 fv-az72-309 40grub2[3090]: debug: parsing: save_env initrdfail
Jun 16 07:27:30 fv-az72-309 40grub2[3091]: debug: parsing: fi; fi
Jun 16 07:27:30 fv-az72-309 40grub2[3092]: debug: parsing: }
Jun 16 07:27:30 fv-az72-309 40grub2[3093]: debug: parsing: function recordfail {
Jun 16 07:27:30 fv-az72-309 40grub2[3094]: debug: parsing: set recordfail=1
Jun 16 07:27:30 fv-az72-309 40grub2[3095]: debug: parsing: if [ -n "${have_grubenv}" ]; then if [ -z "${boot_once}" ]; then save_env recordfail; fi; fi
Jun 16 07:27:30 fv-az72-309 40grub2[3096]: debug: parsing: }
Jun 16 07:27:30 fv-az72-309 40grub2[3097]: debug: parsing: function load_video {
Jun 16 07:27:30 fv-az72-309 40grub2[3098]: debug: parsing: if [ x$feature_all_video_module = xy ]; then
Jun 16 07:27:30 fv-az72-309 40grub2[3099]: debug: parsing: insmod all_video
Jun 16 07:27:30 fv-az72-309 40grub2[3100]: debug: parsing: else
Jun 16 07:27:30 fv-az72-309 40grub2[3101]: debug: parsing: insmod efi_gop
Jun 16 07:27:30 fv-az72-309 40grub2[3102]: debug: parsing: insmod efi_uga
Jun 16 07:27:30 fv-az72-309 40grub2[3103]: debug: parsing: insmod ieee1275_fb
Jun 16 07:27:30 fv-az72-309 40grub2[3104]: debug: parsing: insmod vbe
Jun 16 07:27:30 fv-az72-309 40grub2[3105]: debug: parsing: insmod vga
Jun 16 07:27:30 fv-az72-309 40grub2[3106]: debug: parsing: insmod video_bochs
Jun 16 07:27:30 fv-az72-309 40grub2[3107]: debug: parsing: insmod video_cirrus
Jun 16 07:27:30 fv-az72-309 40grub2[3108]: debug: parsing: fi
Jun 16 07:27:30 fv-az72-309 40grub2[3109]: debug: parsing: }
Jun 16 07:27:30 fv-az72-309 40grub2[3110]: debug: parsing:
Jun 16 07:27:30 fv-az72-309 40grub2[3111]: debug: parsing: serial --speed=9600 --unit=0 --word=8 --parity=no --stop=1
Jun 16 07:27:30 fv-az72-309 40grub2[3112]: debug: parsing: terminal_input serial
Jun 16 07:27:30 fv-az72-309 40grub2[3113]: debug: parsing: terminal_output serial
Jun 16 07:27:30 fv-az72-309 40grub2[3114]: debug: parsing: if [ "${recordfail}" = 1 ] ; then
Jun 16 07:27:30 fv-az72-309 40grub2[3115]: debug: parsing: set timeout=30
Jun 16 07:27:30 fv-az72-309 40grub2[3116]: debug: parsing: else
Jun 16 07:27:30 fv-az72-309 40grub2[3117]: debug: parsing: if [ x$feature_timeout_style = xy ] ; then
Jun 16 07:27:30 fv-az72-309 40grub2[3118]: debug: parsing: set timeout_style=countdown
Jun 16 07:27:30 fv-az72-309 40grub2[3119]: debug: parsing: set timeout=1
Jun 16 07:27:30 fv-az72-309 40grub2[3120]: debug: parsing: # Fallback hidden-timeout code in case the timeout_style feature is
Jun 16 07:27:30 fv-az72-309 40grub2[3121]: debug: parsing: # unavailable.
Jun 16 07:27:30 fv-az72-309 40grub2[3122]: debug: parsing: elif sleep --interruptible 1 ; then
Jun 16 07:27:30 fv-az72-309 40grub2[3123]: debug: parsing: set timeout=0
Jun 16 07:27:30 fv-az72-309 40grub2[3124]: debug: parsing: fi
Jun 16 07:27:30 fv-az72-309 40grub2[3125]: debug: parsing: fi
Jun 16 07:27:30 fv-az72-309 40grub2[3126]: debug: parsing: ### END /etc/grub.d/00_header ###
Jun 16 07:27:30 fv-az72-309 40grub2[3127]: debug: parsing:
Jun 16 07:27:30 fv-az72-309 40grub2[3128]: debug: parsing: ### BEGIN /etc/grub.d/01_track_initrdless_boot_fallback ###
Jun 16 07:27:30 fv-az72-309 40grub2[3129]: debug: parsing: if [ -n "${have_grubenv}" ]; then
Jun 16 07:27:30 fv-az72-309 40grub2[3130]: debug: parsing: if [ -n "${initrdfail}" ]; then
Jun 16 07:27:30 fv-az72-309 40grub2[3131]: debug: parsing: set initrdless_boot_fallback_triggered="${initrdfail}"
Jun 16 07:27:30 fv-az72-309 40grub2[3132]: debug: parsing: else
Jun 16 07:27:30 fv-az72-309 40grub2[3133]: debug: parsing: unset initrdless_boot_fallback_triggered
Jun 16 07:27:30 fv-az72-309 40grub2[3134]: debug: parsing: fi
Jun 16 07:27:30 fv-az72-309 40grub2[3135]: debug: parsing: save_env initrdless_boot_fallback_triggered
Jun 16 07:27:30 fv-az72-309 40grub2[3136]: debug: parsing: fi
Jun 16 07:27:30 fv-az72-309 40grub2[3137]: debug: parsing: ### END /etc/grub.d/01_track_initrdless_boot_fallback ###
Jun 16 07:27:30 fv-az72-309 40grub2[3138]: debug: parsing:
Jun 16 07:27:30 fv-az72-309 40grub2[3139]: debug: parsing: ### BEGIN /etc/grub.d/05_debian_theme ###
Jun 16 07:27:30 fv-az72-309 40grub2[3140]: debug: parsing: set menu_color_normal=white/black
Jun 16 07:27:30 fv-az72-309 40grub2[3141]: debug: parsing: set menu_color_highlight=black/light-gray
Jun 16 07:27:30 fv-az72-309 40grub2[3142]: debug: parsing: ### END /etc/grub.d/05_debian_theme ###
Jun 16 07:27:30 fv-az72-309 40grub2[3143]: debug: parsing:
Jun 16 07:27:30 fv-az72-309 40grub2[3144]: debug: parsing: ### BEGIN /etc/grub.d/10_linux ###
Jun 16 07:27:30 fv-az72-309 40grub2[3145]: debug: parsing: #
Jun 16 07:27:30 fv-az72-309 40grub2[3146]: debug: parsing: # GRUB_FORCE_PARTUUID is set, will attempt initrdless boot
Jun 16 07:27:30 fv-az72-309 40grub2[3147]: debug: parsing: # Upon panic fallback to booting with initrd
Jun 16 07:27:30 fv-az72-309 40grub2[3148]: debug: parsing: set partuuid=2b1f5b8e-4041-4065-af1c-792f94a6d205
Jun 16 07:27:30 fv-az72-309 40grub2[3149]: debug: parsing: function gfxmode {
Jun 16 07:27:30 fv-az72-309 40grub2[3150]: debug: parsing: set gfxpayload="${1}"
Jun 16 07:27:30 fv-az72-309 40grub2[3151]: debug: parsing: if [ "${1}" = "keep" ]; then
Jun 16 07:27:30 fv-az72-309 40grub2[3152]: debug: parsing: set vt_handoff=vt.handoff=7
Jun 16 07:27:30 fv-az72-309 40grub2[3153]: debug: parsing: else
Jun 16 07:27:30 fv-az72-309 40grub2[3154]: debug: parsing: set vt_handoff=
Jun 16 07:27:30 fv-az72-309 40grub2[3155]: debug: parsing: fi
Jun 16 07:27:30 fv-az72-309 40grub2[3156]: debug: parsing: }
Jun 16 07:27:30 fv-az72-309 40grub2[3157]: debug: parsing: if [ "${recordfail}" != 1 ]; then
Jun 16 07:27:30 fv-az72-309 40grub2[3158]: debug: parsing: if [ -e ${prefix}/gfxblacklist.txt ]; then
Jun 16 07:27:30 fv-az72-309 40grub2[3159]: debug: parsing: if [ ${grub_platform} != pc ]; then
Jun 16 07:27:30 fv-az72-309 40grub2[3160]: debug: parsing: set linux_gfx_mode=keep
Jun 16 07:27:30 fv-az72-309 40grub2[3161]: debug: parsing: elif hwmatch ${prefix}/gfxblacklist.txt 3; then
Jun 16 07:27:30 fv-az72-309 40grub2[3162]: debug: parsing: if [ ${match} = 0 ]; then
Jun 16 07:27:30 fv-az72-309 40grub2[3163]: debug: parsing: set linux_gfx_mode=keep
Jun 16 07:27:30 fv-az72-309 40grub2[3164]: debug: parsing: else
Jun 16 07:27:30 fv-az72-309 40grub2[3165]: debug: parsing: set linux_gfx_mode=text
Jun 16 07:27:30 fv-az72-309 40grub2[3166]: debug: parsing: fi
Jun 16 07:27:30 fv-az72-309 40grub2[3167]: debug: parsing: else
Jun 16 07:27:30 fv-az72-309 40grub2[3168]: debug: parsing: set linux_gfx_mode=text
Jun 16 07:27:30 fv-az72-309 40grub2[3169]: debug: parsing: fi
Jun 16 07:27:30 fv-az72-309 40grub2[3170]: debug: parsing: else
Jun 16 07:27:30 fv-az72-309 40grub2[3171]: debug: parsing: set linux_gfx_mode=keep
Jun 16 07:27:30 fv-az72-309 40grub2[3172]: debug: parsing: fi
Jun 16 07:27:30 fv-az72-309 40grub2[3173]: debug: parsing: else
Jun 16 07:27:30 fv-az72-309 40grub2[3174]: debug: parsing: set linux_gfx_mode=text
Jun 16 07:27:30 fv-az72-309 40grub2[3175]: debug: parsing: fi
Jun 16 07:27:30 fv-az72-309 40grub2[3176]: debug: parsing: export linux_gfx_mode
Jun 16 07:27:30 fv-az72-309 40grub2[3177]: debug: parsing: menuentry 'Ubuntu' --class ubuntu --class gnu-linux --class gnu --class os $menuentry_id_option 'gnulinux-simple-3b66cf28-8c39-478b-82b0-294032b5bd9d' {
Jun 16 07:27:30 fv-az72-309 40grub2[3188]: debug: parsing: recordfail
Jun 16 07:27:30 fv-az72-309 40grub2[3189]: debug: parsing: load_video
Jun 16 07:27:30 fv-az72-309 40grub2[3190]: debug: parsing: gfxmode $linux_gfx_mode
Jun 16 07:27:30 fv-az72-309 40grub2[3191]: debug: parsing: insmod gzio
Jun 16 07:27:30 fv-az72-309 40grub2[3192]: debug: parsing: if [ x$grub_platform = xxen ]; then insmod xzio; insmod lzopio; fi
Jun 16 07:27:30 fv-az72-309 40grub2[3193]: debug: parsing: insmod part_gpt
Jun 16 07:27:30 fv-az72-309 40grub2[3194]: debug: parsing: insmod ext2
Jun 16 07:27:30 fv-az72-309 40grub2[3195]: debug: parsing: if [ x$feature_platform_search_hint = xy ]; then
Jun 16 07:27:30 fv-az72-309 40grub2[3196]: debug: parsing: search --no-floppy --fs-uuid --set=root 3b66cf28-8c39-478b-82b0-294032b5bd9d
Jun 16 07:27:30 fv-az72-309 40grub2[3197]: debug: parsing: else
Jun 16 07:27:30 fv-az72-309 40grub2[3198]: debug: parsing: search --no-floppy --fs-uuid --set=root 3b66cf28-8c39-478b-82b0-294032b5bd9d
Jun 16 07:27:30 fv-az72-309 40grub2[3199]: debug: parsing: fi
Jun 16 07:27:30 fv-az72-309 40grub2[3200]: debug: parsing: if [ "${initrdfail}" = 1 ]; then
Jun 16 07:27:30 fv-az72-309 40grub2[3201]: debug: parsing: echo 'GRUB_FORCE_PARTUUID set, initrdless boot failed. Attempting with initrd.'
Jun 16 07:27:30 fv-az72-309 40grub2[3202]: debug: parsing: linux /boot/vmlinuz-5.13.0-1029-azure root=PARTUUID=2b1f5b8e-4041-4065-af1c-792f94a6d205 ro console=tty1 console=ttyS0 earlyprintk=ttyS0
Jun 16 07:27:30 fv-az72-309 40grub2[3206]: debug: parsing: initrd /boot/initrd.img-5.13.0-1029-azure
Jun 16 07:27:30 fv-az72-309 40grub2[3210]: debug: parsing: else
Jun 16 07:27:30 fv-az72-309 40grub2[3211]: debug: parsing: echo 'GRUB_FORCE_PARTUUID set, attempting initrdless boot.'
Jun 16 07:27:30 fv-az72-309 40grub2[3212]: debug: parsing: linux /boot/vmlinuz-5.13.0-1029-azure root=PARTUUID=2b1f5b8e-4041-4065-af1c-792f94a6d205 ro console=tty1 console=ttyS0 earlyprintk=ttyS0 panic=-1
Jun 16 07:27:30 fv-az72-309 40grub2[3216]: debug: parsing: fi
Jun 16 07:27:30 fv-az72-309 40grub2[3217]: debug: parsing: initrdfail
Jun 16 07:27:30 fv-az72-309 40grub2[3218]: debug: parsing: }
Jun 16 07:27:30 fv-az72-309 40grub2[3219]: result: /dev/sdb1:/dev/sdb1:Ubuntu:/boot/vmlinuz-5.13.0-1029-azure:/boot/initrd.img-5.13.0-1029-azure:root=PARTUUID=2b1f5b8e-4041-4065-af1c-792f94a6d205 ro console=tty1 console=ttyS0 earlyprintk=ttyS0 panic=-1
Jun 16 07:27:30 fv-az72-309 40grub2[3220]: debug: parsing: submenu 'Advanced options for Ubuntu' $menuentry_id_option 'gnulinux-advanced-3b66cf28-8c39-478b-82b0-294032b5bd9d' {
Jun 16 07:27:30 fv-az72-309 40grub2[3221]: debug: parsing: menuentry 'Ubuntu, with Linux 5.13.0-1029-azure' --class ubuntu --class gnu-linux --class gnu --class os $menuentry_id_option 'gnulinux-5.13.0-1029-azure-advanced-3b66cf28-8c39-478b-82b0-294032b5bd9d' {
Jun 16 07:27:30 fv-az72-309 40grub2[3232]: debug: parsing: recordfail
Jun 16 07:27:30 fv-az72-309 40grub2[3233]: debug: parsing: load_video
Jun 16 07:27:30 fv-az72-309 40grub2[3234]: debug: parsing: gfxmode $linux_gfx_mode
Jun 16 07:27:30 fv-az72-309 40grub2[3235]: debug: parsing: insmod gzio
Jun 16 07:27:30 fv-az72-309 40grub2[3236]: debug: parsing: if [ x$grub_platform = xxen ]; then insmod xzio; insmod lzopio; fi
Jun 16 07:27:30 fv-az72-309 40grub2[3237]: debug: parsing: insmod part_gpt
Jun 16 07:27:30 fv-az72-309 40grub2[3238]: debug: parsing: insmod ext2
Jun 16 07:27:30 fv-az72-309 40grub2[3239]: debug: parsing: if [ x$feature_platform_search_hint = xy ]; then
Jun 16 07:27:30 fv-az72-309 40grub2[3240]: debug: parsing: search --no-floppy --fs-uuid --set=root 3b66cf28-8c39-478b-82b0-294032b5bd9d
Jun 16 07:27:30 fv-az72-309 40grub2[3241]: debug: parsing: else
Jun 16 07:27:30 fv-az72-309 40grub2[3242]: debug: parsing: search --no-floppy --fs-uuid --set=root 3b66cf28-8c39-478b-82b0-294032b5bd9d
Jun 16 07:27:30 fv-az72-309 40grub2[3243]: debug: parsing: fi
Jun 16 07:27:30 fv-az72-309 40grub2[3244]: debug: parsing: echo 'Loading Linux 5.13.0-1029-azure ...'
Jun 16 07:27:30 fv-az72-309 40grub2[3245]: debug: parsing: if [ "${initrdfail}" = 1 ]; then
Jun 16 07:27:30 fv-az72-309 40grub2[3246]: debug: parsing: echo 'GRUB_FORCE_PARTUUID set, initrdless boot failed. Attempting with initrd.'
Jun 16 07:27:30 fv-az72-309 40grub2[3247]: debug: parsing: linux /boot/vmlinuz-5.13.0-1029-azure root=PARTUUID=2b1f5b8e-4041-4065-af1c-792f94a6d205 ro console=tty1 console=ttyS0 earlyprintk=ttyS0
Jun 16 07:27:30 fv-az72-309 40grub2[3251]: debug: parsing: echo 'Loading initial ramdisk ...'
Jun 16 07:27:30 fv-az72-309 40grub2[3252]: debug: parsing: initrd /boot/initrd.img-5.13.0-1029-azure
Jun 16 07:27:30 fv-az72-309 40grub2[3256]: debug: parsing: else
Jun 16 07:27:30 fv-az72-309 40grub2[3257]: debug: parsing: echo 'GRUB_FORCE_PARTUUID set, attempting initrdless boot.'
Jun 16 07:27:30 fv-az72-309 40grub2[3258]: debug: parsing: linux /boot/vmlinuz-5.13.0-1029-azure root=PARTUUID=2b1f5b8e-4041-4065-af1c-792f94a6d205 ro console=tty1 console=ttyS0 earlyprintk=ttyS0 panic=-1
Jun 16 07:27:30 fv-az72-309 40grub2[3262]: debug: parsing: fi
Jun 16 07:27:30 fv-az72-309 40grub2[3263]: debug: parsing: initrdfail
Jun 16 07:27:30 fv-az72-309 40grub2[3264]: debug: parsing: }
Jun 16 07:27:30 fv-az72-309 40grub2[3265]: result: /dev/sdb1:/dev/sdb1:Ubuntu, with Linux 5.13.0-1029-azure:/boot/vmlinuz-5.13.0-1029-azure:/boot/initrd.img-5.13.0-1029-azure:root=PARTUUID=2b1f5b8e-4041-4065-af1c-792f94a6d205 ro console=tty1 console=ttyS0 earlyprintk=ttyS0 panic=-1
Jun 16 07:27:30 fv-az72-309 40grub2[3266]: debug: parsing: menuentry 'Ubuntu, with Linux 5.13.0-1029-azure (recovery mode)' --class ubuntu --class gnu-linux --class gnu --class os $menuentry_id_option 'gnulinux-5.13.0-1029-azure-recovery-3b66cf28-8c39-478b-82b0-294032b5bd9d' {
Jun 16 07:27:30 fv-az72-309 40grub2[3277]: debug: parsing: recordfail
Jun 16 07:27:30 fv-az72-309 40grub2[3278]: debug: parsing: load_video
Jun 16 07:27:30 fv-az72-309 40grub2[3279]: debug: parsing: insmod gzio
Jun 16 07:27:30 fv-az72-309 40grub2[3280]: debug: parsing: if [ x$grub_platform = xxen ]; then insmod xzio; insmod lzopio; fi
Jun 16 07:27:30 fv-az72-309 40grub2[3281]: debug: parsing: insmod part_gpt
Jun 16 07:27:30 fv-az72-309 40grub2[3282]: debug: parsing: insmod ext2
Jun 16 07:27:30 fv-az72-309 40grub2[3283]: debug: parsing: if [ x$feature_platform_search_hint = xy ]; then
Jun 16 07:27:30 fv-az72-309 40grub2[3284]: debug: parsing: search --no-floppy --fs-uuid --set=root 3b66cf28-8c39-478b-82b0-294032b5bd9d
Jun 16 07:27:30 fv-az72-309 40grub2[3285]: debug: parsing: else
Jun 16 07:27:30 fv-az72-309 40grub2[3286]: debug: parsing: search --no-floppy --fs-uuid --set=root 3b66cf28-8c39-478b-82b0-294032b5bd9d
Jun 16 07:27:30 fv-az72-309 40grub2[3287]: debug: parsing: fi
Jun 16 07:27:30 fv-az72-309 40grub2[3288]: debug: parsing: echo 'Loading Linux 5.13.0-1029-azure ...'
Jun 16 07:27:30 fv-az72-309 40grub2[3289]: debug: parsing: if [ "${initrdfail}" = 1 ]; then
Jun 16 07:27:30 fv-az72-309 40grub2[3290]: debug: parsing: echo 'GRUB_FORCE_PARTUUID set, initrdless boot failed. Attempting with initrd.'
Jun 16 07:27:30 fv-az72-309 40grub2[3291]: debug: parsing: linux /boot/vmlinuz-5.13.0-1029-azure root=PARTUUID=2b1f5b8e-4041-4065-af1c-792f94a6d205 ro recovery nomodeset dis_ucode_ldr console=tty1 console=ttyS0 earlyprintk=ttyS0
Jun 16 07:27:30 fv-az72-309 40grub2[3295]: debug: parsing: echo 'Loading initial ramdisk ...'
Jun 16 07:27:30 fv-az72-309 40grub2[3296]: debug: parsing: initrd /boot/initrd.img-5.13.0-1029-azure
Jun 16 07:27:30 fv-az72-309 40grub2[3300]: debug: parsing: else
Jun 16 07:27:30 fv-az72-309 40grub2[3301]: debug: parsing: echo 'GRUB_FORCE_PARTUUID set, attempting initrdless boot.'
Jun 16 07:27:30 fv-az72-309 40grub2[3302]: debug: parsing: linux /boot/vmlinuz-5.13.0-1029-azure root=PARTUUID=2b1f5b8e-4041-4065-af1c-792f94a6d205 ro recovery nomodeset dis_ucode_ldr console=tty1 console=ttyS0 earlyprintk=ttyS0 panic=-1
Jun 16 07:27:30 fv-az72-309 40grub2[3306]: debug: parsing: fi
Jun 16 07:27:30 fv-az72-309 40grub2[3307]: debug: parsing: initrdfail
Jun 16 07:27:30 fv-az72-309 40grub2[3308]: debug: parsing: }
Jun 16 07:27:30 fv-az72-309 40grub2[3309]: result: /dev/sdb1:/dev/sdb1:Ubuntu, with Linux 5.13.0-1029-azure (recovery mode):/boot/vmlinuz-5.13.0-1029-azure:/boot/initrd.img-5.13.0-1029-azure:root=PARTUUID=2b1f5b8e-4041-4065-af1c-792f94a6d205 ro recovery nomodeset dis_ucode_ldr console=tty1 console=ttyS0 earlyprintk=ttyS0 panic=-1
Jun 16 07:27:30 fv-az72-309 40grub2[3310]: debug: parsing: }
Jun 16 07:27:30 fv-az72-309 40grub2[3311]: debug: parsing:
Jun 16 07:27:30 fv-az72-309 40grub2[3312]: debug: parsing: ### END /etc/grub.d/10_linux ###
Jun 16 07:27:30 fv-az72-309 40grub2[3313]: debug: parsing:
Jun 16 07:27:30 fv-az72-309 40grub2[3314]: debug: parsing: ### BEGIN /etc/grub.d/10_linux_zfs ###
Jun 16 07:27:30 fv-az72-309 40grub2[3315]: debug: parsing: ### END /etc/grub.d/10_linux_zfs ###
Jun 16 07:27:30 fv-az72-309 40grub2[3316]: debug: parsing:
Jun 16 07:27:30 fv-az72-309 40grub2[3317]: debug: parsing: ### BEGIN /etc/grub.d/20_linux_xen ###
Jun 16 07:27:30 fv-az72-309 40grub2[3318]: debug: parsing: ### END /etc/grub.d/20_linux_xen ###
Jun 16 07:27:30 fv-az72-309 40grub2[3319]: debug: parsing:
Jun 16 07:27:30 fv-az72-309 40grub2[3320]: debug: parsing: ### BEGIN /etc/grub.d/30_uefi-firmware ###
Jun 16 07:27:30 fv-az72-309 40grub2[3321]: debug: parsing: ### END /etc/grub.d/30_uefi-firmware ###
Jun 16 07:27:30 fv-az72-309 40grub2[3322]: debug: parsing:
Jun 16 07:27:30 fv-az72-309 40grub2[3323]: debug: parsing: ### BEGIN /etc/grub.d/35_fwupd ###
Jun 16 07:27:30 fv-az72-309 40grub2[3324]: debug: parsing: ### END /etc/grub.d/35_fwupd ###
Jun 16 07:27:30 fv-az72-309 40grub2[3325]: debug: parsing:
Jun 16 07:27:30 fv-az72-309 40grub2[3326]: debug: parsing: ### BEGIN /etc/grub.d/40_custom ###
Jun 16 07:27:30 fv-az72-309 40grub2[3327]: debug: parsing: # This file provides an easy way to add custom menu entries. Simply type the
Jun 16 07:27:30 fv-az72-309 40grub2[3328]: debug: parsing: # menu entries you want to add after this comment. Be careful not to change
Jun 16 07:27:30 fv-az72-309 40grub2[3329]: debug: parsing: # the 'exec tail' line above.
Jun 16 07:27:30 fv-az72-309 40grub2[3330]: debug: parsing: ### END /etc/grub.d/40_custom ###
Jun 16 07:27:30 fv-az72-309 40grub2[3331]: debug: parsing:
Jun 16 07:27:30 fv-az72-309 40grub2[3332]: debug: parsing: ### BEGIN /etc/grub.d/41_custom ###
Jun 16 07:27:30 fv-az72-309 40grub2[3333]: debug: parsing: if [ -f ${config_directory}/custom.cfg ]; then
Jun 16 07:27:30 fv-az72-309 40grub2[3334]: debug: parsing: source ${config_directory}/custom.cfg
Jun 16 07:27:30 fv-az72-309 40grub2[3335]: debug: parsing: elif [ -z "${config_directory}" -a -f $prefix/custom.cfg ]; then
Jun 16 07:27:30 fv-az72-309 40grub2[3336]: debug: parsing: source $prefix/custom.cfg;
Jun 16 07:27:30 fv-az72-309 40grub2[3337]: debug: parsing: fi
Jun 16 07:27:30 fv-az72-309 40grub2[3338]: debug: parsing: ### END /etc/grub.d/41_custom ###
Jun 16 07:27:30 fv-az72-309 50mounted-tests[3339]: debug: /usr/lib/linux-boot-probes/mounted/40grub2 succeeded
Jun 16 07:27:30 fv-az72-309 systemd[1]: var-lib-os\x2dprober-mount.mount: Succeeded.
Jun 16 07:27:30 fv-az72-309 linux-boot-prober[3343]: debug: linux detected by /usr/lib/linux-boot-probes/50mounted-tests
Jun 16 07:27:31 fv-az72-309 provisioner[1638]: [07:27:31] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:31 fv-az72-309 provisioner[1638]: done
Jun 16 07:27:31 fv-az72-309 provisioner[1638]: [07:27:31] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:31 fv-az72-309 provisioner[1638]: Writing provisioner version to imagedata.json
Jun 16 07:27:31 fv-az72-309 provisioner[1638]: [07:27:31] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:31 fv-az72-309 provisioner[1638]: The file /imagegeneration/imagedata.json has been updated with the provisioner version
Jun 16 07:27:31 fv-az72-309 provisioner[1638]: [07:27:31] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:31 fv-az72-309 provisioner[1638]: Copy /imagegeneration/imagedata.json file to the /home/runner/runners/2.292.0/.setup_info target
Jun 16 07:27:31 fv-az72-309 provisioner[1638]: [07:27:31] info: DistributedTask.MachineProvisioning.LinuxTaskAgentMachineRuntime[0]
Jun 16 07:27:31 fv-az72-309 provisioner[1638]: Copy /imagegeneration/imagedata.json file to the /home/runner/runners/2.293.0/.setup_info target
Jun 16 07:27:31 fv-az72-309 provisioner[1638]: [07:27:31] info: MachineManagement.Provisioning.Linux.LinuxMachineEnvironment[0]
Jun 16 07:27:31 fv-az72-309 provisioner[1638]: Running Factory Provisioning for DevFabric
Jun 16 07:27:31 fv-az72-309 provisioner[1638]: [07:27:31] info: MachineManagement.Provisioning.Linux.LinuxMachineEnvironment[0]
Jun 16 07:27:31 fv-az72-309 provisioner[1638]: Creating marker file to indicate machine is clean, will be deleted when first request is executed
Jun 16 07:27:31 fv-az72-309 provisioner[1638]: [07:27:31] info: MachineManagement.Provisioning.MachineProvisioner[0]
Jun 16 07:27:31 fv-az72-309 provisioner[1638]: Removing provisioning setting
Jun 16 07:27:31 fv-az72-309 provisioner[1638]: [07:27:31] info: MachineManagement.Provisioning.Linux.LinuxMachineEnvironment[0]
Jun 16 07:27:31 fv-az72-309 provisioner[1638]: Reading settings from settings file at /opt/runner/provisioner/.settings...
Jun 16 07:27:31 fv-az72-309 provisioner[1638]: [07:27:31] info: MachineManagement.Provisioning.Linux.LinuxMachineEnvironment[0]
Jun 16 07:27:31 fv-az72-309 provisioner[1638]: Settings does not contain the Provisioning key...
Jun 16 07:27:31 fv-az72-309 provisioner[1638]: [07:27:31] info: MachineManagement.Provisioning.Linux.LinuxMachineEnvironment[0]
Jun 16 07:27:31 fv-az72-309 provisioner[1638]: Waiting for AuthorizationTokenFile to be dropped by Custom Script Extension...
Jun 16 07:27:31 fv-az72-309 provisioner[1638]: [07:27:31] info: MachineManagement.Provisioning.Linux.LinuxMachineEnvironment[0]
Jun 16 07:27:31 fv-az72-309 provisioner[1638]: Last edit of access token: 06/16/2022 07:24:40
Jun 16 07:27:31 fv-az72-309 provisioner[1638]: [07:27:31] info: MachineManagement.Provisioning.MachineProvisioner[0]
Jun 16 07:27:31 fv-az72-309 provisioner[1638]: Provisioning successful
Jun 16 07:27:31 fv-az72-309 provisioner[1638]: [07:27:31] info: MachineManagement.Provisioning.MachineProvisioner[0]
Jun 16 07:27:31 fv-az72-309 provisioner[1638]: Finishing Provisioning Mode for fv-az72-309
Jun 16 07:27:31 fv-az72-309 provisioner[1638]: [07:27:31] info: Microsoft.AzureDevOps.Provisioner.Framework.Monitoring.MonitoringService[0]
Jun 16 07:27:31 fv-az72-309 provisioner[1638]: Canceling monitoring tasks
Jun 16 07:27:31 fv-az72-309 provisioner[1638]: [07:27:31] info: Microsoft.AzureDevOps.Provisioner.Framework.Monitoring.MonitoringService[0]
Jun 16 07:27:31 fv-az72-309 provisioner[1638]: All Monitor jobs are cancelled.
Jun 16 07:27:31 fv-az72-309 provisioner[1638]: [07:27:31] info: MachineManagement.Provisioning.MachineProvisioner[1001]
Jun 16 07:27:31 fv-az72-309 provisioner[1638]: The provisioner has stopped
Jun 16 07:27:31 fv-az72-309 systemd[1]: runner-provisioner.service: Succeeded.
Jun 16 07:27:36 fv-az72-309 python3[967]: 2022-06-16T07:27:36.649443Z INFO ExtHandler ExtHandler Fetching goal state [incarnation 3]
Jun 16 07:27:36 fv-az72-309 python3[967]: 2022-06-16T07:27:36.691835Z INFO ExtHandler ExtHandler Found private key matching thumbprint 0D28CC35EFA6D75FF8C5DEB4A1432FB1CD3E8FD9
Jun 16 07:27:36 fv-az72-309 python3[967]: 2022-06-16T07:27:36.723933Z INFO ExtHandler ExtHandler Fetch goal state completed
Jun 16 07:27:42 fv-az72-309 python3[967]: 2022-06-16T07:27:42.739094Z INFO ExtHandler ExtHandler Fetched new vmSettings [correlation ID: 4a2fa388-8e18-4dcf-84c9-2c893784f22a New eTag: 14864452798849322451]
Jun 16 07:27:42 fv-az72-309 python3[967]: 2022-06-16T07:27:42.741103Z INFO ExtHandler ExtHandler ProcessExtensionsInGoalState started [Incarnation: 3; Activity Id: 4bc43ecc-4bc2-4426-b3c0-5bb2577170f3; Correlation Id: 0801e8a7-1dfe-4a0f-985a-00367c0d4ea1; GS Creation Time: 2022-06-16T07:27:32.251274Z]
Jun 16 07:27:42 fv-az72-309 python3[967]: 2022-06-16T07:27:42.741637Z INFO ExtHandler ExtHandler No extension/run-time settings settings found for Microsoft.Azure.Extensions.CustomScript
Jun 16 07:27:42 fv-az72-309 python3[967]: 2022-06-16T07:27:42.771432Z INFO ExtHandler [Microsoft.Azure.Extensions.CustomScript-2.1.6] Target handler state: uninstall [incarnation 3]
Jun 16 07:27:42 fv-az72-309 python3[967]: 2022-06-16T07:27:42.771758Z INFO ExtHandler [Microsoft.Azure.Extensions.CustomScript-2.1.6] [Uninstall] current handler state is: enabled
Jun 16 07:27:42 fv-az72-309 python3[967]: 2022-06-16T07:27:42.772010Z INFO ExtHandler [Microsoft.Azure.Extensions.CustomScript-2.1.6] Disable extension: [bin/custom-script-shim disable]
Jun 16 07:27:42 fv-az72-309 python3[967]: 2022-06-16T07:27:42.772375Z INFO ExtHandler [Microsoft.Azure.Extensions.CustomScript-2.1.6] Executing command: /var/lib/waagent/Microsoft.Azure.Extensions.CustomScript-2.1.6/bin/custom-script-shim disable with environment variables: {"AZURE_GUEST_AGENT_EXTENSION_PATH": "/var/lib/waagent/Microsoft.Azure.Extensions.CustomScript-2.1.6", "AZURE_GUEST_AGENT_EXTENSION_VERSION": "2.1.6", "AZURE_GUEST_AGENT_WIRE_PROTOCOL_ADDRESS": "168.63.129.16", "ConfigSequenceNumber": "0", "AZURE_GUEST_AGENT_EXTENSION_SUPPORTED_FEATURES": "[{\"Key\": \"ExtensionTelemetryPipeline\", \"Value\": \"1.0\"}]"}
Jun 16 07:27:42 fv-az72-309 python3[967]: 2022-06-16T07:27:42.775840Z INFO ExtHandler ExtHandler Started extension in unit 'disable_a07754bb-7e8d-4fda-8f3f-ee686d3905a7.scope'
Jun 16 07:27:42 fv-az72-309 systemd[1]: Started /var/lib/waagent/Microsoft.Azure.Extensions.CustomScript-2.1.6/bin/custom-script-shim disable.
Jun 16 07:27:42 fv-az72-309 systemd[1]: disable_a07754bb-7e8d-4fda-8f3f-ee686d3905a7.scope: Succeeded.
Jun 16 07:27:44 fv-az72-309 python3[967]: 2022-06-16T07:27:44.779170Z INFO ExtHandler [Microsoft.Azure.Extensions.CustomScript-2.1.6] Command: bin/custom-script-shim disable
Jun 16 07:27:44 fv-az72-309 python3[967]: [stdout]
Jun 16 07:27:44 fv-az72-309 python3[967]: + /var/lib/waagent/Microsoft.Azure.Extensions.CustomScript-2.1.6/bin/custom-script-extension disable
Jun 16 07:27:44 fv-az72-309 python3[967]: time=2022-06-16T07:27:42Z version=v2.1.6/git@fc181d8-dirty operation=disable seq=0 event=start
Jun 16 07:27:44 fv-az72-309 python3[967]: time=2022-06-16T07:27:42Z version=v2.1.6/git@fc181d8-dirty operation=disable seq=0 event=noop
Jun 16 07:27:44 fv-az72-309 python3[967]: time=2022-06-16T07:27:42Z version=v2.1.6/git@fc181d8-dirty operation=disable seq=0 event=end
Jun 16 07:27:44 fv-az72-309 python3[967]: [stderr]
Jun 16 07:27:44 fv-az72-309 python3[967]: Running scope as unit: disable_a07754bb-7e8d-4fda-8f3f-ee686d3905a7.scope
Jun 16 07:27:44 fv-az72-309 python3[967]: 2022-06-16T07:27:44.780765Z INFO ExtHandler [Microsoft.Azure.Extensions.CustomScript-2.1.6] Uninstall extension [bin/custom-script-shim uninstall]
Jun 16 07:27:44 fv-az72-309 python3[967]: 2022-06-16T07:27:44.781186Z INFO ExtHandler [Microsoft.Azure.Extensions.CustomScript-2.1.6] Executing command: /var/lib/waagent/Microsoft.Azure.Extensions.CustomScript-2.1.6/bin/custom-script-shim uninstall with environment variables: {"AZURE_GUEST_AGENT_EXTENSION_PATH": "/var/lib/waagent/Microsoft.Azure.Extensions.CustomScript-2.1.6", "AZURE_GUEST_AGENT_EXTENSION_VERSION": "2.1.6", "AZURE_GUEST_AGENT_WIRE_PROTOCOL_ADDRESS": "168.63.129.16", "ConfigSequenceNumber": "0", "AZURE_GUEST_AGENT_EXTENSION_SUPPORTED_FEATURES": "[{\"Key\": \"ExtensionTelemetryPipeline\", \"Value\": \"1.0\"}]"}
Jun 16 07:27:44 fv-az72-309 python3[967]: 2022-06-16T07:27:44.785031Z INFO ExtHandler ExtHandler Started extension in unit 'uninstall_dd9159ab-fa53-499c-9e0b-98a49c82d09f.scope'
Jun 16 07:27:44 fv-az72-309 systemd[1]: Started /var/lib/waagent/Microsoft.Azure.Extensions.CustomScript-2.1.6/bin/custom-script-shim uninstall.
Jun 16 07:27:44 fv-az72-309 systemd[1]: uninstall_dd9159ab-fa53-499c-9e0b-98a49c82d09f.scope: Succeeded.
Jun 16 07:27:46 fv-az72-309 python3[967]: 2022-06-16T07:27:46.788551Z INFO ExtHandler [Microsoft.Azure.Extensions.CustomScript-2.1.6] Command: bin/custom-script-shim uninstall
Jun 16 07:27:46 fv-az72-309 python3[967]: [stdout]
Jun 16 07:27:46 fv-az72-309 python3[967]: + /var/lib/waagent/Microsoft.Azure.Extensions.CustomScript-2.1.6/bin/custom-script-extension uninstall
Jun 16 07:27:46 fv-az72-309 python3[967]: time=2022-06-16T07:27:44Z version=v2.1.6/git@fc181d8-dirty operation=uninstall seq=0 event=start
Jun 16 07:27:46 fv-az72-309 python3[967]: time=2022-06-16T07:27:44Z version=v2.1.6/git@fc181d8-dirty operation=uninstall seq=0 status="not reported for operation (by design)"
Jun 16 07:27:46 fv-az72-309 python3[967]: time=2022-06-16T07:27:44Z version=v2.1.6/git@fc181d8-dirty operation=uninstall seq=0 path=/var/lib/waagent/custom-script event="removing data dir" path=/var/lib/waagent/custom-script
Jun 16 07:27:46 fv-az72-309 python3[967]: time=2022-06-16T07:27:44Z version=v2.1.6/git@fc181d8-dirty operation=uninstall seq=0 path=/var/lib/waagent/custom-script event="removed data dir"
Jun 16 07:27:46 fv-az72-309 python3[967]: time=2022-06-16T07:27:44Z version=v2.1.6/git@fc181d8-dirty operation=uninstall seq=0 path=/var/lib/waagent/custom-script event=uninstalled
Jun 16 07:27:46 fv-az72-309 python3[967]: time=2022-06-16T07:27:44Z version=v2.1.6/git@fc181d8-dirty operation=uninstall seq=0 status="not reported for operation (by design)"
Jun 16 07:27:46 fv-az72-309 python3[967]: time=2022-06-16T07:27:44Z version=v2.1.6/git@fc181d8-dirty operation=uninstall seq=0 event=end
Jun 16 07:27:46 fv-az72-309 python3[967]: [stderr]
Jun 16 07:27:46 fv-az72-309 python3[967]: Running scope as unit: uninstall_dd9159ab-fa53-499c-9e0b-98a49c82d09f.scope
Jun 16 07:27:46 fv-az72-309 python3[967]: 2022-06-16T07:27:46.790365Z INFO ExtHandler [Microsoft.Azure.Extensions.CustomScript-2.1.6] Remove extension handler directory: /var/lib/waagent/Microsoft.Azure.Extensions.CustomScript-2.1.6
Jun 16 07:27:46 fv-az72-309 python3[967]: 2022-06-16T07:27:46.793715Z INFO ExtHandler [Microsoft.Azure.Extensions.CustomScript-2.1.6] Remove the extension slice: Microsoft.Azure.Extensions.CustomScript-2.1.6
Jun 16 07:27:46 fv-az72-309 python3[967]: 2022-06-16T07:27:46.794064Z INFO ExtHandler ExtHandler Stopped tracking cgroup Microsoft.Azure.Extensions.CustomScript-2.1.6 [/sys/fs/cgroup/cpu,cpuacct/azure.slice/azure-vmextensions.slice/azure-vmextensions-Microsoft.Azure.Extensions.CustomScript_2.1.6.slice]
Jun 16 07:27:46 fv-az72-309 python3[967]: 2022-06-16T07:27:46.795163Z INFO ExtHandler ExtHandler [CGI] Removed /lib/systemd/system/azure-vmextensions-Microsoft.Azure.Extensions.CustomScript_2.1.6.slice
Jun 16 07:27:46 fv-az72-309 python3[967]: 2022-06-16T07:27:46.796633Z INFO ExtHandler ExtHandler ProcessExtensionsInGoalState completed [Incarnation: 3; 4055 ms; Activity Id: 4bc43ecc-4bc2-4426-b3c0-5bb2577170f3; Correlation Id: 0801e8a7-1dfe-4a0f-985a-00367c0d4ea1; GS Creation Time: 2022-06-16T07:27:32.251274Z]
Jun 16 07:28:03 fv-az72-309 kernel: hv_utils: Shutdown request received - graceful shutdown initiated
Jun 16 07:28:03 fv-az72-309 systemd[1]: mnt.mount: Found ordering cycle on mnt-swapfile.swap/stop
Jun 16 07:28:03 fv-az72-309 systemd[1]: mnt.mount: Found dependency on swap.target/stop
Jun 16 07:28:03 fv-az72-309 systemd[1]: mnt.mount: Found dependency on run-snapd-ns.mount/stop
Jun 16 07:28:03 fv-az72-309 systemd[1]: mnt.mount: Found dependency on local-fs.target/stop
Jun 16 07:28:03 fv-az72-309 systemd[1]: mnt.mount: Found dependency on systemd-tmpfiles-setup.service/stop
Jun 16 07:28:03 fv-az72-309 systemd[1]: mnt.mount: Found dependency on systemd-resolved.service/stop
Jun 16 07:28:03 fv-az72-309 systemd[1]: mnt.mount: Found dependency on network.target/stop
Jun 16 07:28:03 fv-az72-309 systemd[1]: mnt.mount: Found dependency on network-online.target/stop
Jun 16 07:28:03 fv-az72-309 systemd[1]: mnt.mount: Found dependency on mnt.mount/stop
Jun 16 07:28:03 fv-az72-309 systemd[1]: mnt.mount: Job mnt-swapfile.swap/stop deleted to break ordering cycle starting with mnt.mount/stop
Jun 16 07:28:03 fv-az72-309 systemd[1]: Removed slice system-modprobe.slice.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Removed slice system-walinuxagent.extensions.slice.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped target Cloud-init target.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped target Graphical Interface.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped target Host and Network Name Lookups.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped target Timers.
Jun 16 07:28:03 fv-az72-309 systemd[1]: e2scrub_all.timer: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped Periodic ext4 Online Metadata Check for All Filesystems.
Jun 16 07:28:03 fv-az72-309 systemd[1]: fstrim.timer: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped Discard unused blocks once a week.
Jun 16 07:28:03 fv-az72-309 systemd[1]: logrotate.timer: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped Daily rotation of log files.
Jun 16 07:28:03 fv-az72-309 systemd[1]: man-db.timer: Succeeded.
Jun 16 07:28:03 fv-az72-309 ModemManager[728]: <info> caught signal, shutting down...
Jun 16 07:28:03 fv-az72-309 dockerd[1061]: time="2022-06-16T07:28:03.490671391Z" level=info msg="Processing signal 'terminated'"
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped Daily man-db regeneration.
Jun 16 07:28:03 fv-az72-309 rsyslogd[689]: [origin software="rsyslogd" swVersion="8.2001.0" x-pid="689" x-info="https://www.rsyslog.com"] exiting on signal 15.
Jun 16 07:28:03 fv-az72-309 snapd[695]: main.go:155: Exiting on terminated signal.
Jun 16 07:28:03 fv-az72-309 snapd[695]: overlord.go:504: Released state lock file
Jun 16 07:28:03 fv-az72-309 systemd[1]: motd-news.timer: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped Message of the Day.
Jun 16 07:28:03 fv-az72-309 systemd[1]: phpsessionclean.timer: Succeeded.
Jun 16 07:28:03 fv-az72-309 ModemManager[728]: <info> ModemManager is shut down
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped Clean PHP session files every 30 mins.
Jun 16 07:28:03 fv-az72-309 systemd[1]: systemd-tmpfiles-clean.timer: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped Daily Cleanup of Temporary Directories.
Jun 16 07:28:03 fv-az72-309 systemd[1]: ua-timer.timer: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped Ubuntu Advantage Timer for running repeated jobs.
Jun 16 07:28:03 fv-az72-309 systemd[1]: lvm2-lvmpolld.socket: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Closed LVM2 poll daemon socket.
Jun 16 07:28:03 fv-az72-309 systemd[1]: systemd-rfkill.socket: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Closed Load/Save RF Kill Switch Status /dev/rfkill Watch.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopping Accounts Service...
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopping Availability of block devices...
Jun 16 07:28:03 fv-az72-309 systemd[1]: cloud-final.service: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped Execute cloud user/final scripts.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped target Multi-User System.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped target Login Prompts.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopping Modem Manager...
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopping LSB: automatic crash report generation...
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopping Deferred execution scheduler...
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopping Regular background program processing daemon...
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopping D-Bus System Message Bus...
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopping Docker Application Container Engine...
Jun 16 07:28:03 fv-az72-309 systemd[1]: ephemeral-disk-warning.service: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped Write warning to Azure ephemeral disk.
Jun 16 07:28:03 fv-az72-309 systemd[1]: cloud-config.service: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped Apply the settings specified in cloud-config.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped target Cloud-config availability.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopping Create final runtime dir for shutdown pivot root...
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopping Getty on tty1...
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopping irqbalance daemon...
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopping LSB: Mono XSP4...
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopping Dispatcher daemon for systemd-networkd...
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopping PackageKit Daemon...
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopping The PHP 7.4 FastCGI Process Manager...
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopping The PHP 8.0 FastCGI Process Manager...
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopping The PHP 8.1 FastCGI Process Manager...
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopping System Logging Service...
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopping Serial Getty on ttyS0...
Jun 16 07:28:03 fv-az72-309 systemd[1]: snapd.seeded.service: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped Wait until snapd is fully seeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopping Snap Daemon...
Jun 16 07:28:03 fv-az72-309 systemd[1]: Condition check resulted in Ubuntu core (all-snaps) system shutdown helper setup service being skipped.
Jun 16 07:28:03 fv-az72-309 blkdeactivate[3616]: Deactivating block devices:
Jun 16 07:28:03 fv-az72-309 python3[706]: 2022-06-16T07:28:03.533246Z INFO Daemon Agent WALinuxAgent-2.2.46 forwarding signal 15 to WALinuxAgent-2.7.1.0
Jun 16 07:28:03 fv-az72-309 sshd[1163]: Received signal 15; terminating.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopping LSB: Fast standalone full-text SQL search engine...
Jun 16 07:28:03 fv-az72-309 sphinxsearch[3630]: To enable sphinxsearch, edit /etc/default/sphinxsearch and set START=yes
Jun 16 07:28:03 fv-az72-309 php-fpm7.4[685]: aster process (/etc/php/7.4/fpm/php-fpm.conf): DIGEST-MD5 common mech free
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopping OpenBSD Secure Shell server...
Jun 16 07:28:03 fv-az72-309 dockerd[1061]: time="2022-06-16T07:28:03.577094019Z" level=info msg="Daemon shutdown complete"
Jun 16 07:28:03 fv-az72-309 php-fpm8.1[687]: aster process (/etc/php/8.1/fpm/php-fpm.conf): DIGEST-MD5 common mech free
Jun 16 07:28:03 fv-az72-309 systemd[1]: sysstat.service: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped Resets System Activity Data Collector.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopping Login Service...
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopping Disk Manager...
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopping Azure Linux Agent...
Jun 16 07:28:03 fv-az72-309 systemd[1]: accounts-daemon.service: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped Accounts Service.
Jun 16 07:28:03 fv-az72-309 systemd[1]: cron.service: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped Regular background program processing daemon.
Jun 16 07:28:03 fv-az72-309 systemd[1]: irqbalance.service: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped irqbalance daemon.
Jun 16 07:28:03 fv-az72-309 systemd[1]: networkd-dispatcher.service: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped Dispatcher daemon for systemd-networkd.
Jun 16 07:28:03 fv-az72-309 systemd[1]: rsyslog.service: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped System Logging Service.
Jun 16 07:28:03 fv-az72-309 systemd[1]: snapd.service: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped Snap Daemon.
Jun 16 07:28:03 fv-az72-309 udisksd[705]: udisks daemon version 2.8.4 exiting
Jun 16 07:28:03 fv-az72-309 systemd[1]: walinuxagent.service: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped Azure Linux Agent.
Jun 16 07:28:03 fv-az72-309 systemd[1]: walinuxagent.service: Consumed 2.466s CPU time.
Jun 16 07:28:03 fv-az72-309 systemd[1]: atd.service: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped Deferred execution scheduler.
Jun 16 07:28:03 fv-az72-309 systemd[1]: ModemManager.service: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped Modem Manager.
Jun 16 07:28:03 fv-az72-309 systemd[1]: serial-getty@ttyS0.service: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped Serial Getty on ttyS0.
Jun 16 07:28:03 fv-az72-309 systemd[1]: getty@tty1.service: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped Getty on tty1.
Jun 16 07:28:03 fv-az72-309 systemd[1]: docker.service: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped Docker Application Container Engine.
Jun 16 07:28:03 fv-az72-309 systemd[1]: ssh.service: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped OpenBSD Secure Shell server.
Jun 16 07:28:03 fv-az72-309 php-fpm8.0[686]: aster process (/etc/php/8.0/fpm/php-fpm.conf): DIGEST-MD5 common mech free
Jun 16 07:28:03 fv-az72-309 systemd[1]: packagekit.service: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped PackageKit Daemon.
Jun 16 07:28:03 fv-az72-309 systemd[1]: blk-availability.service: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped Availability of block devices.
Jun 16 07:28:03 fv-az72-309 systemd[1]: dbus.service: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped D-Bus System Message Bus.
Jun 16 07:28:03 fv-az72-309 systemd[1]: sphinxsearch.service: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped LSB: Fast standalone full-text SQL search engine.
Jun 16 07:28:03 fv-az72-309 systemd[1]: php7.4-fpm.service: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped The PHP 7.4 FastCGI Process Manager.
Jun 16 07:28:03 fv-az72-309 systemd[1]: systemd-logind.service: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped Login Service.
Jun 16 07:28:03 fv-az72-309 systemd[1]: php8.0-fpm.service: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped The PHP 8.0 FastCGI Process Manager.
Jun 16 07:28:03 fv-az72-309 mono-xsp4[3623]: * Stopping XSP 4.0 WebServer mono-xsp4
Jun 16 07:28:03 fv-az72-309 systemd[1]: udisks2.service: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped Disk Manager.
Jun 16 07:28:03 fv-az72-309 systemd[1]: php8.1-fpm.service: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped The PHP 8.1 FastCGI Process Manager.
Jun 16 07:28:03 fv-az72-309 apport[3617]: * Stopping automatic crash report generation: apport
Jun 16 07:28:03 fv-az72-309 systemd[1]: Removed slice system-getty.slice.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Removed slice system-serial\x2dgetty.slice.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped target Network is Online.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped target System Time Synchronized.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped target System Time Set.
Jun 16 07:28:03 fv-az72-309 chronyd[676]: chronyd exiting
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopping chrony, an NTP client/server...
Jun 16 07:28:03 fv-az72-309 containerd[709]: time="2022-06-16T07:28:03.720961156Z" level=info msg="Stop CRI service"
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopping containerd container runtime...
Jun 16 07:28:03 fv-az72-309 systemd[1]: Condition check resulted in Show Plymouth Power Off Screen being skipped.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopping Authorization Manager...
Jun 16 07:28:03 fv-az72-309 systemd[1]: systemd-networkd-wait-online.service: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped Wait for Network to be Configured.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopping Permit User Sessions...
Jun 16 07:28:03 fv-az72-309 systemd[1]: containerd.service: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped containerd container runtime.
Jun 16 07:28:03 fv-az72-309 systemd[1]: systemd-user-sessions.service: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped Permit User Sessions.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped target User and Group Name Lookups.
Jun 16 07:28:03 fv-az72-309 systemd[1]: polkit.service: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped Authorization Manager.
Jun 16 07:28:03 fv-az72-309 systemd[1]: chrony.service: Succeeded.
Jun 16 07:28:04 fv-az72-309 apport[3617]: ...done.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped chrony, an NTP client/server.
Jun 16 07:28:03 fv-az72-309 systemd[1]: apport.service: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped LSB: automatic crash report generation.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped target Network.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopping Network Name Resolution...
Jun 16 07:28:03 fv-az72-309 systemd[1]: systemd-resolved.service: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped Network Name Resolution.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopping Network Service...
Jun 16 07:28:03 fv-az72-309 systemd-networkd[529]: eth0: DHCP lease lost
Jun 16 07:28:03 fv-az72-309 systemd[1]: systemd-networkd.service: Succeeded.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped Network Service.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Requested transaction contradicts existing jobs: Transaction for systemd-networkd.service/start is destructive (poweroff.target has 'start' job queued, but 'stop' is included in transaction).
Jun 16 07:28:03 fv-az72-309 systemd[1]: systemd-networkd.socket: Failed to queue service startup job (Maybe the service file is missing or not a non-template unit?): Transaction for systemd-networkd.service/start is destructive (poweroff.target has 'start' job queued, but 'stop' is included in transaction).
Jun 16 07:28:03 fv-az72-309 systemd[1]: systemd-networkd.socket: Failed with result 'resources'.
Jun 16 07:28:03 fv-az72-309 systemd[1]: Stopped target Network (Pre).
Jun 16 07:28:04 fv-az72-309 mono-xsp4[3623]: ...done.
Jun 16 07:28:04 fv-az72-309 systemd[1]: mono-xsp4.service: Succeeded.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Stopped LSB: Mono XSP4.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Stopped target Basic System.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Stopped target Paths.
Jun 16 07:28:04 fv-az72-309 systemd[1]: ua-license-check.path: Succeeded.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Stopped Trigger to poll for Ubuntu Pro licenses (Only enabled on GCP LTS non-pro).
Jun 16 07:28:04 fv-az72-309 systemd[1]: Stopped target Remote File Systems.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Stopped target Remote File Systems (Pre).
Jun 16 07:28:04 fv-az72-309 systemd[1]: Stopped target Slices.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Removed slice User and Session Slice.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Stopped target Sockets.
Jun 16 07:28:04 fv-az72-309 systemd[1]: cloud-init-hotplugd.socket: Succeeded.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Closed cloud-init hotplug hook socket.
Jun 16 07:28:04 fv-az72-309 systemd[1]: dbus.socket: Succeeded.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Closed D-Bus System Message Bus Socket.
Jun 16 07:28:04 fv-az72-309 systemd[1]: docker.socket: Succeeded.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Closed Docker Socket for the API.
Jun 16 07:28:04 fv-az72-309 systemd[1]: iscsid.socket: Succeeded.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Closed Open-iSCSI iscsid Socket.
Jun 16 07:28:04 fv-az72-309 systemd[1]: snap.lxd.daemon.unix.socket: Succeeded.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Closed Socket unix for snap application lxd.daemon.
Jun 16 07:28:04 fv-az72-309 systemd[1]: snapd.socket: Succeeded.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Closed Socket activation for snappy daemon.
Jun 16 07:28:04 fv-az72-309 systemd[1]: syslog.socket: Succeeded.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Closed Syslog Socket.
Jun 16 07:28:04 fv-az72-309 systemd[1]: uuidd.socket: Succeeded.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Closed UUID daemon activation socket.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Stopped target System Initialization.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Stopped target Local Encrypted Volumes.
Jun 16 07:28:04 fv-az72-309 systemd[1]: systemd-ask-password-console.path: Succeeded.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Stopped Dispatch Password Requests to Console Directory Watch.
Jun 16 07:28:04 fv-az72-309 systemd[1]: systemd-ask-password-wall.path: Succeeded.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Stopped Forward Password Requests to Wall Directory Watch.
Jun 16 07:28:04 fv-az72-309 systemd[1]: cloud-init-local.service: Succeeded.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Stopped Initial cloud-init job (pre-networking).
Jun 16 07:28:04 fv-az72-309 systemd[1]: Stopping Entropy daemon using the HAVEGE algorithm...
Jun 16 07:28:04 fv-az72-309 haveged[469]: haveged: Stopping due to signal 15
Jun 16 07:28:04 fv-az72-309 systemd[1]: Stopping Hyper-V KVP Protocol Daemon...
Jun 16 07:28:04 fv-az72-309 systemd[1]: systemd-sysctl.service: Succeeded.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Stopped Apply Kernel Variables.
Jun 16 07:28:04 fv-az72-309 systemd[1]: systemd-modules-load.service: Succeeded.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Stopped Load Kernel Modules.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Stopping Update UTMP about System Boot/Shutdown...
Jun 16 07:28:04 fv-az72-309 systemd[1]: haveged.service: Succeeded.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Stopped Entropy daemon using the HAVEGE algorithm.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Stopping Load/Save Random Seed...
Jun 16 07:28:04 fv-az72-309 systemd[1]: hv-kvp-daemon.service: Succeeded.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Stopped Hyper-V KVP Protocol Daemon.
Jun 16 07:28:04 fv-az72-309 systemd[1]: systemd-update-utmp.service: Succeeded.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Stopped Update UTMP about System Boot/Shutdown.
Jun 16 07:28:04 fv-az72-309 systemd[1]: systemd-tmpfiles-setup.service: Succeeded.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Stopped Create Volatile Files and Directories.
Jun 16 07:28:04 fv-az72-309 systemd[1]: systemd-random-seed.service: Succeeded.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Stopped Load/Save Random Seed.
Jun 16 07:28:04 fv-az72-309 systemd-tmpfiles[3814]: [/run/finalrd-libs.conf:9] Duplicate line for path "/run/initramfs/lib64", ignoring.
Jun 16 07:28:04 fv-az72-309 finalrd[3815]: run-parts: executing /usr/share/finalrd/mdadm.finalrd setup
Jun 16 07:28:04 fv-az72-309 finalrd[3815]: run-parts: executing /usr/share/finalrd/open-iscsi.finalrd setup
Jun 16 07:28:04 fv-az72-309 systemd[1]: finalrd.service: Succeeded.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Stopped Create final runtime dir for shutdown pivot root.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Stopped target Local File Systems.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Unmounting /boot/efi...
Jun 16 07:28:04 fv-az72-309 systemd[1]: Unmounting /run/snapd/ns/lxd.mnt...
Jun 16 07:28:04 fv-az72-309 systemd[1]: Unmounting Mount unit for core20, revision 1518...
Jun 16 07:28:04 fv-az72-309 systemd[1]: Unmounting Mount unit for lxd, revision 22753...
Jun 16 07:28:04 fv-az72-309 systemd[1]: Unmounting Mount unit for snapd, revision 16010...
Jun 16 07:28:04 fv-az72-309 systemd[1]: boot-efi.mount: Succeeded.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Unmounted /boot/efi.
Jun 16 07:28:04 fv-az72-309 systemd[1]: run-snapd-ns-lxd.mnt.mount: Succeeded.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Unmounted /run/snapd/ns/lxd.mnt.
Jun 16 07:28:04 fv-az72-309 systemd[1]: snap-core20-1518.mount: Succeeded.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Unmounted Mount unit for core20, revision 1518.
Jun 16 07:28:04 fv-az72-309 systemd[1]: snap-lxd-22753.mount: Succeeded.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Unmounted Mount unit for lxd, revision 22753.
Jun 16 07:28:04 fv-az72-309 systemd[1]: snap-snapd-16010.mount: Succeeded.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Unmounted Mount unit for snapd, revision 16010.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Unmounting /run/snapd/ns...
Jun 16 07:28:04 fv-az72-309 systemd[1]: systemd-fsck@dev-disk-by\x2duuid-7167\x2d9500.service: Succeeded.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Stopped File System Check on /dev/disk/by-uuid/7167-9500.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Removed slice system-systemd\x2dfsck.slice.
Jun 16 07:28:04 fv-az72-309 systemd[1]: run-snapd-ns.mount: Succeeded.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Unmounted /run/snapd/ns.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Stopped target Local File Systems (Pre).
Jun 16 07:28:04 fv-az72-309 systemd[1]: Stopped target Swap.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Reached target Unmount All Filesystems.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Stopping Monitoring of LVM2 mirrors, snapshots etc. using dmeventd or progress polling...
Jun 16 07:28:04 fv-az72-309 multipathd[397]: exit (signal)
Jun 16 07:28:04 fv-az72-309 multipathd[397]: --------shut down-------
Jun 16 07:28:04 fv-az72-309 systemd[1]: Stopping Device-Mapper Multipath Device Controller...
Jun 16 07:28:04 fv-az72-309 systemd[1]: systemd-tmpfiles-setup-dev.service: Succeeded.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Stopped Create Static Device Nodes in /dev.
Jun 16 07:28:04 fv-az72-309 systemd[1]: systemd-sysusers.service: Succeeded.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Stopped Create System Users.
Jun 16 07:28:04 fv-az72-309 systemd[1]: systemd-remount-fs.service: Succeeded.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Stopped Remount Root and Kernel File Systems.
Jun 16 07:28:04 fv-az72-309 systemd[1]: systemd-fsck-root.service: Succeeded.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Stopped File System Check on Root Device.
Jun 16 07:28:04 fv-az72-309 systemd[1]: multipathd.service: Succeeded.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Stopped Device-Mapper Multipath Device Controller.
Jun 16 07:28:04 fv-az72-309 systemd[1]: lvm2-monitor.service: Succeeded.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Stopped Monitoring of LVM2 mirrors, snapshots etc. using dmeventd or progress polling.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Reached target Shutdown.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Reached target Final Step.
Jun 16 07:28:04 fv-az72-309 systemd[1]: systemd-poweroff.service: Succeeded.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Finished Power-Off.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Reached target Power-Off.
Jun 16 07:28:04 fv-az72-309 systemd[1]: Shutting down.
Jun 16 07:28:04 fv-az72-309 systemd-shutdown[1]: Syncing filesystems and block devices.
Jun 16 07:28:05 fv-az72-309 systemd-shutdown[1]: Sending SIGTERM to remaining processes...
Jun 16 07:28:05 fv-az72-309 systemd-journald[189]: Journal stopped
-- Reboot --
Jun 20 00:09:31 fv-az72-309 kernel: Linux version 5.13.0-1029-azure (buildd@lcy02-amd64-051) (gcc (Ubuntu 9.4.0-1ubuntu1~20.04.1) 9.4.0, GNU ld (GNU Binutils for Ubuntu) 2.34) #34~20.04.1-Ubuntu SMP Thu Jun 9 12:37:07 UTC 2022 (Ubuntu 5.13.0-1029.34~20.04.1-azure 5.13.19)
Jun 20 00:09:31 fv-az72-309 kernel: Command line: BOOT_IMAGE=/boot/vmlinuz-5.13.0-1029-azure root=PARTUUID=2b1f5b8e-4041-4065-af1c-792f94a6d205 ro console=tty1 console=ttyS0 earlyprintk=ttyS0 panic=-1
Jun 20 00:09:31 fv-az72-309 kernel: KERNEL supported cpus:
Jun 20 00:09:31 fv-az72-309 kernel: Intel GenuineIntel
Jun 20 00:09:31 fv-az72-309 kernel: AMD AuthenticAMD
Jun 20 00:09:31 fv-az72-309 kernel: Hygon HygonGenuine
Jun 20 00:09:31 fv-az72-309 kernel: Centaur CentaurHauls
Jun 20 00:09:31 fv-az72-309 kernel: zhaoxin Shanghai
Jun 20 00:09:31 fv-az72-309 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers'
Jun 20 00:09:31 fv-az72-309 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers'
Jun 20 00:09:31 fv-az72-309 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers'
Jun 20 00:09:31 fv-az72-309 kernel: x86/fpu: Supporting XSAVE feature 0x008: 'MPX bounds registers'
Jun 20 00:09:31 fv-az72-309 kernel: x86/fpu: Supporting XSAVE feature 0x010: 'MPX CSR'
Jun 20 00:09:31 fv-az72-309 kernel: x86/fpu: Supporting XSAVE feature 0x020: 'AVX-512 opmask'
Jun 20 00:09:31 fv-az72-309 kernel: x86/fpu: Supporting XSAVE feature 0x040: 'AVX-512 Hi256'
Jun 20 00:09:31 fv-az72-309 kernel: x86/fpu: Supporting XSAVE feature 0x080: 'AVX-512 ZMM_Hi256'
Jun 20 00:09:31 fv-az72-309 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256
Jun 20 00:09:31 fv-az72-309 kernel: x86/fpu: xstate_offset[3]: 832, xstate_sizes[3]: 64
Jun 20 00:09:31 fv-az72-309 kernel: x86/fpu: xstate_offset[4]: 896, xstate_sizes[4]: 64
Jun 20 00:09:31 fv-az72-309 kernel: x86/fpu: xstate_offset[5]: 960, xstate_sizes[5]: 64
Jun 20 00:09:31 fv-az72-309 kernel: x86/fpu: xstate_offset[6]: 1024, xstate_sizes[6]: 512
Jun 20 00:09:31 fv-az72-309 kernel: x86/fpu: xstate_offset[7]: 1536, xstate_sizes[7]: 1024
Jun 20 00:09:31 fv-az72-309 kernel: x86/fpu: Enabled xstate features 0xff, context size is 2560 bytes, using 'compacted' format.
Jun 20 00:09:31 fv-az72-309 kernel: BIOS-provided physical RAM map:
Jun 20 00:09:31 fv-az72-309 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable
Jun 20 00:09:31 fv-az72-309 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved
Jun 20 00:09:31 fv-az72-309 kernel: BIOS-e820: [mem 0x00000000000e0000-0x00000000000fffff] reserved
Jun 20 00:09:31 fv-az72-309 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000003ffeffff] usable
Jun 20 00:09:31 fv-az72-309 kernel: BIOS-e820: [mem 0x000000003fff0000-0x000000003fffefff] ACPI data
Jun 20 00:09:31 fv-az72-309 kernel: BIOS-e820: [mem 0x000000003ffff000-0x000000003fffffff] ACPI NVS
Jun 20 00:09:31 fv-az72-309 kernel: BIOS-e820: [mem 0x0000000100000000-0x000000027fffffff] usable
Jun 20 00:09:31 fv-az72-309 kernel: printk: bootconsole [earlyser0] enabled
Jun 20 00:09:31 fv-az72-309 kernel: NX (Execute Disable) protection: active
Jun 20 00:09:31 fv-az72-309 kernel: SMBIOS 2.3 present.
Jun 20 00:09:31 fv-az72-309 kernel: DMI: Microsoft Corporation Virtual Machine/Virtual Machine, BIOS 090008 12/07/2018
Jun 20 00:09:31 fv-az72-309 kernel: Hypervisor detected: Microsoft Hyper-V
Jun 20 00:09:31 fv-az72-309 kernel: Hyper-V: privilege flags low 0x2e7f, high 0x3880b0, hints 0x60c2c, misc 0xed7b2
Jun 20 00:09:31 fv-az72-309 kernel: Hyper-V Host Build:18362-10.0-3-0.3446
Jun 20 00:09:31 fv-az72-309 kernel: Hyper-V: LAPIC Timer Frequency: 0xc3500
Jun 20 00:09:31 fv-az72-309 kernel: Hyper-V: Using hypercall for remote TLB flush
Jun 20 00:09:31 fv-az72-309 kernel: clocksource: hyperv_clocksource_tsc_page: mask: 0xffffffffffffffff max_cycles: 0x24e6a1710, max_idle_ns: 440795202120 ns
Jun 20 00:09:31 fv-az72-309 kernel: tsc: Marking TSC unstable due to running on Hyper-V
Jun 20 00:09:31 fv-az72-309 kernel: tsc: Detected 2095.196 MHz processor
Jun 20 00:09:31 fv-az72-309 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved
Jun 20 00:09:31 fv-az72-309 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable
Jun 20 00:09:31 fv-az72-309 kernel: last_pfn = 0x280000 max_arch_pfn = 0x400000000
Jun 20 00:09:31 fv-az72-309 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT
Jun 20 00:09:31 fv-az72-309 kernel: e820: update [mem 0x40000000-0xffffffff] usable ==> reserved
Jun 20 00:09:31 fv-az72-309 kernel: last_pfn = 0x3fff0 max_arch_pfn = 0x400000000
Jun 20 00:09:31 fv-az72-309 kernel: found SMP MP-table at [mem 0x000ff780-0x000ff78f]
Jun 20 00:09:31 fv-az72-309 kernel: Using GB pages for direct mapping
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: Early table checksum verification disabled
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: RSDP 0x00000000000F5C00 000014 (v00 ACPIAM)
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: RSDT 0x000000003FFF0000 000040 (v01 VRTUAL MICROSFT 12001807 MSFT 00000097)
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: FACP 0x000000003FFF0200 000081 (v02 VRTUAL MICROSFT 12001807 MSFT 00000097)
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: DSDT 0x000000003FFF1D24 003CD5 (v01 MSFTVM MSFTVM02 00000002 INTL 02002026)
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: FACS 0x000000003FFFF000 000040
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: WAET 0x000000003FFF1A80 000028 (v01 VRTUAL MICROSFT 12001807 MSFT 00000097)
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: SLIC 0x000000003FFF1AC0 000176 (v01 VRTUAL MICROSFT 12001807 MSFT 00000097)
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: OEM0 0x000000003FFF1CC0 000064 (v01 VRTUAL MICROSFT 12001807 MSFT 00000097)
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: SRAT 0x000000003FFF0800 000140 (v02 VRTUAL MICROSFT 00000001 MSFT 00000001)
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: APIC 0x000000003FFF0300 000062 (v01 VRTUAL MICROSFT 12001807 MSFT 00000097)
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: OEMB 0x000000003FFFF040 000064 (v01 VRTUAL MICROSFT 12001807 MSFT 00000097)
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: Reserving FACP table memory at [mem 0x3fff0200-0x3fff0280]
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: Reserving DSDT table memory at [mem 0x3fff1d24-0x3fff59f8]
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: Reserving FACS table memory at [mem 0x3ffff000-0x3ffff03f]
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: Reserving WAET table memory at [mem 0x3fff1a80-0x3fff1aa7]
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: Reserving SLIC table memory at [mem 0x3fff1ac0-0x3fff1c35]
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: Reserving OEM0 table memory at [mem 0x3fff1cc0-0x3fff1d23]
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: Reserving SRAT table memory at [mem 0x3fff0800-0x3fff093f]
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: Reserving APIC table memory at [mem 0x3fff0300-0x3fff0361]
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: Reserving OEMB table memory at [mem 0x3ffff040-0x3ffff0a3]
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: Local APIC address 0xfee00000
Jun 20 00:09:31 fv-az72-309 kernel: SRAT: PXM 0 -> APIC 0x00 -> Node 0
Jun 20 00:09:31 fv-az72-309 kernel: SRAT: PXM 0 -> APIC 0x01 -> Node 0
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x3fffffff] hotplug
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000-0x27fffffff] hotplug
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x280200000-0xfdfffffff] hotplug
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x1000000000-0xffffffffff] hotplug
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x10000200000-0x1ffffffffff] hotplug
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x20000200000-0x3ffffffffff] hotplug
Jun 20 00:09:31 fv-az72-309 kernel: NUMA: Node 0 [mem 0x00000000-0x3fffffff] + [mem 0x100000000-0x27fffffff] -> [mem 0x00000000-0x27fffffff]
Jun 20 00:09:31 fv-az72-309 kernel: NODE_DATA(0) allocated [mem 0x27ffd6000-0x27fffffff]
Jun 20 00:09:31 fv-az72-309 kernel: Zone ranges:
Jun 20 00:09:31 fv-az72-309 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff]
Jun 20 00:09:31 fv-az72-309 kernel: DMA32 [mem 0x0000000001000000-0x00000000ffffffff]
Jun 20 00:09:31 fv-az72-309 kernel: Normal [mem 0x0000000100000000-0x000000027fffffff]
Jun 20 00:09:31 fv-az72-309 kernel: Device empty
Jun 20 00:09:31 fv-az72-309 kernel: Movable zone start for each node
Jun 20 00:09:31 fv-az72-309 kernel: Early memory node ranges
Jun 20 00:09:31 fv-az72-309 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff]
Jun 20 00:09:31 fv-az72-309 kernel: node 0: [mem 0x0000000000100000-0x000000003ffeffff]
Jun 20 00:09:31 fv-az72-309 kernel: node 0: [mem 0x0000000100000000-0x000000027fffffff]
Jun 20 00:09:31 fv-az72-309 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000027fffffff]
Jun 20 00:09:31 fv-az72-309 kernel: On node 0 totalpages: 1834894
Jun 20 00:09:31 fv-az72-309 kernel: DMA zone: 64 pages used for memmap
Jun 20 00:09:31 fv-az72-309 kernel: DMA zone: 158 pages reserved
Jun 20 00:09:31 fv-az72-309 kernel: DMA zone: 3998 pages, LIFO batch:0
Jun 20 00:09:31 fv-az72-309 kernel: DMA32 zone: 4032 pages used for memmap
Jun 20 00:09:31 fv-az72-309 kernel: DMA32 zone: 258032 pages, LIFO batch:63
Jun 20 00:09:31 fv-az72-309 kernel: Normal zone: 24576 pages used for memmap
Jun 20 00:09:31 fv-az72-309 kernel: Normal zone: 1572864 pages, LIFO batch:63
Jun 20 00:09:31 fv-az72-309 kernel: On node 0, zone DMA: 1 pages in unavailable ranges
Jun 20 00:09:31 fv-az72-309 kernel: On node 0, zone DMA: 97 pages in unavailable ranges
Jun 20 00:09:31 fv-az72-309 kernel: On node 0, zone Normal: 16 pages in unavailable ranges
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: PM-Timer IO Port: 0x408
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: Local APIC address 0xfee00000
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1])
Jun 20 00:09:31 fv-az72-309 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl)
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level)
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: IRQ0 used by override.
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: IRQ9 used by override.
Jun 20 00:09:31 fv-az72-309 kernel: Using ACPI (MADT) for SMP configuration information
Jun 20 00:09:31 fv-az72-309 kernel: smpboot: Allowing 2 CPUs, 0 hotplug CPUs
Jun 20 00:09:31 fv-az72-309 kernel: PM: hibernation: Registered nosave memory: [mem 0x00000000-0x00000fff]
Jun 20 00:09:31 fv-az72-309 kernel: PM: hibernation: Registered nosave memory: [mem 0x0009f000-0x0009ffff]
Jun 20 00:09:31 fv-az72-309 kernel: PM: hibernation: Registered nosave memory: [mem 0x000a0000-0x000dffff]
Jun 20 00:09:31 fv-az72-309 kernel: PM: hibernation: Registered nosave memory: [mem 0x000e0000-0x000fffff]
Jun 20 00:09:31 fv-az72-309 kernel: PM: hibernation: Registered nosave memory: [mem 0x3fff0000-0x3fffefff]
Jun 20 00:09:31 fv-az72-309 kernel: PM: hibernation: Registered nosave memory: [mem 0x3ffff000-0x3fffffff]
Jun 20 00:09:31 fv-az72-309 kernel: PM: hibernation: Registered nosave memory: [mem 0x40000000-0xffffffff]
Jun 20 00:09:31 fv-az72-309 kernel: [mem 0x40000000-0xffffffff] available for PCI devices
Jun 20 00:09:31 fv-az72-309 kernel: Booting paravirtualized kernel on Hyper-V
Jun 20 00:09:31 fv-az72-309 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 7645519600211568 ns
Jun 20 00:09:31 fv-az72-309 kernel: setup_percpu: NR_CPUS:8192 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1
Jun 20 00:09:31 fv-az72-309 kernel: percpu: Embedded 63 pages/cpu s221184 r8192 d28672 u1048576
Jun 20 00:09:31 fv-az72-309 kernel: pcpu-alloc: s221184 r8192 d28672 u1048576 alloc=1*2097152
Jun 20 00:09:31 fv-az72-309 kernel: pcpu-alloc: [0] 0 1
Jun 20 00:09:31 fv-az72-309 kernel: Hyper-V: PV spinlocks enabled
Jun 20 00:09:31 fv-az72-309 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear)
Jun 20 00:09:31 fv-az72-309 kernel: Built 1 zonelists, mobility grouping on. Total pages: 1806064
Jun 20 00:09:31 fv-az72-309 kernel: Policy zone: Normal
Jun 20 00:09:31 fv-az72-309 kernel: Kernel command line: BOOT_IMAGE=/boot/vmlinuz-5.13.0-1029-azure root=PARTUUID=2b1f5b8e-4041-4065-af1c-792f94a6d205 ro console=tty1 console=ttyS0 earlyprintk=ttyS0 panic=-1
Jun 20 00:09:31 fv-az72-309 kernel: Dentry cache hash table entries: 1048576 (order: 11, 8388608 bytes, linear)
Jun 20 00:09:31 fv-az72-309 kernel: Inode-cache hash table entries: 524288 (order: 10, 4194304 bytes, linear)
Jun 20 00:09:31 fv-az72-309 kernel: mem auto-init: stack:off, heap alloc:on, heap free:off
Jun 20 00:09:31 fv-az72-309 kernel: Memory: 7105684K/7339576K available (14346K kernel code, 3432K rwdata, 9780K rodata, 2608K init, 6104K bss, 233632K reserved, 0K cma-reserved)
Jun 20 00:09:31 fv-az72-309 kernel: random: get_random_u64 called from __kmem_cache_create+0x2d/0x440 with crng_init=0
Jun 20 00:09:31 fv-az72-309 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1
Jun 20 00:09:31 fv-az72-309 kernel: Kernel/User page tables isolation: enabled
Jun 20 00:09:31 fv-az72-309 kernel: ftrace: allocating 46678 entries in 183 pages
Jun 20 00:09:31 fv-az72-309 kernel: ftrace: allocated 183 pages with 6 groups
Jun 20 00:09:31 fv-az72-309 kernel: rcu: Hierarchical RCU implementation.
Jun 20 00:09:31 fv-az72-309 kernel: rcu: RCU restricting CPUs from NR_CPUS=8192 to nr_cpu_ids=2.
Jun 20 00:09:31 fv-az72-309 kernel: Rude variant of Tasks RCU enabled.
Jun 20 00:09:31 fv-az72-309 kernel: Tracing variant of Tasks RCU enabled.
Jun 20 00:09:31 fv-az72-309 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 25 jiffies.
Jun 20 00:09:31 fv-az72-309 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2
Jun 20 00:09:31 fv-az72-309 kernel: NR_IRQS: 524544, nr_irqs: 440, preallocated irqs: 16
Jun 20 00:09:31 fv-az72-309 kernel: random: crng done (trusting CPU's manufacturer)
Jun 20 00:09:31 fv-az72-309 kernel: Console: colour VGA+ 80x25
Jun 20 00:09:31 fv-az72-309 kernel: printk: console [tty1] enabled
Jun 20 00:09:31 fv-az72-309 kernel: printk: console [ttyS0] enabled
Jun 20 00:09:31 fv-az72-309 kernel: printk: bootconsole [earlyser0] disabled
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: Core revision 20210331
Jun 20 00:09:31 fv-az72-309 kernel: APIC: Switch to symmetric I/O mode setup
Jun 20 00:09:31 fv-az72-309 kernel: Hyper-V: Using IPI hypercalls
Jun 20 00:09:31 fv-az72-309 kernel: Hyper-V: Using enlightened APIC (xapic mode)
Jun 20 00:09:31 fv-az72-309 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1
Jun 20 00:09:31 fv-az72-309 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 4190.39 BogoMIPS (lpj=8380784)
Jun 20 00:09:31 fv-az72-309 kernel: pid_max: default: 32768 minimum: 301
Jun 20 00:09:31 fv-az72-309 kernel: LSM: Security Framework initializing
Jun 20 00:09:31 fv-az72-309 kernel: Yama: becoming mindful.
Jun 20 00:09:31 fv-az72-309 kernel: AppArmor: AppArmor initialized
Jun 20 00:09:31 fv-az72-309 kernel: Mount-cache hash table entries: 16384 (order: 5, 131072 bytes, linear)
Jun 20 00:09:31 fv-az72-309 kernel: Mountpoint-cache hash table entries: 16384 (order: 5, 131072 bytes, linear)
Jun 20 00:09:31 fv-az72-309 kernel: Last level iTLB entries: 4KB 64, 2MB 8, 4MB 8
Jun 20 00:09:31 fv-az72-309 kernel: Last level dTLB entries: 4KB 64, 2MB 0, 4MB 0, 1GB 4
Jun 20 00:09:31 fv-az72-309 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization
Jun 20 00:09:31 fv-az72-309 kernel: Spectre V2 : Mitigation: Retpolines
Jun 20 00:09:31 fv-az72-309 kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch
Jun 20 00:09:31 fv-az72-309 kernel: Speculative Store Bypass: Vulnerable
Jun 20 00:09:31 fv-az72-309 kernel: TAA: Mitigation: Clear CPU buffers
Jun 20 00:09:31 fv-az72-309 kernel: MDS: Mitigation: Clear CPU buffers
Jun 20 00:09:31 fv-az72-309 kernel: Freeing SMP alternatives memory: 40K
Jun 20 00:09:31 fv-az72-309 kernel: smpboot: CPU0: Intel(R) Xeon(R) Platinum 8171M CPU @ 2.60GHz (family: 0x6, model: 0x55, stepping: 0x4)
Jun 20 00:09:31 fv-az72-309 kernel: Performance Events: unsupported p6 CPU model 85 no PMU driver, software events only.
Jun 20 00:09:31 fv-az72-309 kernel: rcu: Hierarchical SRCU implementation.
Jun 20 00:09:31 fv-az72-309 kernel: NMI watchdog: Perf NMI watchdog permanently disabled
Jun 20 00:09:31 fv-az72-309 kernel: smp: Bringing up secondary CPUs ...
Jun 20 00:09:31 fv-az72-309 kernel: x86: Booting SMP configuration:
Jun 20 00:09:31 fv-az72-309 kernel: .... node #0, CPUs: #1
Jun 20 00:09:31 fv-az72-309 kernel: smp: Brought up 1 node, 2 CPUs
Jun 20 00:09:31 fv-az72-309 kernel: smpboot: Max logical packages: 1
Jun 20 00:09:31 fv-az72-309 kernel: smpboot: Total of 2 processors activated (8380.78 BogoMIPS)
Jun 20 00:09:31 fv-az72-309 kernel: devtmpfs: initialized
Jun 20 00:09:31 fv-az72-309 kernel: x86/mm: Memory block size: 128MB
Jun 20 00:09:31 fv-az72-309 kernel: PM: Registering ACPI NVS region [mem 0x3ffff000-0x3fffffff] (4096 bytes)
Jun 20 00:09:31 fv-az72-309 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 7645041785100000 ns
Jun 20 00:09:31 fv-az72-309 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear)
Jun 20 00:09:31 fv-az72-309 kernel: pinctrl core: initialized pinctrl subsystem
Jun 20 00:09:31 fv-az72-309 kernel: PM: RTC time: 00:09:29, date: 2022-06-20
Jun 20 00:09:31 fv-az72-309 kernel: NET: Registered protocol family 16
Jun 20 00:09:31 fv-az72-309 kernel: DMA: preallocated 1024 KiB GFP_KERNEL pool for atomic allocations
Jun 20 00:09:31 fv-az72-309 kernel: DMA: preallocated 1024 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations
Jun 20 00:09:31 fv-az72-309 kernel: DMA: preallocated 1024 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations
Jun 20 00:09:31 fv-az72-309 kernel: audit: initializing netlink subsys (disabled)
Jun 20 00:09:31 fv-az72-309 kernel: audit: type=2000 audit(1655683769.152:1): state=initialized audit_enabled=0 res=1
Jun 20 00:09:31 fv-az72-309 kernel: thermal_sys: Registered thermal governor 'fair_share'
Jun 20 00:09:31 fv-az72-309 kernel: thermal_sys: Registered thermal governor 'bang_bang'
Jun 20 00:09:31 fv-az72-309 kernel: thermal_sys: Registered thermal governor 'step_wise'
Jun 20 00:09:31 fv-az72-309 kernel: thermal_sys: Registered thermal governor 'user_space'
Jun 20 00:09:31 fv-az72-309 kernel: thermal_sys: Registered thermal governor 'power_allocator'
Jun 20 00:09:31 fv-az72-309 kernel: EISA bus registered
Jun 20 00:09:31 fv-az72-309 kernel: cpuidle: using governor ladder
Jun 20 00:09:31 fv-az72-309 kernel: cpuidle: using governor menu
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: bus type PCI registered
Jun 20 00:09:31 fv-az72-309 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5
Jun 20 00:09:31 fv-az72-309 kernel: PCI: Using configuration type 1 for base access
Jun 20 00:09:31 fv-az72-309 kernel: Kprobes globally optimized
Jun 20 00:09:31 fv-az72-309 kernel: HugeTLB registered 1.00 GiB page size, pre-allocated 0 pages
Jun 20 00:09:31 fv-az72-309 kernel: HugeTLB registered 2.00 MiB page size, pre-allocated 0 pages
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: Added _OSI(Module Device)
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: Added _OSI(Processor Device)
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: Added _OSI(3.0 _SCP Extensions)
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: Added _OSI(Processor Aggregator Device)
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: Added _OSI(Linux-Dell-Video)
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: Added _OSI(Linux-Lenovo-NV-HDMI-Audio)
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: Added _OSI(Linux-HPI-Hybrid-Graphics)
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: Interpreter enabled
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: (supports S0 S5)
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: Using IOAPIC for interrupt routing
Jun 20 00:09:31 fv-az72-309 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: Enabled 1 GPEs in block 00 to 0F
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff])
Jun 20 00:09:31 fv-az72-309 kernel: acpi PNP0A03:00: _OSC: OS supports [ASPM ClockPM Segments MSI HPX-Type3]
Jun 20 00:09:31 fv-az72-309 kernel: acpi PNP0A03:00: fail to add MMCONFIG information, can't access extended PCI configuration space under this bridge.
Jun 20 00:09:31 fv-az72-309 kernel: PCI host bridge to bus 0000:00
Jun 20 00:09:31 fv-az72-309 kernel: pci_bus 0000:00: root bus resource [bus 00-ff]
Jun 20 00:09:31 fv-az72-309 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window]
Jun 20 00:09:31 fv-az72-309 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window]
Jun 20 00:09:31 fv-az72-309 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window]
Jun 20 00:09:31 fv-az72-309 kernel: pci_bus 0000:00: root bus resource [mem 0x40000000-0xfffbffff window]
Jun 20 00:09:31 fv-az72-309 kernel: pci_bus 0000:00: root bus resource [mem 0xfe0000000-0xfffffffff window]
Jun 20 00:09:31 fv-az72-309 kernel: pci 0000:00:00.0: [8086:7192] type 00 class 0x060000
Jun 20 00:09:31 fv-az72-309 kernel: pci 0000:00:07.0: [8086:7110] type 00 class 0x060100
Jun 20 00:09:31 fv-az72-309 kernel: pci 0000:00:07.1: [8086:7111] type 00 class 0x010180
Jun 20 00:09:31 fv-az72-309 kernel: pci 0000:00:07.1: reg 0x20: [io 0xffa0-0xffaf]
Jun 20 00:09:31 fv-az72-309 kernel: pci 0000:00:07.1: legacy IDE quirk: reg 0x10: [io 0x01f0-0x01f7]
Jun 20 00:09:31 fv-az72-309 kernel: pci 0000:00:07.1: legacy IDE quirk: reg 0x14: [io 0x03f6]
Jun 20 00:09:31 fv-az72-309 kernel: pci 0000:00:07.1: legacy IDE quirk: reg 0x18: [io 0x0170-0x0177]
Jun 20 00:09:31 fv-az72-309 kernel: pci 0000:00:07.1: legacy IDE quirk: reg 0x1c: [io 0x0376]
Jun 20 00:09:31 fv-az72-309 kernel: * Found PM-Timer Bug on the chipset. Due to workarounds for a bug,
* this clock source is slow. Consider trying other clock sources
Jun 20 00:09:31 fv-az72-309 kernel: pci 0000:00:07.3: acpi_pm_check_blacklist+0x0/0x20 took 11718 usecs
Jun 20 00:09:31 fv-az72-309 kernel: pci 0000:00:07.3: [8086:7113] type 00 class 0x068000
Jun 20 00:09:31 fv-az72-309 kernel: pci 0000:00:07.3: quirk: [io 0x0400-0x043f] claimed by PIIX4 ACPI
Jun 20 00:09:31 fv-az72-309 kernel: pci 0000:00:08.0: [1414:5353] type 00 class 0x030000
Jun 20 00:09:31 fv-az72-309 kernel: pci 0000:00:08.0: reg 0x10: [mem 0xf8000000-0xfbffffff]
Jun 20 00:09:31 fv-az72-309 kernel: pci 0000:00:08.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff]
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 11
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 0
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: PCI: Interrupt link LNKB disabled
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 0
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: PCI: Interrupt link LNKC disabled
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 0
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: PCI: Interrupt link LNKD disabled
Jun 20 00:09:31 fv-az72-309 kernel: iommu: Default domain type: Translated
Jun 20 00:09:31 fv-az72-309 kernel: SCSI subsystem initialized
Jun 20 00:09:31 fv-az72-309 kernel: libata version 3.00 loaded.
Jun 20 00:09:31 fv-az72-309 kernel: pci 0000:00:08.0: vgaarb: setting as boot VGA device
Jun 20 00:09:31 fv-az72-309 kernel: pci 0000:00:08.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none
Jun 20 00:09:31 fv-az72-309 kernel: pci 0000:00:08.0: vgaarb: bridge control possible
Jun 20 00:09:31 fv-az72-309 kernel: vgaarb: loaded
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: bus type USB registered
Jun 20 00:09:31 fv-az72-309 kernel: usbcore: registered new interface driver usbfs
Jun 20 00:09:31 fv-az72-309 kernel: usbcore: registered new interface driver hub
Jun 20 00:09:31 fv-az72-309 kernel: usbcore: registered new device driver usb
Jun 20 00:09:31 fv-az72-309 kernel: pps_core: LinuxPPS API ver. 1 registered
Jun 20 00:09:31 fv-az72-309 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti <giometti@linux.it>
Jun 20 00:09:31 fv-az72-309 kernel: PTP clock support registered
Jun 20 00:09:31 fv-az72-309 kernel: EDAC MC: Ver: 3.0.0
Jun 20 00:09:31 fv-az72-309 kernel: hv_vmbus: Vmbus version:4.0
Jun 20 00:09:31 fv-az72-309 kernel: hv_vmbus: Unknown GUID: c376c1c3-d276-48d2-90a9-c04748072c60
Jun 20 00:09:31 fv-az72-309 kernel: NetLabel: Initializing
Jun 20 00:09:31 fv-az72-309 kernel: NetLabel: domain hash size = 128
Jun 20 00:09:31 fv-az72-309 kernel: NetLabel: protocols = UNLABELED CIPSOv4 CALIPSO
Jun 20 00:09:31 fv-az72-309 kernel: NetLabel: unlabeled traffic allowed by default
Jun 20 00:09:31 fv-az72-309 kernel: PCI: Using ACPI for IRQ routing
Jun 20 00:09:31 fv-az72-309 kernel: PCI: pci_cache_line_size set to 64 bytes
Jun 20 00:09:31 fv-az72-309 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff]
Jun 20 00:09:31 fv-az72-309 kernel: e820: reserve RAM buffer [mem 0x3fff0000-0x3fffffff]
Jun 20 00:09:31 fv-az72-309 kernel: clocksource: Switched to clocksource hyperv_clocksource_tsc_page
Jun 20 00:09:31 fv-az72-309 kernel: VFS: Disk quotas dquot_6.6.0
Jun 20 00:09:31 fv-az72-309 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes)
Jun 20 00:09:31 fv-az72-309 kernel: AppArmor: AppArmor Filesystem Enabled
Jun 20 00:09:31 fv-az72-309 kernel: pnp: PnP ACPI init
Jun 20 00:09:31 fv-az72-309 kernel: pnp 00:00: Plug and Play ACPI device, IDs PNP0b00 (active)
Jun 20 00:09:31 fv-az72-309 kernel: pnp 00:01: Plug and Play ACPI device, IDs PNP0303 PNP030b (active)
Jun 20 00:09:31 fv-az72-309 kernel: pnp 00:02: Plug and Play ACPI device, IDs PNP0f03 PNP0f13 (active)
Jun 20 00:09:31 fv-az72-309 kernel: pnp 00:03: [dma 0 disabled]
Jun 20 00:09:31 fv-az72-309 kernel: pnp 00:03: Plug and Play ACPI device, IDs PNP0501 (active)
Jun 20 00:09:31 fv-az72-309 kernel: pnp 00:04: [dma 0 disabled]
Jun 20 00:09:31 fv-az72-309 kernel: pnp 00:04: Plug and Play ACPI device, IDs PNP0501 (active)
Jun 20 00:09:31 fv-az72-309 kernel: pnp 00:05: [dma 2]
Jun 20 00:09:31 fv-az72-309 kernel: pnp 00:05: Plug and Play ACPI device, IDs PNP0700 (active)
Jun 20 00:09:31 fv-az72-309 kernel: system 00:06: [io 0x01e0-0x01ef] has been reserved
Jun 20 00:09:31 fv-az72-309 kernel: system 00:06: [io 0x0160-0x016f] has been reserved
Jun 20 00:09:31 fv-az72-309 kernel: system 00:06: [io 0x0278-0x027f] has been reserved
Jun 20 00:09:31 fv-az72-309 kernel: system 00:06: [io 0x0378-0x037f] has been reserved
Jun 20 00:09:31 fv-az72-309 kernel: system 00:06: [io 0x0678-0x067f] has been reserved
Jun 20 00:09:31 fv-az72-309 kernel: system 00:06: [io 0x0778-0x077f] has been reserved
Jun 20 00:09:31 fv-az72-309 kernel: system 00:06: [io 0x04d0-0x04d1] has been reserved
Jun 20 00:09:31 fv-az72-309 kernel: system 00:06: Plug and Play ACPI device, IDs PNP0c02 (active)
Jun 20 00:09:31 fv-az72-309 kernel: system 00:07: [io 0x0400-0x043f] has been reserved
Jun 20 00:09:31 fv-az72-309 kernel: system 00:07: [io 0x0370-0x0371] has been reserved
Jun 20 00:09:31 fv-az72-309 kernel: system 00:07: [io 0x0440-0x044f] has been reserved
Jun 20 00:09:31 fv-az72-309 kernel: system 00:07: [mem 0xfec00000-0xfec00fff] could not be reserved
Jun 20 00:09:31 fv-az72-309 kernel: system 00:07: [mem 0xfee00000-0xfee00fff] has been reserved
Jun 20 00:09:31 fv-az72-309 kernel: system 00:07: Plug and Play ACPI device, IDs PNP0c02 (active)
Jun 20 00:09:31 fv-az72-309 kernel: system 00:08: [mem 0x00000000-0x0009ffff] could not be reserved
Jun 20 00:09:31 fv-az72-309 kernel: system 00:08: [mem 0x000c0000-0x000dffff] could not be reserved
Jun 20 00:09:31 fv-az72-309 kernel: system 00:08: [mem 0x000e0000-0x000fffff] could not be reserved
Jun 20 00:09:31 fv-az72-309 kernel: system 00:08: [mem 0x00100000-0x3fffffff] could not be reserved
Jun 20 00:09:31 fv-az72-309 kernel: system 00:08: [mem 0xfffc0000-0xffffffff] has been reserved
Jun 20 00:09:31 fv-az72-309 kernel: system 00:08: Plug and Play ACPI device, IDs PNP0c01 (active)
Jun 20 00:09:31 fv-az72-309 kernel: pnp: PnP ACPI: found 9 devices
Jun 20 00:09:31 fv-az72-309 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns
Jun 20 00:09:31 fv-az72-309 kernel: NET: Registered protocol family 2
Jun 20 00:09:31 fv-az72-309 kernel: IP idents hash table entries: 131072 (order: 8, 1048576 bytes, linear)
Jun 20 00:09:31 fv-az72-309 kernel: tcp_listen_portaddr_hash hash table entries: 4096 (order: 4, 65536 bytes, linear)
Jun 20 00:09:31 fv-az72-309 kernel: TCP established hash table entries: 65536 (order: 7, 524288 bytes, linear)
Jun 20 00:09:31 fv-az72-309 kernel: TCP bind hash table entries: 65536 (order: 8, 1048576 bytes, linear)
Jun 20 00:09:31 fv-az72-309 kernel: TCP: Hash tables configured (established 65536 bind 65536)
Jun 20 00:09:31 fv-az72-309 kernel: MPTCP token hash table entries: 8192 (order: 5, 196608 bytes, linear)
Jun 20 00:09:31 fv-az72-309 kernel: UDP hash table entries: 4096 (order: 5, 131072 bytes, linear)
Jun 20 00:09:31 fv-az72-309 kernel: UDP-Lite hash table entries: 4096 (order: 5, 131072 bytes, linear)
Jun 20 00:09:31 fv-az72-309 kernel: NET: Registered protocol family 1
Jun 20 00:09:31 fv-az72-309 kernel: NET: Registered protocol family 44
Jun 20 00:09:31 fv-az72-309 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window]
Jun 20 00:09:31 fv-az72-309 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window]
Jun 20 00:09:31 fv-az72-309 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window]
Jun 20 00:09:31 fv-az72-309 kernel: pci_bus 0000:00: resource 7 [mem 0x40000000-0xfffbffff window]
Jun 20 00:09:31 fv-az72-309 kernel: pci_bus 0000:00: resource 8 [mem 0xfe0000000-0xfffffffff window]
Jun 20 00:09:31 fv-az72-309 kernel: pci 0000:00:00.0: Limiting direct PCI/PCI transfers
Jun 20 00:09:31 fv-az72-309 kernel: PCI: CLS 0 bytes, default 64
Jun 20 00:09:31 fv-az72-309 kernel: PCI-DMA: Using software bounce buffering for IO (SWIOTLB)
Jun 20 00:09:31 fv-az72-309 kernel: software IO TLB: mapped [mem 0x000000003bff0000-0x000000003fff0000] (64MB)
Jun 20 00:09:31 fv-az72-309 kernel: Initialise system trusted keyrings
Jun 20 00:09:31 fv-az72-309 kernel: Key type blacklist registered
Jun 20 00:09:31 fv-az72-309 kernel: workingset: timestamp_bits=36 max_order=21 bucket_order=0
Jun 20 00:09:31 fv-az72-309 kernel: zbud: loaded
Jun 20 00:09:31 fv-az72-309 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher
Jun 20 00:09:31 fv-az72-309 kernel: fuse: init (API version 7.34)
Jun 20 00:09:31 fv-az72-309 kernel: integrity: Platform Keyring initialized
Jun 20 00:09:31 fv-az72-309 kernel: Key type asymmetric registered
Jun 20 00:09:31 fv-az72-309 kernel: Asymmetric key parser 'x509' registered
Jun 20 00:09:31 fv-az72-309 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 244)
Jun 20 00:09:31 fv-az72-309 kernel: io scheduler mq-deadline registered
Jun 20 00:09:31 fv-az72-309 kernel: shpchp: Standard Hot Plug PCI Controller Driver version: 0.4
Jun 20 00:09:31 fv-az72-309 kernel: hv_vmbus: registering driver hv_pci
Jun 20 00:09:31 fv-az72-309 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input0
Jun 20 00:09:31 fv-az72-309 kernel: ACPI: button: Power Button [PWRF]
Jun 20 00:09:31 fv-az72-309 kernel: Serial: 8250/16550 driver, 32 ports, IRQ sharing enabled
Jun 20 00:09:31 fv-az72-309 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A
Jun 20 00:09:31 fv-az72-309 kernel: 00:04: ttyS1 at I/O 0x2f8 (irq = 3, base_baud = 115200) is a 16550A
Jun 20 00:09:31 fv-az72-309 kernel: Linux agpgart interface v0.103
Jun 20 00:09:31 fv-az72-309 kernel: loop: module loaded
Jun 20 00:09:31 fv-az72-309 kernel: hv_vmbus: registering driver hv_storvsc
Jun 20 00:09:31 fv-az72-309 kernel: ata_piix 0000:00:07.1: version 2.13
Jun 20 00:09:31 fv-az72-309 kernel: ata_piix 0000:00:07.1: Hyper-V Virtual Machine detected, ATA device ignore set
Jun 20 00:09:31 fv-az72-309 kernel: scsi host0: storvsc_host_t
Jun 20 00:09:31 fv-az72-309 kernel: scsi host3: storvsc_host_t
Jun 20 00:09:31 fv-az72-309 kernel: scsi host4: ata_piix
Jun 20 00:09:31 fv-az72-309 kernel: scsi host2: storvsc_host_t
Jun 20 00:09:31 fv-az72-309 kernel: scsi 0:0:0:0: Direct-Access Msft Virtual Disk 1.0 PQ: 0 ANSI: 5
Jun 20 00:09:31 fv-az72-309 kernel: scsi host1: storvsc_host_t
Jun 20 00:09:31 fv-az72-309 kernel: scsi host5: ata_piix
Jun 20 00:09:31 fv-az72-309 kernel: scsi: waiting for bus probes to complete ...
Jun 20 00:09:31 fv-az72-309 kernel: ata1: PATA max UDMA/33 cmd 0x1f0 ctl 0x3f6 bmdma 0xffa0 irq 14
Jun 20 00:09:31 fv-az72-309 kernel: ata2: PATA max UDMA/33 cmd 0x170 ctl 0x376 bmdma 0xffa8 irq 15
Jun 20 00:09:31 fv-az72-309 kernel: sd 0:0:0:0: Attached scsi generic sg0 type 0
Jun 20 00:09:31 fv-az72-309 kernel: tun: Universal TUN/TAP device driver, 1.6
Jun 20 00:09:31 fv-az72-309 kernel: PPP generic driver version 2.4.2
Jun 20 00:09:31 fv-az72-309 kernel: sd 0:0:0:0: [sda] 180355072 512-byte logical blocks: (92.3 GB/86.0 GiB)
Jun 20 00:09:31 fv-az72-309 kernel: VFIO - User Level meta-driver version: 0.3
Jun 20 00:09:31 fv-az72-309 kernel: sd 0:0:0:0: [sda] 4096-byte physical blocks
Jun 20 00:09:31 fv-az72-309 kernel: i8042: PNP: PS/2 Controller [PNP0303:PS2K,PNP0f03:PS2M] at 0x60,0x64 irq 1,12
Jun 20 00:09:31 fv-az72-309 kernel: sd 0:0:0:0: [sda] Write Protect is off
Jun 20 00:09:31 fv-az72-309 kernel: sd 0:0:0:0: [sda] Mode Sense: 0f 00 10 00
Jun 20 00:09:31 fv-az72-309 kernel: sd 0:0:0:0: [sda] Write cache: disabled, read cache: enabled, supports DPO and FUA
Jun 20 00:09:31 fv-az72-309 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1
Jun 20 00:09:31 fv-az72-309 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12
Jun 20 00:09:31 fv-az72-309 kernel: scsi 1:0:1:0: Direct-Access Msft Virtual Disk 1.0 PQ: 0 ANSI: 5
Jun 20 00:09:31 fv-az72-309 kernel: mousedev: PS/2 mouse device common for all mice
Jun 20 00:09:31 fv-az72-309 kernel: sda: sda1 sda14 sda15
Jun 20 00:09:31 fv-az72-309 kernel: rtc_cmos 00:00: RTC can wake from S4
Jun 20 00:09:31 fv-az72-309 kernel: sd 1:0:1:0: Attached scsi generic sg1 type 0
Jun 20 00:09:31 fv-az72-309 kernel: rtc_cmos 00:00: registered as rtc0
Jun 20 00:09:31 fv-az72-309 kernel: rtc_cmos 00:00: setting system clock to 2022-06-20T00:09:30 UTC (1655683770)
Jun 20 00:09:31 fv-az72-309 kernel: sd 1:0:1:0: [sdb] 29360128 512-byte logical blocks: (15.0 GB/14.0 GiB)
Jun 20 00:09:31 fv-az72-309 kernel: rtc_cmos 00:00: alarms up to one month, 114 bytes nvram
Jun 20 00:09:31 fv-az72-309 kernel: device-mapper: uevent: version 1.0.3
Jun 20 00:09:31 fv-az72-309 kernel: sd 1:0:1:0: [sdb] Write Protect is off
Jun 20 00:09:31 fv-az72-309 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input1
Jun 20 00:09:31 fv-az72-309 kernel: sd 1:0:1:0: [sdb] Mode Sense: 0f 00 10 00
Jun 20 00:09:31 fv-az72-309 kernel: device-mapper: ioctl: 4.45.0-ioctl (2021-03-22) initialised: dm-devel@redhat.com
Jun 20 00:09:31 fv-az72-309 kernel: sd 0:0:0:0: [sda] Attached SCSI disk
Jun 20 00:09:31 fv-az72-309 kernel: platform eisa.0: Probing EISA bus 0
Jun 20 00:09:31 fv-az72-309 kernel: platform eisa.0: EISA: Cannot allocate resource for mainboard
Jun 20 00:09:31 fv-az72-309 kernel: sd 1:0:1:0: [sdb] Write cache: disabled, read cache: enabled, supports DPO and FUA
Jun 20 00:09:31 fv-az72-309 kernel: platform eisa.0: Cannot allocate resource for EISA slot 1
Jun 20 00:09:31 fv-az72-309 kernel: platform eisa.0: Cannot allocate resource for EISA slot 2
Jun 20 00:09:31 fv-az72-309 kernel: platform eisa.0: Cannot allocate resource for EISA slot 3
Jun 20 00:09:31 fv-az72-309 kernel: platform eisa.0: Cannot allocate resource for EISA slot 4
Jun 20 00:09:31 fv-az72-309 kernel: platform eisa.0: Cannot allocate resource for EISA slot 5
Jun 20 00:09:31 fv-az72-309 kernel: platform eisa.0: Cannot allocate resource for EISA slot 6
Jun 20 00:09:31 fv-az72-309 kernel: platform eisa.0: Cannot allocate resource for EISA slot 7
Jun 20 00:09:31 fv-az72-309 kernel: ata1.01: host indicates ignore ATA devices, ignored
Jun 20 00:09:31 fv-az72-309 kernel: platform eisa.0: Cannot allocate resource for EISA slot 8
Jun 20 00:09:31 fv-az72-309 kernel: platform eisa.0: EISA: Detected 0 cards
Jun 20 00:09:31 fv-az72-309 kernel: intel_pstate: CPU model not supported
Jun 20 00:09:31 fv-az72-309 kernel: drop_monitor: Initializing network drop monitor service
Jun 20 00:09:31 fv-az72-309 kernel: ata1.00: host indicates ignore ATA devices, ignored
Jun 20 00:09:31 fv-az72-309 kernel: NET: Registered protocol family 10
Jun 20 00:09:31 fv-az72-309 kernel: Segment Routing with IPv6
Jun 20 00:09:31 fv-az72-309 kernel: sdb: sdb1
Jun 20 00:09:31 fv-az72-309 kernel: NET: Registered protocol family 17
Jun 20 00:09:31 fv-az72-309 kernel: Key type dns_resolver registered
Jun 20 00:09:31 fv-az72-309 kernel: No MBM correction factor available
Jun 20 00:09:31 fv-az72-309 kernel: IPI shorthand broadcast: enabled
Jun 20 00:09:31 fv-az72-309 kernel: sched_clock: Marking stable (1918415800, 133997700)->(2135540900, -83127400)
Jun 20 00:09:31 fv-az72-309 kernel: registered taskstats version 1
Jun 20 00:09:31 fv-az72-309 kernel: Loading compiled-in X.509 certificates
Jun 20 00:09:31 fv-az72-309 kernel: Loaded X.509 cert 'Build time autogenerated kernel key: 07f5640c3d7bf043074dc27d0b5799302e473486'
Jun 20 00:09:31 fv-az72-309 kernel: Loaded X.509 cert 'Canonical Ltd. Live Patch Signing: 14df34d1a87cf37625abec039ef2bf521249b969'
Jun 20 00:09:31 fv-az72-309 kernel: Loaded X.509 cert 'Canonical Ltd. Kernel Module Signing: 88f752e560a1e0737e31163a466ad7b70a850c19'
Jun 20 00:09:31 fv-az72-309 kernel: blacklist: Loading compiled-in revocation X.509 certificates
Jun 20 00:09:31 fv-az72-309 kernel: Loaded X.509 cert 'Canonical Ltd. Secure Boot Signing: 61482aa2830d0ab2ad5af10b7250da9033ddcef0'
Jun 20 00:09:31 fv-az72-309 kernel: zswap: loaded using pool lzo/zbud
Jun 20 00:09:31 fv-az72-309 kernel: Key type ._fscrypt registered
Jun 20 00:09:31 fv-az72-309 kernel: Key type .fscrypt registered
Jun 20 00:09:31 fv-az72-309 kernel: Key type fscrypt-provisioning registered
Jun 20 00:09:31 fv-az72-309 kernel: sd 1:0:1:0: [sdb] Attached SCSI disk
Jun 20 00:09:31 fv-az72-309 kernel: Key type encrypted registered
Jun 20 00:09:31 fv-az72-309 kernel: AppArmor: AppArmor sha1 policy hashing enabled
Jun 20 00:09:31 fv-az72-309 kernel: ima: No TPM chip found, activating TPM-bypass!
Jun 20 00:09:31 fv-az72-309 kernel: Loading compiled-in module X.509 certificates
Jun 20 00:09:31 fv-az72-309 kernel: Loaded X.509 cert 'Build time autogenerated kernel key: 07f5640c3d7bf043074dc27d0b5799302e473486'
Jun 20 00:09:31 fv-az72-309 kernel: ima: Allocated hash algorithm: sha1
Jun 20 00:09:31 fv-az72-309 kernel: ima: No architecture policies found
Jun 20 00:09:31 fv-az72-309 kernel: evm: Initialising EVM extended attributes:
Jun 20 00:09:31 fv-az72-309 kernel: evm: security.selinux
Jun 20 00:09:31 fv-az72-309 kernel: evm: security.SMACK64
Jun 20 00:09:31 fv-az72-309 kernel: evm: security.SMACK64EXEC
Jun 20 00:09:31 fv-az72-309 kernel: evm: security.SMACK64TRANSMUTE
Jun 20 00:09:31 fv-az72-309 kernel: evm: security.SMACK64MMAP
Jun 20 00:09:31 fv-az72-309 kernel: evm: security.apparmor
Jun 20 00:09:31 fv-az72-309 kernel: evm: security.ima
Jun 20 00:09:31 fv-az72-309 kernel: evm: security.capability
Jun 20 00:09:31 fv-az72-309 kernel: evm: HMAC attrs: 0x1
Jun 20 00:09:31 fv-az72-309 kernel: PM: Magic number: 10:116:153
Jun 20 00:09:31 fv-az72-309 kernel: RAS: Correctable Errors collector initialized.
Jun 20 00:09:31 fv-az72-309 kernel: md: Waiting for all devices to be available before autodetect
Jun 20 00:09:31 fv-az72-309 kernel: md: If you don't use raid, use raid=noautodetect
Jun 20 00:09:31 fv-az72-309 kernel: md: Autodetecting RAID arrays.
Jun 20 00:09:31 fv-az72-309 kernel: md: autorun ...
Jun 20 00:09:31 fv-az72-309 kernel: md: ... autorun DONE.
Jun 20 00:09:31 fv-az72-309 kernel: EXT4-fs (sda1): mounted filesystem with ordered data mode. Opts: (null). Quota mode: none.
Jun 20 00:09:31 fv-az72-309 kernel: VFS: Mounted root (ext4 filesystem) readonly on device 8:1.
Jun 20 00:09:31 fv-az72-309 kernel: devtmpfs: mounted
Jun 20 00:09:31 fv-az72-309 kernel: Freeing unused decrypted memory: 2036K
Jun 20 00:09:31 fv-az72-309 kernel: Freeing unused kernel image (initmem) memory: 2608K
Jun 20 00:09:31 fv-az72-309 kernel: Write protecting the kernel read-only data: 26624k
Jun 20 00:09:31 fv-az72-309 kernel: Freeing unused kernel image (text/rodata gap) memory: 2036K
Jun 20 00:09:31 fv-az72-309 kernel: Freeing unused kernel image (rodata/data gap) memory: 460K
Jun 20 00:09:31 fv-az72-309 kernel: x86/mm: Checked W+X mappings: passed, no W+X pages found.
Jun 20 00:09:31 fv-az72-309 kernel: x86/mm: Checking user space page tables
Jun 20 00:09:31 fv-az72-309 kernel: x86/mm: Checked W+X mappings: passed, no W+X pages found.
Jun 20 00:09:31 fv-az72-309 kernel: Run /sbin/init as init process
Jun 20 00:09:31 fv-az72-309 kernel: with arguments:
Jun 20 00:09:31 fv-az72-309 kernel: /sbin/init
Jun 20 00:09:31 fv-az72-309 kernel: with environment:
Jun 20 00:09:31 fv-az72-309 kernel: HOME=/
Jun 20 00:09:31 fv-az72-309 kernel: TERM=linux
Jun 20 00:09:31 fv-az72-309 kernel: BOOT_IMAGE=/boot/vmlinuz-5.13.0-1029-azure
Jun 20 00:09:31 fv-az72-309 systemd[1]: Inserted module 'autofs4'
Jun 20 00:09:31 fv-az72-309 systemd[1]: systemd 245.4-4ubuntu3.17 running in system mode. (+PAM +AUDIT +SELINUX +IMA +APPARMOR +SMACK +SYSVINIT +UTMP +LIBCRYPTSETUP +GCRYPT +GNUTLS +ACL +XZ +LZ4 +SECCOMP +BLKID +ELFUTILS +KMOD +IDN2 -IDN +PCRE2 default-hierarchy=hybrid)
Jun 20 00:09:31 fv-az72-309 systemd[1]: Detected virtualization microsoft.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Detected architecture x86-64.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Set hostname to <fv-az72-309>.
Jun 20 00:09:31 fv-az72-309 systemd[1]: /etc/systemd/system/runner-provisioner.service:3: Invalid URL, ignoring: None
Jun 20 00:09:31 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network-online.target ignored (device units cannot be delayed).
Jun 20 00:09:31 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network.target ignored (device units cannot be delayed).
Jun 20 00:09:31 fv-az72-309 systemd[1]: Unnecessary job for /sys/devices/virtual/misc/vmbus!hv_fcopy was removed.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Unnecessary job for /sys/devices/virtual/misc/vmbus!hv_vss was removed.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Created slice Slice for Azure VM Agent and Extensions.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Created slice system-modprobe.slice.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Created slice system-serial\x2dgetty.slice.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Created slice system-systemd\x2dfsck.slice.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Created slice User and Session Slice.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Started Forward Password Requests to Wall Directory Watch.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Set up automount Arbitrary Executable File Formats File System Automount Point.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Reached target User and Group Name Lookups.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Reached target Slices.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Reached target Swap.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Reached target System Time Set.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Listening on Device-mapper event daemon FIFOs.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Listening on LVM2 poll daemon socket.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Listening on multipathd control socket.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Listening on Syslog Socket.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Listening on fsck to fsckd communication Socket.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Listening on initctl Compatibility Named Pipe.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Listening on Journal Audit Socket.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Listening on Journal Socket (/dev/log).
Jun 20 00:09:31 fv-az72-309 systemd[1]: Listening on Journal Socket.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Listening on Network Service Netlink Socket.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Listening on udev Control Socket.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Listening on udev Kernel Socket.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Mounting Huge Pages File System...
Jun 20 00:09:31 fv-az72-309 systemd[1]: Mounting POSIX Message Queue File System...
Jun 20 00:09:31 fv-az72-309 systemd[1]: Mounting Kernel Debug File System...
Jun 20 00:09:31 fv-az72-309 systemd[1]: Mounting Kernel Trace File System...
Jun 20 00:09:31 fv-az72-309 systemd[1]: Starting Journal Service...
Jun 20 00:09:31 fv-az72-309 systemd[1]: Starting Set the console keyboard layout...
Jun 20 00:09:31 fv-az72-309 systemd[1]: Starting Create list of static device nodes for the current kernel...
Jun 20 00:09:31 fv-az72-309 systemd[1]: Starting Monitoring of LVM2 mirrors, snapshots etc. using dmeventd or progress polling...
Jun 20 00:09:31 fv-az72-309 systemd[1]: Starting Load Kernel Module drm...
Jun 20 00:09:31 fv-az72-309 systemd[1]: Condition check resulted in OpenVSwitch configuration for cleanup being skipped.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Condition check resulted in Set Up Additional Binary Formats being skipped.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Starting File System Check on Root Device...
Jun 20 00:09:31 fv-az72-309 systemd[1]: Starting Load Kernel Modules...
Jun 20 00:09:31 fv-az72-309 systemd[1]: Starting udev Coldplug all Devices...
Jun 20 00:09:31 fv-az72-309 systemd[1]: Starting Uncomplicated firewall...
Jun 20 00:09:31 fv-az72-309 systemd[1]: Starting Setup network rules for WALinuxAgent...
Jun 20 00:09:31 fv-az72-309 systemd[1]: Mounted Huge Pages File System.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Mounted POSIX Message Queue File System.
Jun 20 00:09:31 fv-az72-309 systemd-journald[189]: Journal started
Jun 20 00:09:31 fv-az72-309 systemd-journald[189]: Runtime Journal (/run/log/journal/3d8d945fc71147a483fa20cb6792de9d) is 8.0M, max 69.4M, 61.4M free.
Jun 20 00:09:31 fv-az72-309 kernel: IPMI message handler: version 39.2
Jun 20 00:09:31 fv-az72-309 systemd[1]: Mounted Kernel Debug File System.
Jun 20 00:09:31 fv-az72-309 systemd-modules-load[201]: Inserted module 'msr'
Jun 20 00:09:31 fv-az72-309 systemd-fsck[202]: cloudimg-rootfs: clean, 1072970/11096064 files, 14540324/22515963 blocks
Jun 20 00:09:31 fv-az72-309 systemd[1]: Started Journal Service.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Mounted Kernel Trace File System.
Jun 20 00:09:31 fv-az72-309 systemd-modules-load[201]: Inserted module 'ipmi_devintf'
Jun 20 00:09:31 fv-az72-309 kernel: ipmi device interface
Jun 20 00:09:31 fv-az72-309 systemd[1]: Finished Create list of static device nodes for the current kernel.
Jun 20 00:09:31 fv-az72-309 systemd[1]: modprobe@drm.service: Succeeded.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Finished Load Kernel Module drm.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Finished File System Check on Root Device.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Finished Load Kernel Modules.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Finished Uncomplicated firewall.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Mounting FUSE Control File System...
Jun 20 00:09:31 fv-az72-309 systemd[1]: Mounting Kernel Configuration File System...
Jun 20 00:09:31 fv-az72-309 systemd[1]: Started File System Check Daemon to report status.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Starting Remount Root and Kernel File Systems...
Jun 20 00:09:31 fv-az72-309 systemd[1]: Starting Apply Kernel Variables...
Jun 20 00:09:31 fv-az72-309 systemd-sysctl[214]: Not setting net/ipv4/conf/all/promote_secondaries (explicit setting exists).
Jun 20 00:09:31 fv-az72-309 systemd-sysctl[214]: Not setting net/ipv4/conf/default/promote_secondaries (explicit setting exists).
Jun 20 00:09:31 fv-az72-309 systemd[1]: Finished Set the console keyboard layout.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Mounted FUSE Control File System.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Mounted Kernel Configuration File System.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Finished Apply Kernel Variables.
Jun 20 00:09:31 fv-az72-309 kernel: EXT4-fs (sda1): re-mounted. Opts: discard. Quota mode: none.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Finished Remount Root and Kernel File Systems.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Condition check resulted in Rebuild Hardware Database being skipped.
Jun 20 00:09:31 fv-az72-309 systemd[1]: Starting Flush Journal to Persistent Storage...
Jun 20 00:09:32 fv-az72-309 systemd-journald[189]: Time spent on flushing to /var/log/journal/3d8d945fc71147a483fa20cb6792de9d is 35.886ms for 559 entries.
Jun 20 00:09:32 fv-az72-309 systemd-journald[189]: System Journal (/var/log/journal/3d8d945fc71147a483fa20cb6792de9d) is 8.0M, max 4.0G, 3.9G free.
Jun 20 00:09:32 fv-az72-309 systemd[1]: Condition check resulted in Platform Persistent Storage Archival being skipped.
Jun 20 00:09:32 fv-az72-309 systemd[1]: Starting Load/Save Random Seed...
Jun 20 00:09:32 fv-az72-309 systemd[1]: Starting Create System Users...
Jun 20 00:09:32 fv-az72-309 systemd[1]: Finished udev Coldplug all Devices.
Jun 20 00:09:32 fv-az72-309 systemd[1]: Starting udev Wait for Complete Device Initialization...
Jun 20 00:09:32 fv-az72-309 systemd[1]: Finished Load/Save Random Seed.
Jun 20 00:09:32 fv-az72-309 systemd[1]: Finished Create System Users.
Jun 20 00:09:32 fv-az72-309 systemd[1]: Starting Create Static Device Nodes in /dev...
Jun 20 00:09:32 fv-az72-309 systemd[1]: Finished Create Static Device Nodes in /dev.
Jun 20 00:09:32 fv-az72-309 systemd[1]: Starting udev Kernel Device Manager...
Jun 20 00:09:32 fv-az72-309 systemd[1]: Finished Flush Journal to Persistent Storage.
Jun 20 00:09:32 fv-az72-309 systemd[1]: Started udev Kernel Device Manager.
Jun 20 00:09:32 fv-az72-309 systemd[1]: Condition check resulted in Show Plymouth Boot Screen being skipped.
Jun 20 00:09:32 fv-az72-309 systemd[1]: Started Dispatch Password Requests to Console Directory Watch.
Jun 20 00:09:32 fv-az72-309 systemd[1]: Condition check resulted in Forward Password Requests to Plymouth Directory Watch being skipped.
Jun 20 00:09:32 fv-az72-309 systemd[1]: Reached target Local Encrypted Volumes.
Jun 20 00:09:32 fv-az72-309 kernel: hv_vmbus: registering driver hyperv_fb
Jun 20 00:09:32 fv-az72-309 kernel: hyperv_fb: Synthvid Version major 3, minor 5
Jun 20 00:09:32 fv-az72-309 kernel: hyperv_fb: Screen resolution: 1024x768, Color depth: 32, Frame buffer size: 8388608
Jun 20 00:09:32 fv-az72-309 kernel: hyperv_fb: Unable to allocate enough contiguous physical memory on Gen 1 VM. Using MMIO instead.
Jun 20 00:09:32 fv-az72-309 kernel: Console: switching to colour frame buffer device 128x48
Jun 20 00:09:32 fv-az72-309 kernel: hid: raw HID events driver (C) Jiri Kosina
Jun 20 00:09:32 fv-az72-309 kernel: hv_vmbus: registering driver hyperv_keyboard
Jun 20 00:09:32 fv-az72-309 kernel: input: AT Translated Set 2 keyboard as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0A03:00/device:07/VMBUS:01/d34b2567-b9b6-42b9-8778-0a4ec0b955bf/serio2/input/input3
Jun 20 00:09:32 fv-az72-309 kernel: hv_vmbus: registering driver hid_hyperv
Jun 20 00:09:32 fv-az72-309 kernel: input: Microsoft Vmbus HID-compliant Mouse as /devices/0006:045E:0621.0001/input/input4
Jun 20 00:09:32 fv-az72-309 kernel: hid 0006:045E:0621.0001: input: VIRTUAL HID v0.01 Mouse [Microsoft Vmbus HID-compliant Mouse] on
Jun 20 00:09:32 fv-az72-309 kernel: hv_vmbus: registering driver hv_balloon
Jun 20 00:09:32 fv-az72-309 kernel: hv_balloon: Using Dynamic Memory protocol version 2.0
Jun 20 00:09:32 fv-az72-309 kernel: hv_utils: Registering HyperV Utility Driver
Jun 20 00:09:32 fv-az72-309 kernel: hv_vmbus: registering driver hv_utils
Jun 20 00:09:32 fv-az72-309 kernel: hv_utils: Heartbeat IC version 3.0
Jun 20 00:09:32 fv-az72-309 kernel: hv_utils: TimeSync IC version 4.0
Jun 20 00:09:32 fv-az72-309 kernel: hv_utils: Shutdown IC version 3.2
Jun 20 00:09:35 fv-az72-309 kernel: hv_vmbus: registering driver hv_netvsc
Jun 20 00:09:35 fv-az72-309 systemd[1]: Found device /dev/ttyS0.
Jun 20 00:09:35 fv-az72-309 kernel: cryptd: max_cpu_qlen set to 1000
Jun 20 00:09:35 fv-az72-309 kernel: AVX2 version of gcm_enc/dec engaged.
Jun 20 00:09:35 fv-az72-309 kernel: AES CTR mode by8 optimization enabled
Jun 20 00:09:35 fv-az72-309 systemd[1]: Listening on Load/Save RF Kill Switch Status /dev/rfkill Watch.
Jun 20 00:09:35 fv-az72-309 systemd-udevd[225]: ethtool: autonegotiation is unset or enabled, the speed and duplex are not writable.
Jun 20 00:09:35 fv-az72-309 udevadm[221]: systemd-udev-settle.service is deprecated.
Jun 20 00:09:35 fv-az72-309 kernel: bpfilter: Loaded bpfilter_umh pid 297
Jun 20 00:09:35 fv-az72-309 unknown: Started bpfilter
Jun 20 00:09:35 fv-az72-309 systemd-udevd[227]: Using default interface naming scheme 'v245'.
Jun 20 00:09:35 fv-az72-309 systemd[1]: Found device /sys/devices/virtual/misc/vmbus!hv_kvp.
Jun 20 00:09:35 fv-az72-309 systemd[1]: Started Hyper-V KVP Protocol Daemon.
Jun 20 00:09:35 fv-az72-309 python3[217]: Setting up firewall for the WALinux Agent with args: {'dst_ip': '168.63.129.16', 'uid': '0', 'wait': '-w'}
Jun 20 00:09:35 fv-az72-309 python3[217]: Successfully set the firewall rules
Jun 20 00:09:35 fv-az72-309 KVP[317]: KVP starting; pid is:317
Jun 20 00:09:35 fv-az72-309 KVP[317]: KVP LIC Version: 3.1
Jun 20 00:09:35 fv-az72-309 kernel: hv_utils: KVP IC version 4.0
Jun 20 00:09:35 fv-az72-309 systemd[1]: walinuxagent-network-setup.service: Succeeded.
Jun 20 00:09:35 fv-az72-309 systemd[1]: Finished Setup network rules for WALinuxAgent.
Jun 20 00:09:35 fv-az72-309 systemd[1]: Found device Virtual_Disk UEFI.
Jun 20 00:09:35 fv-az72-309 systemd[1]: Found device Virtual_Disk Temporary_Storage.
Jun 20 00:09:35 fv-az72-309 systemd[1]: Finished Monitoring of LVM2 mirrors, snapshots etc. using dmeventd or progress polling.
Jun 20 00:09:36 fv-az72-309 systemd-udevd[227]: ethtool: autonegotiation is unset or enabled, the speed and duplex are not writable.
Jun 20 00:09:36 fv-az72-309 systemd[1]: Finished udev Wait for Complete Device Initialization.
Jun 20 00:09:36 fv-az72-309 systemd[1]: Starting Device-Mapper Multipath Device Controller...
Jun 20 00:09:36 fv-az72-309 kernel: alua: device handler registered
Jun 20 00:09:36 fv-az72-309 kernel: emc: device handler registered
Jun 20 00:09:36 fv-az72-309 kernel: rdac: device handler registered
Jun 20 00:09:36 fv-az72-309 multipathd[399]: --------start up--------
Jun 20 00:09:36 fv-az72-309 multipathd[399]: read /etc/multipath.conf
Jun 20 00:09:36 fv-az72-309 multipathd[399]: path checkers start up
Jun 20 00:09:36 fv-az72-309 systemd[1]: Started Device-Mapper Multipath Device Controller.
Jun 20 00:09:36 fv-az72-309 systemd[1]: Reached target Local File Systems (Pre).
Jun 20 00:09:36 fv-az72-309 systemd[1]: Mounting Mount unit for core20, revision 1518...
Jun 20 00:09:36 fv-az72-309 systemd[1]: Mounting Mount unit for lxd, revision 22753...
Jun 20 00:09:36 fv-az72-309 systemd[1]: Mounting Mount unit for snapd, revision 16010...
Jun 20 00:09:36 fv-az72-309 kernel: loop0: detected capacity change from 0 to 126824
Jun 20 00:09:36 fv-az72-309 systemd[1]: Starting File System Check on /dev/disk/by-uuid/7167-9500...
Jun 20 00:09:36 fv-az72-309 systemd[1]: Starting File System Check on /dev/disk/cloud/azure_resource-part1...
Jun 20 00:09:36 fv-az72-309 kernel: loop1: detected capacity change from 0 to 96160
Jun 20 00:09:36 fv-az72-309 systemd-fsck[411]: sdb1: fsck.ntfs doesn't exist, not checking file system.
Jun 20 00:09:36 fv-az72-309 systemd[1]: Mounted Mount unit for core20, revision 1518.
Jun 20 00:09:36 fv-az72-309 systemd[1]: Finished File System Check on /dev/disk/cloud/azure_resource-part1.
Jun 20 00:09:36 fv-az72-309 kernel: loop2: detected capacity change from 0 to 138880
Jun 20 00:09:36 fv-az72-309 systemd[1]: Mounted Mount unit for snapd, revision 16010.
Jun 20 00:09:36 fv-az72-309 systemd[1]: Mounted Mount unit for lxd, revision 22753.
Jun 20 00:09:36 fv-az72-309 systemd-fsck[417]: fsck.fat 4.1 (2017-01-24)
Jun 20 00:09:36 fv-az72-309 systemd-fsck[417]: /dev/sda15: 12 files, 10642/213716 clusters
Jun 20 00:09:36 fv-az72-309 systemd[1]: Finished File System Check on /dev/disk/by-uuid/7167-9500.
Jun 20 00:09:36 fv-az72-309 systemd[1]: Mounting /boot/efi...
Jun 20 00:09:36 fv-az72-309 systemd[1]: Mounted /boot/efi.
Jun 20 00:09:36 fv-az72-309 systemd[1]: Reached target Local File Systems.
Jun 20 00:09:36 fv-az72-309 systemd[1]: Starting Load AppArmor profiles...
Jun 20 00:09:36 fv-az72-309 systemd[1]: Starting Enable support for additional executable binary formats...
Jun 20 00:09:36 fv-az72-309 systemd[1]: Starting Set console font and keymap...
Jun 20 00:09:36 fv-az72-309 systemd[1]: Starting Create final runtime dir for shutdown pivot root...
Jun 20 00:09:36 fv-az72-309 systemd[1]: Condition check resulted in LXD - agent - 9p mount being skipped.
Jun 20 00:09:36 fv-az72-309 systemd[1]: Condition check resulted in LXD - agent being skipped.
Jun 20 00:09:36 fv-az72-309 apparmor.systemd[425]: Restarting AppArmor
Jun 20 00:09:36 fv-az72-309 systemd[1]: Starting Tell Plymouth To Write Out Runtime Data...
Jun 20 00:09:36 fv-az72-309 apparmor.systemd[425]: Reloading AppArmor profiles
Jun 20 00:09:36 fv-az72-309 systemd[1]: Condition check resulted in Store a System Token in an EFI Variable being skipped.
Jun 20 00:09:36 fv-az72-309 systemd[1]: Condition check resulted in Commit a transient machine-id on disk being skipped.
Jun 20 00:09:36 fv-az72-309 systemd[1]: Starting Create Volatile Files and Directories...
Jun 20 00:09:36 fv-az72-309 systemd[1]: Finished Set console font and keymap.
Jun 20 00:09:36 fv-az72-309 systemd[1]: Finished Create final runtime dir for shutdown pivot root.
Jun 20 00:09:36 fv-az72-309 systemd[1]: plymouth-read-write.service: Succeeded.
Jun 20 00:09:36 fv-az72-309 systemd[1]: Finished Tell Plymouth To Write Out Runtime Data.
Jun 20 00:09:36 fv-az72-309 systemd[1]: Finished Create Volatile Files and Directories.
Jun 20 00:09:36 fv-az72-309 apparmor.systemd[439]: Skipping profile in /etc/apparmor.d/disable: usr.sbin.rsyslogd
Jun 20 00:09:36 fv-az72-309 apparmor.systemd[440]: Skipping profile in /etc/apparmor.d/disable: usr.bin.firefox
Jun 20 00:09:36 fv-az72-309 audit[438]: AVC apparmor="STATUS" operation="profile_load" profile="unconfined" name="lsb_release" pid=438 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 systemd[1]: proc-sys-fs-binfmt_misc.automount: Got automount request for /proc/sys/fs/binfmt_misc, triggered by 427 (update-binfmts)
Jun 20 00:09:36 fv-az72-309 systemd[1]: Mounting Arbitrary Executable File Formats File System...
Jun 20 00:09:36 fv-az72-309 kernel: audit: type=1400 audit(1655683776.240:2): apparmor="STATUS" operation="profile_load" profile="unconfined" name="lsb_release" pid=438 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 systemd[1]: Starting Update UTMP about System Boot/Shutdown...
Jun 20 00:09:36 fv-az72-309 audit[437]: AVC apparmor="STATUS" operation="profile_load" profile="unconfined" name="/usr/lib/NetworkManager/nm-dhcp-client.action" pid=437 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 audit[437]: AVC apparmor="STATUS" operation="profile_load" profile="unconfined" name="/usr/lib/NetworkManager/nm-dhcp-helper" pid=437 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 audit[437]: AVC apparmor="STATUS" operation="profile_load" profile="unconfined" name="/usr/lib/connman/scripts/dhclient-script" pid=437 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 audit[437]: AVC apparmor="STATUS" operation="profile_load" profile="unconfined" name="/{,usr/}sbin/dhclient" pid=437 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 kernel: audit: type=1400 audit(1655683776.248:3): apparmor="STATUS" operation="profile_load" profile="unconfined" name="/usr/lib/NetworkManager/nm-dhcp-client.action" pid=437 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 kernel: audit: type=1400 audit(1655683776.248:4): apparmor="STATUS" operation="profile_load" profile="unconfined" name="/usr/lib/NetworkManager/nm-dhcp-helper" pid=437 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 kernel: audit: type=1400 audit(1655683776.248:5): apparmor="STATUS" operation="profile_load" profile="unconfined" name="/usr/lib/connman/scripts/dhclient-script" pid=437 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 kernel: audit: type=1400 audit(1655683776.248:6): apparmor="STATUS" operation="profile_load" profile="unconfined" name="/{,usr/}sbin/dhclient" pid=437 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 audit[441]: AVC apparmor="STATUS" operation="profile_load" profile="unconfined" name="/usr/sbin/haveged" pid=441 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 kernel: audit: type=1400 audit(1655683776.252:7): apparmor="STATUS" operation="profile_load" profile="unconfined" name="/usr/sbin/haveged" pid=441 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 audit[444]: AVC apparmor="STATUS" operation="profile_load" profile="unconfined" name="/usr/sbin/mysqld" pid=444 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 kernel: audit: type=1400 audit(1655683776.260:8): apparmor="STATUS" operation="profile_load" profile="unconfined" name="/usr/sbin/mysqld" pid=444 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 audit[445]: AVC apparmor="STATUS" operation="profile_load" profile="unconfined" name="/usr/bin/man" pid=445 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 audit[445]: AVC apparmor="STATUS" operation="profile_load" profile="unconfined" name="man_filter" pid=445 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 audit[445]: AVC apparmor="STATUS" operation="profile_load" profile="unconfined" name="man_groff" pid=445 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 systemd[1]: Finished Update UTMP about System Boot/Shutdown.
Jun 20 00:09:36 fv-az72-309 kernel: audit: type=1400 audit(1655683776.264:9): apparmor="STATUS" operation="profile_load" profile="unconfined" name="/usr/bin/man" pid=445 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 kernel: audit: type=1400 audit(1655683776.264:10): apparmor="STATUS" operation="profile_load" profile="unconfined" name="man_filter" pid=445 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 audit[448]: AVC apparmor="STATUS" operation="profile_load" profile="unconfined" name="/usr/sbin/chronyd" pid=448 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 systemd[1]: Mounted Arbitrary Executable File Formats File System.
Jun 20 00:09:36 fv-az72-309 audit[447]: AVC apparmor="STATUS" operation="profile_load" profile="unconfined" name="/usr/lib/snapd/snap-confine" pid=447 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 audit[447]: AVC apparmor="STATUS" operation="profile_load" profile="unconfined" name="/usr/lib/snapd/snap-confine//mount-namespace-capture-helper" pid=447 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 audit[450]: AVC apparmor="STATUS" operation="profile_load" profile="unconfined" name="nvidia_modprobe" pid=450 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 audit[450]: AVC apparmor="STATUS" operation="profile_load" profile="unconfined" name="nvidia_modprobe//kmod" pid=450 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 audit[449]: AVC apparmor="STATUS" operation="profile_load" profile="unconfined" name="/usr/sbin/tcpdump" pid=449 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 systemd[1]: Finished Load AppArmor profiles.
Jun 20 00:09:36 fv-az72-309 systemd[1]: Finished Enable support for additional executable binary formats.
Jun 20 00:09:36 fv-az72-309 systemd[1]: Started Entropy daemon using the HAVEGE algorithm.
Jun 20 00:09:36 fv-az72-309 systemd[1]: Starting Load AppArmor profiles managed internally by snapd...
Jun 20 00:09:36 fv-az72-309 systemd[1]: Condition check resulted in Authentication service for virtual machines hosted on VMware being skipped.
Jun 20 00:09:36 fv-az72-309 systemd[1]: Condition check resulted in Service for virtual machines hosted on VMware being skipped.
Jun 20 00:09:36 fv-az72-309 systemd[1]: Starting Initial cloud-init job (pre-networking)...
Jun 20 00:09:36 fv-az72-309 sh[454]: + [ -e /var/lib/cloud/instance/obj.pkl ]
Jun 20 00:09:36 fv-az72-309 sh[454]: + echo cleaning persistent cloud-init object
Jun 20 00:09:36 fv-az72-309 sh[454]: cleaning persistent cloud-init object
Jun 20 00:09:36 fv-az72-309 sh[454]: + rm /var/lib/cloud/instance/obj.pkl
Jun 20 00:09:36 fv-az72-309 haveged[451]: haveged starting up
Jun 20 00:09:36 fv-az72-309 sh[454]: + exit 0
Jun 20 00:09:36 fv-az72-309 audit[465]: AVC apparmor="STATUS" operation="profile_load" profile="unconfined" name="snap-update-ns.lxd" pid=465 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 audit[464]: AVC apparmor="STATUS" operation="profile_load" profile="unconfined" name="/snap/snapd/16010/usr/lib/snapd/snap-confine" pid=464 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 audit[464]: AVC apparmor="STATUS" operation="profile_load" profile="unconfined" name="/snap/snapd/16010/usr/lib/snapd/snap-confine//mount-namespace-capture-helper" pid=464 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 audit[466]: AVC apparmor="STATUS" operation="profile_load" profile="unconfined" name="snap.lxd.activate" pid=466 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 audit[467]: AVC apparmor="STATUS" operation="profile_load" profile="unconfined" name="snap.lxd.benchmark" pid=467 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 audit[468]: AVC apparmor="STATUS" operation="profile_load" profile="unconfined" name="snap.lxd.buginfo" pid=468 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 audit[469]: AVC apparmor="STATUS" operation="profile_load" profile="unconfined" name="snap.lxd.check-kernel" pid=469 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 audit[470]: AVC apparmor="STATUS" operation="profile_load" profile="unconfined" name="snap.lxd.daemon" pid=470 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 audit[471]: AVC apparmor="STATUS" operation="profile_load" profile="unconfined" name="snap.lxd.hook.configure" pid=471 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 audit[472]: AVC apparmor="STATUS" operation="profile_load" profile="unconfined" name="snap.lxd.hook.install" pid=472 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 audit[473]: AVC apparmor="STATUS" operation="profile_load" profile="unconfined" name="snap.lxd.hook.remove" pid=473 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 audit[474]: AVC apparmor="STATUS" operation="profile_load" profile="unconfined" name="snap.lxd.lxc" pid=474 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 audit[475]: AVC apparmor="STATUS" operation="profile_load" profile="unconfined" name="snap.lxd.lxc-to-lxd" pid=475 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 audit[476]: AVC apparmor="STATUS" operation="profile_load" profile="unconfined" name="snap.lxd.lxd" pid=476 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 audit[477]: AVC apparmor="STATUS" operation="profile_load" profile="unconfined" name="snap.lxd.migrate" pid=477 comm="apparmor_parser"
Jun 20 00:09:36 fv-az72-309 systemd[1]: Finished Load AppArmor profiles managed internally by snapd.
Jun 20 00:09:36 fv-az72-309 haveged[451]: haveged: ver: 1.9.1; arch: x86; vend: GenuineIntel; build: (gcc 8.3.0 ITV); collect: 128K
Jun 20 00:09:36 fv-az72-309 haveged[451]: haveged: cpu: (L4 VC); data: 32K (L4 V); inst: 32K (L4 V); idx: 24/40; sz: 31410/52825
Jun 20 00:09:36 fv-az72-309 haveged[451]: haveged: tot tests(BA8): A:1/1 B:1/1 continuous tests(B): last entropy estimate 8.00164
Jun 20 00:09:36 fv-az72-309 haveged[451]: haveged: fills: 0, generated: 0
Jun 20 00:09:37 fv-az72-309 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready
Jun 20 00:09:37 fv-az72-309 dhclient[491]: Internet Systems Consortium DHCP Client 4.4.1
Jun 20 00:09:37 fv-az72-309 dhclient[491]: Copyright 2004-2018 Internet Systems Consortium.
Jun 20 00:09:37 fv-az72-309 dhclient[491]: All rights reserved.
Jun 20 00:09:37 fv-az72-309 dhclient[491]: For info, please visit https://www.isc.org/software/dhcp/
Jun 20 00:09:37 fv-az72-309 dhclient[491]:
Jun 20 00:09:37 fv-az72-309 dhclient[491]: Listening on LPF/eth0/00:0d:3a:13:05:24
Jun 20 00:09:37 fv-az72-309 dhclient[491]: Sending on LPF/eth0/00:0d:3a:13:05:24
Jun 20 00:09:37 fv-az72-309 dhclient[491]: Sending on Socket/fallback
Jun 20 00:09:37 fv-az72-309 dhclient[491]: DHCPDISCOVER on eth0 to 255.255.255.255 port 67 interval 3 (xid=0x13264032)
Jun 20 00:09:37 fv-az72-309 dhclient[491]: DHCPOFFER of 10.1.0.17 from 168.63.129.16
Jun 20 00:09:37 fv-az72-309 dhclient[491]: DHCPREQUEST for 10.1.0.17 on eth0 to 255.255.255.255 port 67 (xid=0x32402613)
Jun 20 00:09:37 fv-az72-309 dhclient[491]: DHCPACK of 10.1.0.17 from 168.63.129.16 (xid=0x13264032)
Jun 20 00:09:37 fv-az72-309 dhclient[491]: Timeout too large reducing to: 2147483646 (TIME_MAX - 1)
Jun 20 00:09:37 fv-az72-309 dhclient[491]: bound to 10.1.0.17 -- renewal in 4294967295 seconds.
Jun 20 00:09:37 fv-az72-309 systemd[1]: Condition check resulted in OpenVSwitch configuration for cleanup being skipped.
Jun 20 00:09:37 fv-az72-309 systemd[1]: Starting Setup network rules for WALinuxAgent...
Jun 20 00:09:37 fv-az72-309 systemd[1]: Condition check resulted in OpenVSwitch configuration for cleanup being skipped.
Jun 20 00:09:37 fv-az72-309 systemd[1]: Condition check resulted in OpenVSwitch configuration for cleanup being skipped.
Jun 20 00:09:37 fv-az72-309 cloud-init[481]: Cloud-init v. 22.2-0ubuntu1~20.04.2 running 'init-local' at Mon, 20 Jun 2022 00:09:36 +0000. Up 5.24 seconds.
Jun 20 00:09:37 fv-az72-309 systemd[1]: Finished Initial cloud-init job (pre-networking).
Jun 20 00:09:37 fv-az72-309 python3[522]: Setting up firewall for the WALinux Agent with args: {'dst_ip': '168.63.129.16', 'uid': '0', 'wait': '-w'}
Jun 20 00:09:37 fv-az72-309 python3[522]: Successfully set the firewall rules
Jun 20 00:09:37 fv-az72-309 systemd[1]: walinuxagent-network-setup.service: Succeeded.
Jun 20 00:09:37 fv-az72-309 systemd[1]: Finished Setup network rules for WALinuxAgent.
Jun 20 00:09:37 fv-az72-309 systemd[1]: Reached target Network (Pre).
Jun 20 00:09:37 fv-az72-309 systemd[1]: Starting Network Service...
Jun 20 00:09:37 fv-az72-309 systemd-networkd[532]: Enumeration completed
Jun 20 00:09:37 fv-az72-309 systemd-networkd[532]: eth0: Link UP
Jun 20 00:09:37 fv-az72-309 systemd[1]: Started Network Service.
Jun 20 00:09:37 fv-az72-309 systemd-networkd[532]: eth0: Gained carrier
Jun 20 00:09:38 fv-az72-309 systemd-networkd[532]: eth0: Link DOWN
Jun 20 00:09:38 fv-az72-309 systemd-networkd[532]: eth0: Lost carrier
Jun 20 00:09:38 fv-az72-309 systemd-networkd[532]: eth0: IPv6 successfully enabled
Jun 20 00:09:38 fv-az72-309 systemd[1]: Starting Wait for Network to be Configured...
Jun 20 00:09:38 fv-az72-309 systemd-networkd[532]: eth0: Link UP
Jun 20 00:09:38 fv-az72-309 systemd[1]: Starting Network Name Resolution...
Jun 20 00:09:38 fv-az72-309 systemd-networkd[532]: eth0: Gained carrier
Jun 20 00:09:38 fv-az72-309 systemd-networkd[532]: eth0: DHCPv4 address 10.1.0.17/16 via 10.1.0.1
Jun 20 00:09:38 fv-az72-309 systemd-resolved[534]: Positive Trust Anchors:
Jun 20 00:09:38 fv-az72-309 systemd-resolved[534]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d
Jun 20 00:09:38 fv-az72-309 systemd-resolved[534]: Negative trust anchors: 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test
Jun 20 00:09:38 fv-az72-309 systemd-resolved[534]: Using system hostname 'fv-az72-309'.
Jun 20 00:09:38 fv-az72-309 systemd[1]: Started Network Name Resolution.
Jun 20 00:09:38 fv-az72-309 systemd[1]: Reached target Network.
Jun 20 00:09:38 fv-az72-309 systemd[1]: Reached target Host and Network Name Lookups.
Jun 20 00:09:39 fv-az72-309 systemd-networkd[532]: eth0: Gained IPv6LL
Jun 20 00:09:39 fv-az72-309 systemd-networkd-wait-online[533]: managing: eth0
Jun 20 00:09:39 fv-az72-309 systemd[1]: Finished Wait for Network to be Configured.
Jun 20 00:09:39 fv-az72-309 systemd[1]: Starting Initial cloud-init job (metadata service crawler)...
Jun 20 00:09:39 fv-az72-309 cloud-init[540]: Cloud-init v. 22.2-0ubuntu1~20.04.2 running 'init' at Mon, 20 Jun 2022 00:09:39 +0000. Up 7.70 seconds.
Jun 20 00:09:39 fv-az72-309 cloud-init[540]: ci-info: +++++++++++++++++++++++++++++++++++++Net device info+++++++++++++++++++++++++++++++++++++
Jun 20 00:09:39 fv-az72-309 cloud-init[540]: ci-info: +--------+------+----------------------------+-------------+--------+-------------------+
Jun 20 00:09:39 fv-az72-309 cloud-init[540]: ci-info: | Device | Up | Address | Mask | Scope | Hw-Address |
Jun 20 00:09:39 fv-az72-309 cloud-init[540]: ci-info: +--------+------+----------------------------+-------------+--------+-------------------+
Jun 20 00:09:39 fv-az72-309 cloud-init[540]: ci-info: | eth0 | True | 10.1.0.17 | 255.255.0.0 | global | 00:0d:3a:13:05:24 |
Jun 20 00:09:39 fv-az72-309 cloud-init[540]: ci-info: | eth0 | True | fe80::20d:3aff:fe13:524/64 | . | link | 00:0d:3a:13:05:24 |
Jun 20 00:09:39 fv-az72-309 cloud-init[540]: ci-info: | lo | True | 127.0.0.1 | 255.0.0.0 | host | . |
Jun 20 00:09:39 fv-az72-309 cloud-init[540]: ci-info: | lo | True | ::1/128 | . | host | . |
Jun 20 00:09:39 fv-az72-309 cloud-init[540]: ci-info: +--------+------+----------------------------+-------------+--------+-------------------+
Jun 20 00:09:39 fv-az72-309 cloud-init[540]: ci-info: ++++++++++++++++++++++++++++++Route IPv4 info+++++++++++++++++++++++++++++++
Jun 20 00:09:39 fv-az72-309 cloud-init[540]: ci-info: +-------+-----------------+----------+-----------------+-----------+-------+
Jun 20 00:09:39 fv-az72-309 cloud-init[540]: ci-info: | Route | Destination | Gateway | Genmask | Interface | Flags |
Jun 20 00:09:39 fv-az72-309 cloud-init[540]: ci-info: +-------+-----------------+----------+-----------------+-----------+-------+
Jun 20 00:09:39 fv-az72-309 cloud-init[540]: ci-info: | 0 | 0.0.0.0 | 10.1.0.1 | 0.0.0.0 | eth0 | UG |
Jun 20 00:09:39 fv-az72-309 cloud-init[540]: ci-info: | 1 | 10.1.0.0 | 0.0.0.0 | 255.255.0.0 | eth0 | U |
Jun 20 00:09:39 fv-az72-309 cloud-init[540]: ci-info: | 2 | 168.63.129.16 | 10.1.0.1 | 255.255.255.255 | eth0 | UGH |
Jun 20 00:09:39 fv-az72-309 cloud-init[540]: ci-info: | 3 | 169.254.169.254 | 10.1.0.1 | 255.255.255.255 | eth0 | UGH |
Jun 20 00:09:39 fv-az72-309 cloud-init[540]: ci-info: +-------+-----------------+----------+-----------------+-----------+-------+
Jun 20 00:09:39 fv-az72-309 cloud-init[540]: ci-info: +++++++++++++++++++Route IPv6 info+++++++++++++++++++
Jun 20 00:09:39 fv-az72-309 cloud-init[540]: ci-info: +-------+-------------+---------+-----------+-------+
Jun 20 00:09:39 fv-az72-309 cloud-init[540]: ci-info: | Route | Destination | Gateway | Interface | Flags |
Jun 20 00:09:39 fv-az72-309 cloud-init[540]: ci-info: +-------+-------------+---------+-----------+-------+
Jun 20 00:09:39 fv-az72-309 cloud-init[540]: ci-info: | 1 | fe80::/64 | :: | eth0 | U |
Jun 20 00:09:39 fv-az72-309 cloud-init[540]: ci-info: | 3 | local | :: | eth0 | U |
Jun 20 00:09:39 fv-az72-309 cloud-init[540]: ci-info: | 4 | multicast | :: | eth0 | U |
Jun 20 00:09:39 fv-az72-309 cloud-init[540]: ci-info: +-------+-------------+---------+-----------+-------+
Jun 20 00:09:39 fv-az72-309 ntfs-3g[551]: Version 2017.3.23AR.3 integrated FUSE 28
Jun 20 00:09:39 fv-az72-309 ntfs-3g[551]: Mounted /dev/sdb1 (Read-Only, label "Temporary Storage", NTFS 3.1)
Jun 20 00:09:39 fv-az72-309 ntfs-3g[551]: Cmdline options: ro
Jun 20 00:09:39 fv-az72-309 ntfs-3g[551]: Mount options: ro,allow_other,nonempty,relatime,fsname=/dev/sdb1,blkdev,blksize=4096
Jun 20 00:09:39 fv-az72-309 ntfs-3g[551]: Ownership and permissions disabled, configuration type 7
Jun 20 00:09:39 fv-az72-309 ntfs-3g[551]: Unmounting /dev/sdb1 (Temporary Storage)
Jun 20 00:09:39 fv-az72-309 systemd[1]: run-cloud\x2dinit-tmp-tmp5pbcsij7.mount: Succeeded.
Jun 20 00:09:39 fv-az72-309 kernel: sdb: sdb1
Jun 20 00:09:39 fv-az72-309 systemd[1]: systemd-fsck@dev-disk-cloud-azure_resource\x2dpart1.service: Succeeded.
Jun 20 00:09:39 fv-az72-309 systemd[1]: Stopped File System Check on /dev/disk/cloud/azure_resource-part1.
Jun 20 00:09:40 fv-az72-309 systemd[1]: Starting Tell Plymouth To Write Out Runtime Data...
Jun 20 00:09:40 fv-az72-309 systemd[1]: Condition check resulted in Show Plymouth Boot Screen being skipped.
Jun 20 00:09:40 fv-az72-309 systemd[1]: Condition check resulted in Forward Password Requests to Plymouth Directory Watch being skipped.
Jun 20 00:09:40 fv-az72-309 systemd[1]: Condition check resulted in Set Up Additional Binary Formats being skipped.
Jun 20 00:09:40 fv-az72-309 systemd[1]: Condition check resulted in Store a System Token in an EFI Variable being skipped.
Jun 20 00:09:40 fv-az72-309 systemd[1]: Condition check resulted in Rebuild Hardware Database being skipped.
Jun 20 00:09:40 fv-az72-309 systemd[1]: Condition check resulted in Commit a transient machine-id on disk being skipped.
Jun 20 00:09:40 fv-az72-309 systemd[1]: Condition check resulted in Platform Persistent Storage Archival being skipped.
Jun 20 00:09:40 fv-az72-309 systemd[1]: plymouth-read-write.service: Succeeded.
Jun 20 00:09:40 fv-az72-309 systemd[1]: Finished Tell Plymouth To Write Out Runtime Data.
Jun 20 00:09:40 fv-az72-309 kernel: EXT4-fs (sdb1): mounted filesystem with ordered data mode. Opts: (null). Quota mode: none.
Jun 20 00:09:40 fv-az72-309 systemd[1]: Reloading.
Jun 20 00:09:40 fv-az72-309 systemd[1]: /etc/systemd/system/runner-provisioner.service:3: Invalid URL, ignoring: None
Jun 20 00:09:41 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network-online.target ignored (device units cannot be delayed).
Jun 20 00:09:41 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network.target ignored (device units cannot be delayed).
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: Generating public/private rsa key pair.
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: Your identification has been saved in /etc/ssh/ssh_host_rsa_key
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: Your public key has been saved in /etc/ssh/ssh_host_rsa_key.pub
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: The key fingerprint is:
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: SHA256:xgHytXr9AoLrM03Ufz16C0Uv9NChaKGn5zsbtoCu2Gg root@fv-az72-309
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: The key's randomart image is:
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: +---[RSA 3072]----+
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: | . . . . . |
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: | o o .. o ...|
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: | ..o. + .+ .|
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: | ..o.o+ o + |
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: | ..o So.. .o o|
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: | ..+..+...o. |
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: | .o . ..=o. . |
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: | E* o oo=.. |
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: | .o.=.. +o... |
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: +----[SHA256]-----+
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: Generating public/private dsa key pair.
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: Your identification has been saved in /etc/ssh/ssh_host_dsa_key
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: Your public key has been saved in /etc/ssh/ssh_host_dsa_key.pub
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: The key fingerprint is:
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: SHA256:XGcH1MwJBdIF6Dewo/X0osa6enb06nIxLFk5/q6UAm8 root@fv-az72-309
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: The key's randomart image is:
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: +---[DSA 1024]----+
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: | .+*Xo. |
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: | o...= |
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: | ..+o . |
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: | . .Oo+. |
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: | . SB * o |
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: | o+ *.o . |
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: | E+o* . |
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: | .+o* o |
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: | .+oO++. |
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: +----[SHA256]-----+
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: Generating public/private ecdsa key pair.
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: Your identification has been saved in /etc/ssh/ssh_host_ecdsa_key
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: Your public key has been saved in /etc/ssh/ssh_host_ecdsa_key.pub
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: The key fingerprint is:
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: SHA256:6mlnK30thyw5xPDk04GqP6xQuDIEULkyXUpZlLeH/us root@fv-az72-309
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: The key's randomart image is:
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: +---[ECDSA 256]---+
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: |...=o. |
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: |. + o . |
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: |.o + . o . |
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: |+ +. + + . |
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: | +. .. OS. . |
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: |. o o.* . |
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: |o o o.+ + o |
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: | o ...=.O = o |
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: | .++*E* o |
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: +----[SHA256]-----+
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: Generating public/private ed25519 key pair.
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: Your identification has been saved in /etc/ssh/ssh_host_ed25519_key
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: Your public key has been saved in /etc/ssh/ssh_host_ed25519_key.pub
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: The key fingerprint is:
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: SHA256:CkHtFnUi6AlzMN/rV+bhb10BzsSd4ft4SZpQoIuutqU root@fv-az72-309
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: The key's randomart image is:
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: +--[ED25519 256]--+
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: | o..o..o ... ..o|
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: | o+o.o. o. .+.o |
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: | =oo.. . +... |
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: | o.o.. . .o ..|
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: | ....S.+. o.|
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: | o.. = .. +.+|
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: | oo. o +.oo|
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: | .+. .. .. |
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: | .E. .. |
Jun 20 00:09:41 fv-az72-309 cloud-init[540]: +----[SHA256]-----+
Jun 20 00:09:41 fv-az72-309 systemd[1]: Finished Initial cloud-init job (metadata service crawler).
Jun 20 00:09:41 fv-az72-309 systemd[1]: Reached target Cloud-config availability.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Reached target Network is Online.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Reached target System Initialization.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Condition check resulted in Process error reports when automatic reporting is enabled (file watch) being skipped.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Started Trigger to poll for Ubuntu Pro licenses (Only enabled on GCP LTS non-pro).
Jun 20 00:09:41 fv-az72-309 systemd[1]: Started Daily Cleanup of Temporary Directories.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Started Ubuntu Advantage Timer for running repeated jobs.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Reached target Paths.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Condition check resulted in Unix socket for apport crash forwarding being skipped.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Listening on cloud-init hotplug hook socket.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Listening on D-Bus System Message Bus Socket.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Starting Docker Socket for the API.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Listening on Open-iSCSI iscsid Socket.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Listening on Socket unix for snap application lxd.daemon.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Starting Socket activation for snappy daemon.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Listening on UUID daemon activation socket.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Condition check resulted in Login to default iSCSI targets being skipped.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Reached target Remote File Systems (Pre).
Jun 20 00:09:41 fv-az72-309 systemd[1]: Reached target Remote File Systems.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Starting Availability of block devices...
Jun 20 00:09:41 fv-az72-309 systemd[1]: Listening on Docker Socket for the API.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Listening on Socket activation for snappy daemon.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Finished Availability of block devices.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Reached target Sockets.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Reached target Basic System.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Starting Accounts Service...
Jun 20 00:09:41 fv-az72-309 systemd[1]: Starting LSB: automatic crash report generation...
Jun 20 00:09:41 fv-az72-309 systemd[1]: Starting Deferred execution scheduler...
Jun 20 00:09:41 fv-az72-309 systemd[1]: Starting chrony, an NTP client/server...
Jun 20 00:09:41 fv-az72-309 systemd[1]: Starting containerd container runtime...
Jun 20 00:09:41 fv-az72-309 systemd[1]: Started Regular background program processing daemon.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Started D-Bus System Message Bus.
Jun 20 00:09:41 fv-az72-309 cron[655]: (CRON) INFO (pidfile fd = 3)
Jun 20 00:09:41 fv-az72-309 cron[655]: (CRON) INFO (Running @reboot jobs)
Jun 20 00:09:41 fv-az72-309 systemd[1]: Started Save initial kernel messages after boot.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Starting Remove Stale Online ext4 Metadata Check Snapshots...
Jun 20 00:09:41 fv-az72-309 systemd[1]: Condition check resulted in getty on tty2-tty6 if dbus and logind are not available being skipped.
Jun 20 00:09:41 fv-az72-309 chronyd-starter.sh[660]: WARNING: libcap needs an update (cap=40 should have a name).
Jun 20 00:09:41 fv-az72-309 systemd[1]: Starting Record successful boot for GRUB...
Jun 20 00:09:41 fv-az72-309 systemd[1]: Condition check resulted in Hyper-V File Copy Protocol Daemon being skipped.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Condition check resulted in Hyper-V VSS Protocol Daemon being skipped.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Started irqbalance daemon.
Jun 20 00:09:41 fv-az72-309 chronyd[666]: chronyd version 3.5 starting (+CMDMON +NTP +REFCLOCK +RTC +PRIVDROP +SCFILTER +SIGND +ASYNCDNS +SECHASH +IPV6 -DEBUG)
Jun 20 00:09:41 fv-az72-309 chronyd[666]: Frequency -44.376 +/- 1.296 ppm read from /var/lib/chrony/chrony.drift
Jun 20 00:09:41 fv-az72-309 chronyd[666]: Loaded seccomp filter
Jun 20 00:09:41 fv-az72-309 systemd[1]: Starting LSB: Mono XSP4...
Jun 20 00:09:41 fv-az72-309 dbus-daemon[658]: [system] AppArmor D-Bus mediation is enabled
Jun 20 00:09:41 fv-az72-309 systemd[1]: Starting Dispatcher daemon for systemd-networkd...
Jun 20 00:09:41 fv-az72-309 systemd[1]: Condition check resulted in Set the CPU Frequency Scaling governor being skipped.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Starting The PHP 7.4 FastCGI Process Manager...
Jun 20 00:09:41 fv-az72-309 systemd[1]: Starting The PHP 8.0 FastCGI Process Manager...
Jun 20 00:09:41 fv-az72-309 systemd[1]: Starting The PHP 8.1 FastCGI Process Manager...
Jun 20 00:09:41 fv-az72-309 systemd[1]: Starting Authorization Manager...
Jun 20 00:09:41 fv-az72-309 systemd[1]: Condition check resulted in Pollinate to seed the pseudo random number generator being skipped.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Condition check resulted in fast remote file copy program daemon being skipped.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Starting System Logging Service...
Jun 20 00:09:41 fv-az72-309 systemd[1]: Started Job that runs the Runner Provisioner.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Condition check resulted in Secure Boot updates for DB and DBX being skipped.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Started session-manager-plugin.
Jun 20 00:09:41 fv-az72-309 apport[649]: * Starting automatic crash report generation: apport
Jun 20 00:09:41 fv-az72-309 systemd[1]: Starting Service for snap application lxd.activate...
Jun 20 00:09:41 fv-az72-309 rsyslogd[682]: imuxsock: Acquired UNIX socket '/run/systemd/journal/syslog' (fd 3) from systemd. [v8.2001.0]
Jun 20 00:09:41 fv-az72-309 systemd[1]: Condition check resulted in Automatically repair incorrect owner/permissions on core devices being skipped.
Jun 20 00:09:41 fv-az72-309 rsyslogd[682]: rsyslogd's groupid changed to 110
Jun 20 00:09:41 fv-az72-309 systemd[1]: Condition check resulted in Wait for the Ubuntu Core chooser trigger being skipped.
Jun 20 00:09:41 fv-az72-309 rsyslogd[682]: rsyslogd's userid changed to 104
Jun 20 00:09:41 fv-az72-309 systemd[1]: Starting Snap Daemon...
Jun 20 00:09:41 fv-az72-309 rsyslogd[682]: [origin software="rsyslogd" swVersion="8.2001.0" x-pid="682" x-info="https://www.rsyslog.com"] start
Jun 20 00:09:41 fv-az72-309 systemd[1]: Starting OpenBSD Secure Shell server...
Jun 20 00:09:41 fv-az72-309 dbus-daemon[658]: [system] Activating via systemd: service name='org.freedesktop.PolicyKit1' unit='polkit.service' requested by ':1.2' (uid=0 pid=647 comm="/usr/lib/accountsservice/accounts-daemon " label="unconfined")
Jun 20 00:09:41 fv-az72-309 systemd[1]: Starting Resets System Activity Data Collector...
Jun 20 00:09:41 fv-az72-309 systemd[1]: Starting Login Service...
Jun 20 00:09:41 fv-az72-309 systemd[1]: Starting Permit User Sessions...
Jun 20 00:09:41 fv-az72-309 python3[704]: /usr/sbin/waagent:27: DeprecationWarning: the imp module is deprecated in favour of importlib; see the module's documentation for alternative uses
Jun 20 00:09:41 fv-az72-309 python3[704]: import imp
Jun 20 00:09:41 fv-az72-309 systemd[1]: Condition check resulted in Ubuntu Advantage reboot cmds being skipped.
Jun 20 00:09:42 fv-az72-309 apport[649]: ...done.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Starting Disk Manager...
Jun 20 00:09:41 fv-az72-309 systemd[1]: Started Azure Linux Agent.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Started Deferred execution scheduler.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Finished Resets System Activity Data Collector.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Finished Permit User Sessions.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Starting Hold until boot process finishes up...
Jun 20 00:09:41 fv-az72-309 systemd[1]: Starting Terminate Plymouth Boot Screen...
Jun 20 00:09:41 fv-az72-309 systemd[1]: Started System Logging Service.
Jun 20 00:09:41 fv-az72-309 systemd[1]: plymouth-quit-wait.service: Succeeded.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Finished Hold until boot process finishes up.
Jun 20 00:09:41 fv-az72-309 udisksd[703]: udisks daemon version 2.8.4 starting
Jun 20 00:09:41 fv-az72-309 systemd[1]: Started Serial Getty on ttyS0.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Starting Set console scheme...
Jun 20 00:09:41 fv-az72-309 systemd[1]: plymouth-quit.service: Succeeded.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Finished Terminate Plymouth Boot Screen.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Finished Set console scheme.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Created slice system-getty.slice.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Started Getty on tty1.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Reached target Login Prompts.
Jun 20 00:09:41 fv-az72-309 systemd[1]: grub-common.service: Succeeded.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Finished Record successful boot for GRUB.
Jun 20 00:09:41 fv-az72-309 systemd[1]: Starting GRUB failed boot detection...
Jun 20 00:09:42 fv-az72-309 systemd[1]: e2scrub_reap.service: Succeeded.
Jun 20 00:09:42 fv-az72-309 systemd[1]: Finished Remove Stale Online ext4 Metadata Check Snapshots.
Jun 20 00:09:42 fv-az72-309 mono-xsp4[671]: * Starting XSP 4.0 WebServer mono-xsp4
Jun 20 00:09:42 fv-az72-309 mono-xsp4[671]: ...done.
Jun 20 00:09:42 fv-az72-309 systemd[1]: Started LSB: automatic crash report generation.
Jun 20 00:09:42 fv-az72-309 polkitd[681]: started daemon version 0.105 using authority implementation `local' version `0.105'
Jun 20 00:09:42 fv-az72-309 systemd[1]: Started LSB: Mono XSP4.
Jun 20 00:09:42 fv-az72-309 dbus-daemon[658]: [system] Successfully activated service 'org.freedesktop.PolicyKit1'
Jun 20 00:09:42 fv-az72-309 systemd[1]: Started Authorization Manager.
Jun 20 00:09:42 fv-az72-309 systemd[1]: Starting Modem Manager...
Jun 20 00:09:42 fv-az72-309 systemd[1]: grub-initrd-fallback.service: Succeeded.
Jun 20 00:09:42 fv-az72-309 systemd[1]: Finished GRUB failed boot detection.
Jun 20 00:09:42 fv-az72-309 sshd[738]: Server listening on 0.0.0.0 port 22.
Jun 20 00:09:42 fv-az72-309 accounts-daemon[647]: started daemon version 0.6.55
Jun 20 00:09:42 fv-az72-309 sshd[738]: Server listening on :: port 22.
Jun 20 00:09:42 fv-az72-309 systemd[1]: Started Accounts Service.
Jun 20 00:09:42 fv-az72-309 systemd[1]: Started OpenBSD Secure Shell server.
Jun 20 00:09:42 fv-az72-309 systemd[1]: Started chrony, an NTP client/server.
Jun 20 00:09:42 fv-az72-309 systemd[1]: Reached target System Time Synchronized.
Jun 20 00:09:42 fv-az72-309 systemd[1]: Started Periodic ext4 Online Metadata Check for All Filesystems.
Jun 20 00:09:42 fv-az72-309 systemd[1]: Started Discard unused blocks once a week.
Jun 20 00:09:42 fv-az72-309 systemd[1]: Started Daily rotation of log files.
Jun 20 00:09:42 fv-az72-309 systemd[1]: Started Daily man-db regeneration.
Jun 20 00:09:42 fv-az72-309 systemd[1]: Started Message of the Day.
Jun 20 00:09:42 fv-az72-309 systemd[1]: Started Clean PHP session files every 30 mins.
Jun 20 00:09:42 fv-az72-309 systemd[1]: Condition check resulted in Timer to automatically fetch and run repair assertions being skipped.
Jun 20 00:09:42 fv-az72-309 systemd[1]: Reached target Timers.
Jun 20 00:09:42 fv-az72-309 systemd[1]: Starting LSB: Fast standalone full-text SQL search engine...
Jun 20 00:09:42 fv-az72-309 systemd[1]: Starting Discard unused blocks on filesystems from /etc/fstab...
Jun 20 00:09:42 fv-az72-309 systemd[1]: Starting Clean php session files...
Jun 20 00:09:42 fv-az72-309 sphinxsearch[759]: To enable sphinxsearch, edit /etc/default/sphinxsearch and set START=yes
Jun 20 00:09:42 fv-az72-309 python3[704]: 2022-06-20T00:09:42.187027Z INFO Daemon Azure Linux Agent Version:2.2.46
Jun 20 00:09:42 fv-az72-309 python3[704]: 2022-06-20T00:09:42.190939Z INFO Daemon OS: ubuntu 20.04
Jun 20 00:09:42 fv-az72-309 python3[704]: 2022-06-20T00:09:42.194146Z INFO Daemon Python: 3.8.10
Jun 20 00:09:42 fv-az72-309 python3[704]: 2022-06-20T00:09:42.197558Z INFO Daemon CGroups Status: The cgroup filesystem is ready to use
Jun 20 00:09:42 fv-az72-309 python3[704]: 2022-06-20T00:09:42.207982Z INFO Daemon Run daemon
Jun 20 00:09:42 fv-az72-309 systemd[1]: Starting Rotate log files...
Jun 20 00:09:42 fv-az72-309 session-manager-plugin[687]: The Session Manager plugin was installed successfully. Use the AWS CLI to start a session.
Jun 20 00:09:42 fv-az72-309 udisksd[703]: failed to load module mdraid: libbd_mdraid.so.2: cannot open shared object file: No such file or directory
Jun 20 00:09:42 fv-az72-309 systemd[1]: Starting Daily man-db regeneration...
Jun 20 00:09:42 fv-az72-309 systemd[1]: Started LSB: Fast standalone full-text SQL search engine.
Jun 20 00:09:42 fv-az72-309 systemd[1]: session-manager-plugin.service: Succeeded.
Jun 20 00:09:42 fv-az72-309 python3[704]: 2022-06-20T00:09:42.226326Z INFO Daemon cloud-init is enabled: True
Jun 20 00:09:42 fv-az72-309 python3[704]: 2022-06-20T00:09:42.230013Z INFO Daemon Using cloud-init for provisioning
Jun 20 00:09:42 fv-az72-309 python3[704]: 2022-06-20T00:09:42.233982Z INFO Daemon Activate resource disk
Jun 20 00:09:42 fv-az72-309 python3[704]: 2022-06-20T00:09:42.236941Z INFO Daemon Searching gen1 prefix 00000000-0001 or gen2 f8b3781a-1e82-4818-a1c3-63d806ec15bb
Jun 20 00:09:42 fv-az72-309 python3[704]: 2022-06-20T00:09:42.244158Z INFO Daemon Found device: sdb
Jun 20 00:09:42 fv-az72-309 python3[704]: 2022-06-20T00:09:42.253998Z INFO Daemon Resource disk [/dev/sdb1] is already mounted [/mnt]
Jun 20 00:09:42 fv-az72-309 python3[704]: 2022-06-20T00:09:42.269361Z INFO Daemon Enable swap
Jun 20 00:09:42 fv-az72-309 python3[704]: 2022-06-20T00:09:42.278833Z INFO Daemon Create swap file
Jun 20 00:09:42 fv-az72-309 udisksd[703]: Failed to load the 'mdraid' libblockdev plugin
Jun 20 00:09:42 fv-az72-309 ModemManager[751]: <info> ModemManager (version 1.16.6) starting in system bus...
Jun 20 00:09:42 fv-az72-309 python3[704]: 2022-06-20T00:09:42.336492Z INFO Daemon Enabled 4194304KB of swap at /mnt/swapfile
Jun 20 00:09:42 fv-az72-309 kernel: Adding 4194300k swap on /mnt/swapfile. Priority:-2 extents:9 across:4505596k FS
Jun 20 00:09:42 fv-az72-309 python3[704]: 2022-06-20T00:09:42.341954Z INFO Daemon Clean protocol and wireserver endpoint
Jun 20 00:09:42 fv-az72-309 python3[704]: 2022-06-20T00:09:42.350334Z WARNING Daemon VM is provisioned, but the VM unique identifier has changed -- clearing cached state
Jun 20 00:09:42 fv-az72-309 python3[704]: WARNING! Cached DHCP leases will be deleted.
Jun 20 00:09:42 fv-az72-309 python3[704]: 2022-06-20T00:09:42.365193Z INFO Daemon Detect protocol endpoints
Jun 20 00:09:42 fv-az72-309 systemd-logind[696]: New seat seat0.
Jun 20 00:09:42 fv-az72-309 python3[704]: 2022-06-20T00:09:42.373259Z INFO Daemon Clean protocol and wireserver endpoint
Jun 20 00:09:42 fv-az72-309 python3[704]: 2022-06-20T00:09:42.380413Z INFO Daemon WireServer endpoint is not found. Rerun dhcp handler
Jun 20 00:09:42 fv-az72-309 python3[704]: 2022-06-20T00:09:42.387481Z INFO Daemon Test for route to 168.63.129.16
Jun 20 00:09:42 fv-az72-309 python3[704]: 2022-06-20T00:09:42.391519Z INFO Daemon Route to 168.63.129.16 exists
Jun 20 00:09:42 fv-az72-309 python3[704]: 2022-06-20T00:09:42.397597Z INFO Daemon Wire server endpoint:168.63.129.16
Jun 20 00:09:42 fv-az72-309 python3[704]: 2022-06-20T00:09:42.408410Z INFO Daemon Fabric preferred wire protocol version:2015-04-05
Jun 20 00:09:42 fv-az72-309 python3[704]: 2022-06-20T00:09:42.416940Z INFO Daemon Wire protocol version:2012-11-30
Jun 20 00:09:42 fv-az72-309 systemd-logind[696]: Watching system buttons on /dev/input/event0 (Power Button)
Jun 20 00:09:42 fv-az72-309 systemd-logind[696]: Watching system buttons on /dev/input/event2 (AT Translated Set 2 keyboard)
Jun 20 00:09:42 fv-az72-309 systemd-logind[696]: Watching system buttons on /dev/input/event1 (AT Translated Set 2 keyboard)
Jun 20 00:09:42 fv-az72-309 systemd[1]: Started Login Service.
Jun 20 00:09:42 fv-az72-309 python3[704]: 2022-06-20T00:09:42.427715Z INFO Daemon Server preferred version:2015-04-05
Jun 20 00:09:42 fv-az72-309 systemd[1]: Started Disk Manager.
Jun 20 00:09:42 fv-az72-309 udisksd[703]: Acquired the name org.freedesktop.UDisks2 on the system message bus
Jun 20 00:09:42 fv-az72-309 python3[704]: 2022-06-20T00:09:42.672474Z INFO Daemon Found private key matching thumbprint 33D7A199EB5A54B287B160824E7C8404A806D41B
Jun 20 00:09:42 fv-az72-309 python3[704]: 2022-06-20T00:09:42.689458Z INFO Daemon Provisioning already completed, skipping.
Jun 20 00:09:42 fv-az72-309 python3[704]: 2022-06-20T00:09:42.694154Z INFO Daemon RDMA capabilities are not enabled, skipping
Jun 20 00:09:42 fv-az72-309 python3[704]: 2022-06-20T00:09:42.718900Z INFO Daemon Determined Agent WALinuxAgent-2.7.1.0 to be the latest agent
Jun 20 00:09:42 fv-az72-309 networkd-dispatcher[674]: No valid path found for iwconfig
Jun 20 00:09:42 fv-az72-309 networkd-dispatcher[674]: No valid path found for iw
Jun 20 00:09:42 fv-az72-309 systemd[1]: Started Modem Manager.
Jun 20 00:09:42 fv-az72-309 snapd[691]: AppArmor status: apparmor is enabled and all features are available
Jun 20 00:09:42 fv-az72-309 rsyslogd[682]: [origin software="rsyslogd" swVersion="8.2001.0" x-pid="682" x-info="https://www.rsyslog.com"] rsyslogd was HUPed
Jun 20 00:09:42 fv-az72-309 systemd[1]: tmp-snap.rootfs_xmixJk.mount: Succeeded.
Jun 20 00:09:42 fv-az72-309 systemd[1]: logrotate.service: Succeeded.
Jun 20 00:09:42 fv-az72-309 systemd[1]: Finished Rotate log files.
Jun 20 00:09:42 fv-az72-309 systemd[1]: Started Dispatcher daemon for systemd-networkd.
Jun 20 00:09:43 fv-az72-309 fstrim[761]: /mnt: 9.7 GiB (10361049088 bytes) trimmed on /dev/disk/cloud/azure_resource-part1
Jun 20 00:09:43 fv-az72-309 fstrim[761]: /boot/efi: 99.2 MiB (103973888 bytes) trimmed on /dev/sda15
Jun 20 00:09:43 fv-az72-309 fstrim[761]: /: 28.1 GiB (30198099968 bytes) trimmed on /dev/sda1
Jun 20 00:09:43 fv-az72-309 systemd[1]: fstrim.service: Succeeded.
Jun 20 00:09:43 fv-az72-309 systemd[1]: Finished Discard unused blocks on filesystems from /etc/fstab.
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.179944401Z" level=info msg="starting containerd" revision=a17ec496a95e55601607ca50828147e8ccaeebf1 version=1.5.13+azure-1
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.302251Z INFO ExtHandler ExtHandler The agent will now check for updates and then will process extensions. Output to /dev/console will be suspended during those operations.
Jun 20 00:09:43 fv-az72-309 php8.1[818]: DIGEST-MD5 common mech free
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.406511055Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.451579Z INFO ExtHandler ExtHandler Agent WALinuxAgent-2.7.1.0 is running as the goal state agent
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.451819Z INFO ExtHandler ExtHandler WireServer endpoint 168.63.129.16 read from file
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.451898Z INFO ExtHandler ExtHandler Wire server endpoint:168.63.129.16
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.454641191Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1
Jun 20 00:09:43 fv-az72-309 kernel: aufs 5.x-rcN-20210809
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.471670046Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.471855Z INFO ExtHandler ExtHandler HostGAPlugin version: 1.0.8.124
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.472369077Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.472410Z INFO ExtHandler ExtHandler Fetched new vmSettings [correlation ID: 6fb7f8f8-f91a-4429-ab63-9fab2c1e8be9 New eTag: 14923850863347568451]
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.472526Z INFO ExtHandler ExtHandler Fetching goal state [incarnation 1]
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.472720293Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.472881100Z" level=warning msg="failed to load plugin io.containerd.snapshotter.v1.devmapper" error="devmapper not configured"
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.473033307Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.474079553Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.474682780Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.475538618Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.475888234Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.476514761Z" level=warning msg="could not use snapshotter devmapper in metadata plugin" error="devmapper not configured"
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.476707770Z" level=info msg="metadata content store policy set" policy=shared
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.486732115Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.486914223Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.490472781Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.490686490Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.490810996Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.490913200Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.491011805Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.491108009Z" level=info msg="loading plugin \"io.containerd.service.v1.leases-service\"..." type=io.containerd.service.v1
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.491201813Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.491306818Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.491401422Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.491593831Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.491756338Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.492188057Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.492393266Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.492556173Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.492655378Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.492759482Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.492992893Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.493099297Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.493189501Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.493284406Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.493401011Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.493495215Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.495787517Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.495956024Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.496069929Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.496175234Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.496482347Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec:} UntrustedWorkloadRuntime:{Type: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[BinaryName: CriuImagePath: CriuPath: CriuWorkPath: IoGid:0 IoUid:0 NoNewKeyring:false NoPivotRoot:false Root: ShimCgroup: SystemdCgroup:false] PrivilegedWithoutHostDevices:false BaseRuntimeSpec:}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginConfTemplate:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:false SelinuxCategoryRange:1024 SandboxImage:k8s.gcr.io/pause:3.5 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}"
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.510855385Z" level=info msg="Connect containerd service"
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.511033893Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\""
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.533932Z INFO ExtHandler ExtHandler Found private key matching thumbprint 33D7A199EB5A54B287B160824E7C8404A806D41B
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.540250990Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config"
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.541337838Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.543173819Z" level=info msg="Start subscribing containerd event"
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.547925130Z" level=info msg="Start recovering state"
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.548150240Z" level=info msg="Start event monitor"
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.548250845Z" level=info msg="Start snapshots syncer"
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.548436453Z" level=info msg="Start cni network conf syncer"
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.548539457Z" level=info msg="Start streaming server"
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.549093582Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.549324792Z" level=info msg=serving... address=/run/containerd/containerd.sock
Jun 20 00:09:43 fv-az72-309 containerd[707]: time="2022-06-20T00:09:43.549677108Z" level=info msg="containerd successfully booted in 0.373259s"
Jun 20 00:09:43 fv-az72-309 systemd[1]: Started containerd container runtime.
Jun 20 00:09:43 fv-az72-309 systemd[1]: Starting Docker Application Container Engine...
Jun 20 00:09:43 fv-az72-309 provisioner[683]: Argument: --agentdirectory
Jun 20 00:09:43 fv-az72-309 provisioner[683]: Value: /home/runner/runners
Jun 20 00:09:43 fv-az72-309 provisioner[683]: Argument: --settings
Jun 20 00:09:43 fv-az72-309 provisioner[683]: Value: /opt/runner/provisioner/.settings
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.587516Z INFO ExtHandler ExtHandler Fetch goal state completed
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.611034Z INFO ExtHandler ExtHandler Distro: ubuntu-20.04; OSUtil: Ubuntu18OSUtil; AgentService: walinuxagent; Python: 3.8.10; systemd: True; LISDrivers: name: hv_vmbus
Jun 20 00:09:43 fv-az72-309 python3[853]: ; logrotate: logrotate 3.14.0;
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.617028Z INFO ExtHandler ExtHandler WALinuxAgent-2.7.1.0 running as process 853
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.625655Z INFO ExtHandler ExtHandler [CGI] systemd version: systemd 245 (245.4-4ubuntu3.17)
Jun 20 00:09:43 fv-az72-309 python3[853]: +PAM +AUDIT +SELINUX +IMA +APPARMOR +SMACK +SYSVINIT +UTMP +LIBCRYPTSETUP +GCRYPT +GNUTLS +ACL +XZ +LZ4 +SECCOMP +BLKID +ELFUTILS +KMOD +IDN2 -IDN +PCRE2 default-hierarchy=hybrid
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.640573Z INFO ExtHandler ExtHandler The CPU cgroup controller is mounted at /sys/fs/cgroup/cpu,cpuacct
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.640794Z INFO ExtHandler ExtHandler The memory cgroup controller is mounted at /sys/fs/cgroup/memory
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.646869Z INFO ExtHandler ExtHandler [CGI] cgroups v2 mounted at /sys/fs/cgroup/unified. Controllers: []
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.665292Z INFO ExtHandler ExtHandler [CGI] CPUAccounting: yes
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.674660Z INFO ExtHandler ExtHandler [CGI] CPUQuota: 750ms
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.684464Z INFO ExtHandler ExtHandler [CGI] MemoryAccounting: yes
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.685211Z INFO ExtHandler ExtHandler [CGI] Agent CPU cgroup: /sys/fs/cgroup/cpu,cpuacct/azure.slice/walinuxagent.service
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.685570Z INFO ExtHandler ExtHandler [CGI] Ensuring the agent's CPUQuota is 75%
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.686616Z INFO ExtHandler ExtHandler Started tracking cgroup walinuxagent.service [/sys/fs/cgroup/cpu,cpuacct/azure.slice/walinuxagent.service]
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.686713Z INFO ExtHandler ExtHandler [CGI] Agent cgroups enabled: True
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.687622Z INFO ExtHandler ExtHandler Starting setup for Persistent firewall rules
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.697371Z INFO ExtHandler ExtHandler Firewalld service not running/unavailable, trying to set up walinuxagent-network-setup.service
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.697704Z INFO ExtHandler ExtHandler Successfully updated the Binary file /var/lib/waagent/waagent-network-setup.py for firewall setup
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.708695Z INFO ExtHandler ExtHandler Unit file version matches with expected version: 1.2, not overwriting unit file
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.708868Z INFO ExtHandler ExtHandler Service: walinuxagent-network-setup.service already enabled. No change needed.
Jun 20 00:09:43 fv-az72-309 provisioner[683]: Starting service provider configuration
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.732779Z INFO ExtHandler ExtHandler Logs from the walinuxagent-network-setup.service since system boot:
Jun 20 00:09:43 fv-az72-309 python3[853]: -- Logs begin at Thu 2022-06-16 07:26:56 UTC, end at Mon 2022-06-20 00:09:43 UTC. --
Jun 20 00:09:43 fv-az72-309 python3[853]: Jun 20 00:09:35 fv-az72-309 python3[217]: Setting up firewall for the WALinux Agent with args: {'dst_ip': '168.63.129.16', 'uid': '0', 'wait': '-w'}
Jun 20 00:09:43 fv-az72-309 python3[853]: Jun 20 00:09:35 fv-az72-309 python3[217]: Successfully set the firewall rules
Jun 20 00:09:43 fv-az72-309 python3[853]: Jun 20 00:09:35 fv-az72-309 systemd[1]: walinuxagent-network-setup.service: Succeeded.
Jun 20 00:09:43 fv-az72-309 python3[853]: Jun 20 00:09:35 fv-az72-309 systemd[1]: Finished Setup network rules for WALinuxAgent.
Jun 20 00:09:43 fv-az72-309 python3[853]: Jun 20 00:09:37 fv-az72-309 systemd[1]: Starting Setup network rules for WALinuxAgent...
Jun 20 00:09:43 fv-az72-309 python3[853]: Jun 20 00:09:37 fv-az72-309 python3[522]: Setting up firewall for the WALinux Agent with args: {'dst_ip': '168.63.129.16', 'uid': '0', 'wait': '-w'}
Jun 20 00:09:43 fv-az72-309 python3[853]: Jun 20 00:09:37 fv-az72-309 python3[522]: Successfully set the firewall rules
Jun 20 00:09:43 fv-az72-309 python3[853]: Jun 20 00:09:37 fv-az72-309 systemd[1]: walinuxagent-network-setup.service: Succeeded.
Jun 20 00:09:43 fv-az72-309 python3[853]: Jun 20 00:09:37 fv-az72-309 systemd[1]: Finished Setup network rules for WALinuxAgent.
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.733627Z INFO ExtHandler ExtHandler Persistent firewall rules setup successfully
Jun 20 00:09:43 fv-az72-309 systemd[1]: Started The PHP 8.1 FastCGI Process Manager.
Jun 20 00:09:43 fv-az72-309 systemd[1]: Started The PHP 8.0 FastCGI Process Manager.
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.764577Z INFO ExtHandler ExtHandler Not setting the firewall rule to allow DNS TCP request to wireserver for a non root user since it already exists
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.765130Z INFO ExtHandler ExtHandler Checking if log collection is allowed at this time [True]. All three conditions must be met: configuration enabled [True], cgroups enabled [True], python supported: [True]
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.766365Z INFO ExtHandler ExtHandler Starting env monitor service.
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.767271Z INFO MonitorHandler ExtHandler WireServer endpoint 168.63.129.16 read from file
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.767379Z INFO MonitorHandler ExtHandler Wire server endpoint:168.63.129.16
Jun 20 00:09:43 fv-az72-309 systemd[1]: Started The PHP 7.4 FastCGI Process Manager.
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.767916Z INFO MonitorHandler ExtHandler Monitor.NetworkConfigurationChanges is disabled.
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.768113Z INFO MonitorHandler ExtHandler Routing table from /proc/net/route:
Jun 20 00:09:43 fv-az72-309 python3[853]: Iface Destination Gateway Flags RefCnt Use Metric Mask MTU Window IRTT
Jun 20 00:09:43 fv-az72-309 python3[853]: eth0 00000000 0100010A 0003 0 0 100 00000000 0 0 0
Jun 20 00:09:43 fv-az72-309 python3[853]: eth0 0000010A 00000000 0001 0 0 0 0000FFFF 0 0 0
Jun 20 00:09:43 fv-az72-309 python3[853]: eth0 10813FA8 0100010A 0007 0 0 100 FFFFFFFF 0 0 0
Jun 20 00:09:43 fv-az72-309 python3[853]: eth0 FEA9FEA9 0100010A 0007 0 0 100 FFFFFFFF 0 0 0
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.770476Z INFO ExtHandler ExtHandler Start SendTelemetryHandler service.
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.772753Z INFO EnvHandler ExtHandler WireServer endpoint 168.63.129.16 read from file
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.772934Z INFO EnvHandler ExtHandler Wire server endpoint:168.63.129.16
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.773837Z INFO EnvHandler ExtHandler Configure routes
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.773972Z INFO EnvHandler ExtHandler Gateway:None
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.774036Z INFO EnvHandler ExtHandler Routes:None
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.782692Z INFO ExtHandler ExtHandler Start Extension Telemetry service.
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.782350Z INFO SendTelemetryHandler ExtHandler Successfully started the SendTelemetryHandler thread
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.784319Z INFO TelemetryEventsCollector ExtHandler Extension Telemetry pipeline enabled: True
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.787686Z INFO TelemetryEventsCollector ExtHandler Successfully started the TelemetryEventsCollector thread
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.795361Z INFO ExtHandler ExtHandler Goal State Period: 6 sec. This indicates how often the agent checks for new goal states and reports status.
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.809158Z INFO MonitorHandler ExtHandler Network interfaces:
Jun 20 00:09:43 fv-az72-309 python3[853]: Executing ['ip', '-a', '-o', 'link']:
Jun 20 00:09:43 fv-az72-309 python3[853]: 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN mode DEFAULT group default qlen 1000\ link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
Jun 20 00:09:43 fv-az72-309 python3[853]: 2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP mode DEFAULT group default qlen 1000\ link/ether 00:0d:3a:13:05:24 brd ff:ff:ff:ff:ff:ff
Jun 20 00:09:43 fv-az72-309 python3[853]: Executing ['ip', '-4', '-a', '-o', 'address']:
Jun 20 00:09:43 fv-az72-309 python3[853]: 1: lo inet 127.0.0.1/8 scope host lo\ valid_lft forever preferred_lft forever
Jun 20 00:09:43 fv-az72-309 python3[853]: 2: eth0 inet 10.1.0.17/16 brd 10.1.255.255 scope global eth0\ valid_lft forever preferred_lft forever
Jun 20 00:09:43 fv-az72-309 python3[853]: Executing ['ip', '-6', '-a', '-o', 'address']:
Jun 20 00:09:43 fv-az72-309 python3[853]: 1: lo inet6 ::1/128 scope host \ valid_lft forever preferred_lft forever
Jun 20 00:09:43 fv-az72-309 python3[853]: 2: eth0 inet6 fe80::20d:3aff:fe13:524/64 scope link \ valid_lft forever preferred_lft forever
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.821550Z INFO ExtHandler ExtHandler Checking for agent updates (family: Prod)
Jun 20 00:09:43 fv-az72-309 php8.1[918]: DIGEST-MD5 common mech free
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.880447Z INFO ExtHandler ExtHandler ProcessExtensionsInGoalState started [Incarnation: 1; Activity Id: 48fbd77a-0509-46ad-81a1-1394da8eb4b4; Correlation Id: 16378185-6322-4308-98a6-03c89673c297; GS Creation Time: 2022-06-20T00:07:55.901245Z]
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.884305Z INFO EnvHandler ExtHandler Set block dev timeout: sdb with timeout: 300
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.884476Z INFO EnvHandler ExtHandler Set block dev timeout: sda with timeout: 300
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.933692Z INFO ExtHandler [Microsoft.Azure.Extensions.CustomScript-2.1.6] Target handler state: enabled [incarnation 1]
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.937169Z INFO ExtHandler [Microsoft.Azure.Extensions.CustomScript-2.1.6] [Enable] current handler state is: notinstalled
Jun 20 00:09:43 fv-az72-309 python3[853]: 2022-06-20T00:09:43.937522Z INFO ExtHandler [Microsoft.Azure.Extensions.CustomScript-2.1.6] Downloading extension package: https://umsa4s5cv5qvrkpxlmjq.blob.core.windows.net/5237dd14-0aad-f051-0fad-1e33e1b63091/5237dd14-0aad-f051-0fad-1e33e1b63091_2.1.6.zip
Jun 20 00:09:44 fv-az72-309 python3[853]: 2022-06-20T00:09:44.054361Z INFO ExtHandler [Microsoft.Azure.Extensions.CustomScript-2.1.6] Unzipping extension package: /var/lib/waagent/Microsoft.Azure.Extensions.CustomScript__2.1.6.zip
Jun 20 00:09:44 fv-az72-309 php8.1[976]: DIGEST-MD5 common mech free
Jun 20 00:09:44 fv-az72-309 python3[853]: 2022-06-20T00:09:44.226239Z INFO ExtHandler [Microsoft.Azure.Extensions.CustomScript-2.1.6] Initializing extension Microsoft.Azure.Extensions.CustomScript-2.1.6
Jun 20 00:09:44 fv-az72-309 python3[853]: 2022-06-20T00:09:44.228256Z INFO ExtHandler ExtHandler [CGI] Created /lib/systemd/system/azure-vmextensions-Microsoft.Azure.Extensions.CustomScript_2.1.6.slice
Jun 20 00:09:44 fv-az72-309 python3[853]: 2022-06-20T00:09:44.229004Z INFO ExtHandler [Microsoft.Azure.Extensions.CustomScript-2.1.6] Update settings file: 99.settings
Jun 20 00:09:44 fv-az72-309 python3[853]: 2022-06-20T00:09:44.229613Z INFO ExtHandler [Microsoft.Azure.Extensions.CustomScript-2.1.6] Install extension [bin/custom-script-shim install]
Jun 20 00:09:44 fv-az72-309 python3[853]: 2022-06-20T00:09:44.229958Z INFO ExtHandler [Microsoft.Azure.Extensions.CustomScript-2.1.6] Executing command: /var/lib/waagent/Microsoft.Azure.Extensions.CustomScript-2.1.6/bin/custom-script-shim install with environment variables: {"AZURE_GUEST_AGENT_UNINSTALL_CMD_EXIT_CODE": "NOT_RUN", "AZURE_GUEST_AGENT_EXTENSION_PATH": "/var/lib/waagent/Microsoft.Azure.Extensions.CustomScript-2.1.6", "AZURE_GUEST_AGENT_EXTENSION_VERSION": "2.1.6", "AZURE_GUEST_AGENT_WIRE_PROTOCOL_ADDRESS": "168.63.129.16", "ConfigSequenceNumber": "99", "AZURE_GUEST_AGENT_EXTENSION_SUPPORTED_FEATURES": "[{\"Key\": \"ExtensionTelemetryPipeline\", \"Value\": \"1.0\"}]"}
Jun 20 00:09:44 fv-az72-309 python3[853]: 2022-06-20T00:09:44.236342Z INFO ExtHandler ExtHandler Started extension in unit 'install_0b3bf6d6-56e1-417f-85ed-1149e224521c.scope'
Jun 20 00:09:44 fv-az72-309 python3[853]: 2022-06-20T00:09:44.237055Z INFO ExtHandler ExtHandler Started tracking cgroup Microsoft.Azure.Extensions.CustomScript-2.1.6 [/sys/fs/cgroup/cpu,cpuacct/azure.slice/azure-vmextensions.slice/azure-vmextensions-Microsoft.Azure.Extensions.CustomScript_2.1.6.slice]
Jun 20 00:09:44 fv-az72-309 systemd[1]: Created slice Slice for Azure VM Extensions.
Jun 20 00:09:44 fv-az72-309 systemd[1]: Created slice Slice for Azure VM extension Microsoft.Azure.Extensions.CustomScript-2.1.6.
Jun 20 00:09:44 fv-az72-309 systemd[1]: Started /var/lib/waagent/Microsoft.Azure.Extensions.CustomScript-2.1.6/bin/custom-script-shim install.
Jun 20 00:09:44 fv-az72-309 systemd[1]: install_0b3bf6d6-56e1-417f-85ed-1149e224521c.scope: Succeeded.
Jun 20 00:09:44 fv-az72-309 php8.0[994]: DIGEST-MD5 common mech free
Jun 20 00:09:44 fv-az72-309 dockerd[913]: time="2022-06-20T00:09:44.525211296Z" level=info msg="Starting up"
Jun 20 00:09:44 fv-az72-309 dockerd[913]: time="2022-06-20T00:09:44.527941717Z" level=info msg="detected 127.0.0.53 nameserver, assuming systemd-resolved, so using resolv.conf: /run/systemd/resolve/resolv.conf"
Jun 20 00:09:44 fv-az72-309 audit[1019]: AVC apparmor="STATUS" operation="profile_load" profile="unconfined" name="docker-default" pid=1019 comm="apparmor_parser"
Jun 20 00:09:44 fv-az72-309 kernel: kauditd_printk_skb: 22 callbacks suppressed
Jun 20 00:09:44 fv-az72-309 kernel: audit: type=1400 audit(1655683784.674:33): apparmor="STATUS" operation="profile_load" profile="unconfined" name="docker-default" pid=1019 comm="apparmor_parser"
Jun 20 00:09:44 fv-az72-309 dockerd[913]: time="2022-06-20T00:09:44.699789443Z" level=info msg="parsed scheme: \"unix\"" module=grpc
Jun 20 00:09:44 fv-az72-309 dockerd[913]: time="2022-06-20T00:09:44.700101757Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc
Jun 20 00:09:44 fv-az72-309 dockerd[913]: time="2022-06-20T00:09:44.700257264Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/containerd/containerd.sock <nil> 0 <nil>}] <nil> <nil>}" module=grpc
Jun 20 00:09:44 fv-az72-309 dockerd[913]: time="2022-06-20T00:09:44.700373469Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc
Jun 20 00:09:44 fv-az72-309 php8.0[1026]: DIGEST-MD5 common mech free
Jun 20 00:09:44 fv-az72-309 dockerd[913]: time="2022-06-20T00:09:44.753493726Z" level=info msg="parsed scheme: \"unix\"" module=grpc
Jun 20 00:09:44 fv-az72-309 dockerd[913]: time="2022-06-20T00:09:44.753533528Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc
Jun 20 00:09:44 fv-az72-309 dockerd[913]: time="2022-06-20T00:09:44.753555729Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/containerd/containerd.sock <nil> 0 <nil>}] <nil> <nil>}" module=grpc
Jun 20 00:09:44 fv-az72-309 dockerd[913]: time="2022-06-20T00:09:44.753571029Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc
Jun 20 00:09:44 fv-az72-309 systemd[1]: var-lib-docker-overlay2-check\x2doverlayfs\x2dsupport4125447870-merged.mount: Succeeded.
Jun 20 00:09:44 fv-az72-309 dockerd[913]: time="2022-06-20T00:09:44.902914356Z" level=info msg="[graphdriver] using prior storage driver: overlay2"
Jun 20 00:09:44 fv-az72-309 php8.0[1040]: DIGEST-MD5 common mech free
Jun 20 00:09:45 fv-az72-309 provisioner[683]: Done configuring service provider
Jun 20 00:09:45 fv-az72-309 dockerd[913]: time="2022-06-20T00:09:45.139899372Z" level=warning msg="Your kernel does not support CPU realtime scheduler"
Jun 20 00:09:45 fv-az72-309 dockerd[913]: time="2022-06-20T00:09:45.140218986Z" level=warning msg="Your kernel does not support cgroup blkio weight"
Jun 20 00:09:45 fv-az72-309 dockerd[913]: time="2022-06-20T00:09:45.140353192Z" level=warning msg="Your kernel does not support cgroup blkio weight_device"
Jun 20 00:09:45 fv-az72-309 dockerd[913]: time="2022-06-20T00:09:45.141350037Z" level=info msg="Loading containers: start."
Jun 20 00:09:45 fv-az72-309 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this.
Jun 20 00:09:45 fv-az72-309 kernel: Bridge firewalling registered
Jun 20 00:09:45 fv-az72-309 systemd[1]: man-db.service: Succeeded.
Jun 20 00:09:45 fv-az72-309 systemd[1]: Finished Daily man-db regeneration.
Jun 20 00:09:45 fv-az72-309 php7.4[1053]: DIGEST-MD5 common mech free
Jun 20 00:09:45 fv-az72-309 php7.4[1076]: DIGEST-MD5 common mech free
Jun 20 00:09:45 fv-az72-309 kernel: Initializing XFRM netlink socket
Jun 20 00:09:45 fv-az72-309 systemd-udevd[813]: Using default interface naming scheme 'v245'.
Jun 20 00:09:45 fv-az72-309 systemd-udevd[813]: ethtool: autonegotiation is unset or enabled, the speed and duplex are not writable.
Jun 20 00:09:45 fv-az72-309 systemd-networkd[532]: rtnl: received neighbor for link '3' we don't know about, ignoring.
Jun 20 00:09:45 fv-az72-309 systemd-networkd[532]: rtnl: received neighbor for link '3' we don't know about, ignoring.
Jun 20 00:09:45 fv-az72-309 systemd-networkd[532]: rtnl: received neighbor for link '3' we don't know about, ignoring.
Jun 20 00:09:45 fv-az72-309 systemd-networkd[532]: rtnl: received neighbor for link '3' we don't know about, ignoring.
Jun 20 00:09:45 fv-az72-309 networkd-dispatcher[674]: WARNING:Unknown index 3 seen, reloading interface list
Jun 20 00:09:45 fv-az72-309 php7.4[1121]: DIGEST-MD5 common mech free
Jun 20 00:09:45 fv-az72-309 systemd[1]: phpsessionclean.service: Succeeded.
Jun 20 00:09:45 fv-az72-309 systemd[1]: Finished Clean php session files.
Jun 20 00:09:45 fv-az72-309 systemd-networkd[532]: docker0: Link UP
Jun 20 00:09:45 fv-az72-309 dockerd[913]: time="2022-06-20T00:09:45.733696421Z" level=info msg="Default bridge (docker0) is assigned with an IP address 172.17.0.0/16. Daemon option --bip can be used to set a preferred IP address"
Jun 20 00:09:45 fv-az72-309 dockerd[913]: time="2022-06-20T00:09:45.846060307Z" level=info msg="Loading containers: done."
Jun 20 00:09:45 fv-az72-309 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck317930894-merged.mount: Succeeded.
Jun 20 00:09:45 fv-az72-309 dockerd[913]: time="2022-06-20T00:09:45.959672249Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2
Jun 20 00:09:45 fv-az72-309 dockerd[913]: time="2022-06-20T00:09:45.959898559Z" level=info msg="Docker daemon" commit=f756502055d2e36a84f2068e6620bea5ecf09058 graphdriver(s)=overlay2 version=20.10.16+azure-2
Jun 20 00:09:45 fv-az72-309 dockerd[913]: time="2022-06-20T00:09:45.960376280Z" level=info msg="Daemon has completed initialization"
Jun 20 00:09:45 fv-az72-309 systemd[1]: Started Docker Application Container Engine.
Jun 20 00:09:46 fv-az72-309 dockerd[913]: time="2022-06-20T00:09:46.009683668Z" level=info msg="API listen on /run/docker.sock"
Jun 20 00:09:46 fv-az72-309 python3[853]: 2022-06-20T00:09:46.238666Z INFO ExtHandler [Microsoft.Azure.Extensions.CustomScript-2.1.6] Command: bin/custom-script-shim install
Jun 20 00:09:46 fv-az72-309 python3[853]: [stdout]
Jun 20 00:09:46 fv-az72-309 python3[853]: + /var/lib/waagent/Microsoft.Azure.Extensions.CustomScript-2.1.6/bin/custom-script-extension install
Jun 20 00:09:46 fv-az72-309 python3[853]: time=2022-06-20T00:09:44Z version=v2.1.6/git@fc181d8-dirty operation=install seq=99 event=start
Jun 20 00:09:46 fv-az72-309 python3[853]: time=2022-06-20T00:09:44Z version=v2.1.6/git@fc181d8-dirty operation=install seq=99 status="not reported for operation (by design)"
Jun 20 00:09:46 fv-az72-309 python3[853]: time=2022-06-20T00:09:44Z version=v2.1.6/git@fc181d8-dirty operation=install seq=99 event="migrate to mrseq" error="Can't find out seqnum from /var/lib/waagent/Microsoft.Azure.Extensions.CustomScript-2.1.6/status, not enough files"
Jun 20 00:09:46 fv-az72-309 python3[853]: time=2022-06-20T00:09:44Z version=v2.1.6/git@fc181d8-dirty operation=install seq=99 event="created data dir" path=/var/lib/waagent/custom-script
Jun 20 00:09:46 fv-az72-309 python3[853]: time=2022-06-20T00:09:44Z version=v2.1.6/git@fc181d8-dirty operation=install seq=99 event=installed
Jun 20 00:09:46 fv-az72-309 python3[853]: time=2022-06-20T00:09:44Z version=v2.1.6/git@fc181d8-dirty operation=install seq=99 status="not reported for operation (by design)"
Jun 20 00:09:46 fv-az72-309 python3[853]: time=2022-06-20T00:09:44Z version=v2.1.6/git@fc181d8-dirty operation=install seq=99 event=end
Jun 20 00:09:46 fv-az72-309 python3[853]: [stderr]
Jun 20 00:09:46 fv-az72-309 python3[853]: Running scope as unit: install_0b3bf6d6-56e1-417f-85ed-1149e224521c.scope
Jun 20 00:09:46 fv-az72-309 python3[853]: 2022-06-20T00:09:46.239839Z INFO ExtHandler [Microsoft.Azure.Extensions.CustomScript-2.1.6] Requested extension state: enabled
Jun 20 00:09:46 fv-az72-309 python3[853]: 2022-06-20T00:09:46.240097Z INFO ExtHandler [Microsoft.Azure.Extensions.CustomScript-2.1.6] Enable extension: [bin/custom-script-shim enable]
Jun 20 00:09:46 fv-az72-309 python3[853]: 2022-06-20T00:09:46.240389Z INFO ExtHandler [Microsoft.Azure.Extensions.CustomScript-2.1.6] Executing command: /var/lib/waagent/Microsoft.Azure.Extensions.CustomScript-2.1.6/bin/custom-script-shim enable with environment variables: {"AZURE_GUEST_AGENT_UNINSTALL_CMD_EXIT_CODE": "NOT_RUN", "AZURE_GUEST_AGENT_EXTENSION_PATH": "/var/lib/waagent/Microsoft.Azure.Extensions.CustomScript-2.1.6", "AZURE_GUEST_AGENT_EXTENSION_VERSION": "2.1.6", "AZURE_GUEST_AGENT_WIRE_PROTOCOL_ADDRESS": "168.63.129.16", "ConfigSequenceNumber": "99", "AZURE_GUEST_AGENT_EXTENSION_SUPPORTED_FEATURES": "[{\"Key\": \"ExtensionTelemetryPipeline\", \"Value\": \"1.0\"}]"}
Jun 20 00:09:46 fv-az72-309 python3[853]: 2022-06-20T00:09:46.244060Z INFO ExtHandler ExtHandler Started extension in unit 'enable_a078213d-e1c9-4073-aa18-bdb3cc076462.scope'
Jun 20 00:09:46 fv-az72-309 systemd[1]: Started /var/lib/waagent/Microsoft.Azure.Extensions.CustomScript-2.1.6/bin/custom-script-shim enable.
Jun 20 00:09:46 fv-az72-309 systemd[1]: enable_a078213d-e1c9-4073-aa18-bdb3cc076462.scope: Succeeded.
Jun 20 00:09:46 fv-az72-309 lxd.activate[689]: => Starting LXD activation
Jun 20 00:09:46 fv-az72-309 lxd.activate[689]: ==> Loading snap configuration
Jun 20 00:09:46 fv-az72-309 lxd.activate[689]: ==> Checking for socket activation support
Jun 20 00:09:46 fv-az72-309 systemd[1]: dmesg.service: Succeeded.
Jun 20 00:09:46 fv-az72-309 snapd[691]: AppArmor status: apparmor is enabled and all features are available
Jun 20 00:09:46 fv-az72-309 snapd[691]: overlord.go:263: Acquiring state lock file
Jun 20 00:09:46 fv-az72-309 snapd[691]: overlord.go:268: Acquired state lock file
Jun 20 00:09:46 fv-az72-309 snapd[691]: daemon.go:247: started snapd/2.56 (series 16; classic) ubuntu/20.04 (amd64) linux/5.13.0-1029-azure.
Jun 20 00:09:46 fv-az72-309 kernel: loop3: detected capacity change from 0 to 8
Jun 20 00:09:46 fv-az72-309 systemd[1]: tmp-syscheck\x2dmountpoint\x2d169127102.mount: Succeeded.
Jun 20 00:09:47 fv-az72-309 snapd[691]: daemon.go:340: adjusting startup timeout by 45s (pessimistic estimate of 30s plus 5s per snap)
Jun 20 00:09:47 fv-az72-309 systemd[1]: Started Snap Daemon.
Jun 20 00:09:47 fv-az72-309 systemd[1]: Starting Wait until snapd is fully seeded...
Jun 20 00:09:47 fv-az72-309 dbus-daemon[658]: [system] Activating via systemd: service name='org.freedesktop.timedate1' unit='dbus-org.freedesktop.timedate1.service' requested by ':1.11' (uid=0 pid=691 comm="/usr/lib/snapd/snapd " label="unconfined")
Jun 20 00:09:47 fv-az72-309 systemd[1]: Starting Time & Date Service...
Jun 20 00:09:47 fv-az72-309 dbus-daemon[658]: [system] Successfully activated service 'org.freedesktop.timedate1'
Jun 20 00:09:47 fv-az72-309 systemd[1]: Started Time & Date Service.
Jun 20 00:09:47 fv-az72-309 systemd[1]: Finished Wait until snapd is fully seeded.
Jun 20 00:09:47 fv-az72-309 systemd[1]: Starting Apply the settings specified in cloud-config...
Jun 20 00:09:47 fv-az72-309 systemd[1]: Condition check resulted in Auto import assertions from block devices being skipped.
Jun 20 00:09:47 fv-az72-309 lxd.activate[689]: ==> Setting LXD socket ownership
Jun 20 00:09:47 fv-az72-309 lxd.activate[689]: ==> LXD never started on this system, no need to start it now
Jun 20 00:09:47 fv-az72-309 systemd[1]: snap.lxd.activate.service: Succeeded.
Jun 20 00:09:47 fv-az72-309 systemd[1]: Finished Service for snap application lxd.activate.
Jun 20 00:09:47 fv-az72-309 cloud-init[1367]: Cloud-init v. 22.2-0ubuntu1~20.04.2 running 'modules:config' at Mon, 20 Jun 2022 00:09:47 +0000. Up 15.74 seconds.
Jun 20 00:09:47 fv-az72-309 systemd[1]: Finished Apply the settings specified in cloud-config.
Jun 20 00:09:47 fv-az72-309 systemd[1]: Starting Write warning to Azure ephemeral disk...
Jun 20 00:09:47 fv-az72-309 systemd[1]: Finished Write warning to Azure ephemeral disk.
Jun 20 00:09:47 fv-az72-309 systemd[1]: Reached target Multi-User System.
Jun 20 00:09:47 fv-az72-309 systemd[1]: Reached target Graphical Interface.
Jun 20 00:09:47 fv-az72-309 systemd[1]: Starting Execute cloud user/final scripts...
Jun 20 00:09:47 fv-az72-309 systemd[1]: Starting Update UTMP about System Runlevel Changes...
Jun 20 00:09:48 fv-az72-309 systemd[1]: systemd-update-utmp-runlevel.service: Succeeded.
Jun 20 00:09:48 fv-az72-309 systemd[1]: Finished Update UTMP about System Runlevel Changes.
Jun 20 00:09:48 fv-az72-309 python3[853]: 2022-06-20T00:09:48.247268Z INFO ExtHandler [Microsoft.Azure.Extensions.CustomScript-2.1.6] Command: bin/custom-script-shim enable
Jun 20 00:09:48 fv-az72-309 python3[853]: [stdout]
Jun 20 00:09:48 fv-az72-309 python3[853]: Writing a placeholder status file indicating progress before forking: /var/lib/waagent/Microsoft.Azure.Extensions.CustomScript-2.1.6/status/99.status
Jun 20 00:09:48 fv-az72-309 python3[853]: + nohup /var/lib/waagent/Microsoft.Azure.Extensions.CustomScript-2.1.6/bin/custom-script-extension enable
Jun 20 00:09:48 fv-az72-309 python3[853]: time=2022-06-20T00:09:46Z version=v2.1.6/git@fc181d8-dirty operation=enable seq=99 event=start
Jun 20 00:09:48 fv-az72-309 python3[853]: time=2022-06-20T00:09:46Z version=v2.1.6/git@fc181d8-dirty operation=enable seq=99 event=pre-check
Jun 20 00:09:48 fv-az72-309 python3[853]: time=2022-06-20T00:09:46Z version=v2.1.6/git@fc181d8-dirty operation=enable seq=99 event="comparing seqnum" path=mrseq
Jun 20 00:09:48 fv-az72-309 python3[853]: time=2022-06-20T00:09:46Z version=v2.1.6/git@fc181d8-dirty operation=enable seq=99 event="seqnum saved" path=mrseq
Jun 20 00:09:48 fv-az72-309 python3[853]: time=2022-06-20T00:09:46Z version=v2.1.6/git@fc181d8-dirty operation=enable seq=99 event="reading configuration"
Jun 20 00:09:48 fv-az72-309 python3[853]: time=2022-06-20T00:09:46Z version=v2.1.6/git@fc181d8-dirty operation=enable seq=99 event="read configuration"
Jun 20 00:09:48 fv-az72-309 python3[853]: time=2022-06-20T00:09:46Z version=v2.1.6/git@fc181d8-dirty operation=enable seq=99 event="validating json schema"
Jun 20 00:09:48 fv-az72-309 python3[853]: time=2022-06-20T00:09:46Z version=v2.1.6/git@fc181d8-dirty operation=enable seq=99 event="json schema valid"
Jun 20 00:09:48 fv-az72-309 python3[853]: time=2022-06-20T00:09:46Z version=v2.1.6/git@fc181d8-dirty operation=enable seq=99 event="parsing configuration json"
Jun 20 00:09:48 fv-az72-309 python3[853]: time=2022-06-20T00:09:46Z version=v2.1.6/git@fc181d8-dirty operation=enable seq=99 event="parsed configuration json"
Jun 20 00:09:48 fv-az72-309 python3[853]: time=2022-06-20T00:09:46Z version=v2.1.6/git@fc181d8-dirty operation=enable seq=99 event="validating configuration logically"
Jun 20 00:09:48 fv-az72-309 python3[853]: time=2022-06-20T00:09:46Z version=v2.1.6/git@fc181d8-dirty operation=enable seq=99 event="validated configuration"
Jun 20 00:09:48 fv-az72-309 python3[853]: time=2022-06-20T00:09:46Z version=v2.1.6/git@fc181d8-dirty operation=enable seq=99 event="creating output directory" path=/var/lib/waagent/custom-script/download/99
Jun 20 00:09:48 fv-az72-309 python3[853]: time=2022-06-20T00:09:46Z version=v2.1.6/git@fc181d8-dirty operation=enable seq=99 event="created output directory"
Jun 20 00:09:48 fv-az72-309 python3[853]: time=2022-06-20T00:09:46Z version=v2.1.6/git@fc181d8-dirty operation=enable seq=99 files=0
Jun 20 00:09:48 fv-az72-309 python3[853]: time=2022-06-20T00:09:46Z version=v2.1.6/git@fc181d8-dirty operation=enable seq=99 event="executing command" output=/var/lib/waagent/custom-script/download/99
Jun 20 00:09:48 fv-az72-309 python3[853]: time=2022-06-20T00:09:46Z version=v2.1.6/git@fc181d8-dirty operation=enable seq=99 event="executing protected commandToExecute" output=/var/lib/waagent/custom-script/download/99
Jun 20 00:09:48 fv-az72-309 python3[853]: time=2022-06-20T00:09:46Z version=v2.1.6/git@fc181d8-dirty operation=enable seq=99 event="executed command" output=/var/lib/waagent/custom-script/download/99
Jun 20 00:09:48 fv-az72-309 python3[853]: time=2022-06-20T00:09:46Z version=v2.1.6/git@fc181d8-dirty operation=enable seq=99 event=enabled
Jun 20 00:09:48 fv-az72-309 python3[853]: time=2022-06-20T00:09:46Z version=v2.1.6/git@fc181d8-dirty operation=enable seq=99 event=end
Jun 20 00:09:48 fv-az72-309 python3[853]: [stderr]
Jun 20 00:09:48 fv-az72-309 python3[853]: Running scope as unit: enable_a078213d-e1c9-4073-aa18-bdb3cc076462.scope
Jun 20 00:09:48 fv-az72-309 python3[853]: 2022-06-20T00:09:48.249383Z INFO ExtHandler ExtHandler ProcessExtensionsInGoalState completed [Incarnation: 1; 4368 ms; Activity Id: 48fbd77a-0509-46ad-81a1-1394da8eb4b4; Correlation Id: 16378185-6322-4308-98a6-03c89673c297; GS Creation Time: 2022-06-20T00:07:55.901245Z]
Jun 20 00:09:48 fv-az72-309 python3[853]: 2022-06-20T00:09:48.279861Z INFO ExtHandler ExtHandler Extension status: [('Microsoft.Azure.Extensions.CustomScript', 'success')]
Jun 20 00:09:48 fv-az72-309 python3[853]: 2022-06-20T00:09:48.280279Z INFO ExtHandler ExtHandler All extensions in the goal state have reached a terminal state: [('Microsoft.Azure.Extensions.CustomScript', 'success')]
Jun 20 00:09:48 fv-az72-309 python3[853]: 2022-06-20T00:09:48.280630Z INFO ExtHandler ExtHandler Looking for existing remote access users.
Jun 20 00:09:48 fv-az72-309 python3[853]: 2022-06-20T00:09:48.291931Z INFO ExtHandler ExtHandler [HEARTBEAT] Agent WALinuxAgent-2.7.1.0 is running as the goal state agent [DEBUG HeartbeatCounter: 0;HeartbeatId: D409D817-B676-42D6-87B2-4FC0E04C82EB;DroppedPackets: 0;UpdateGSErrors: 0;AutoUpdate: 1]
Jun 20 00:09:48 fv-az72-309 cloud-init[1395]: #############################################################
Jun 20 00:09:48 fv-az72-309 cloud-init[1396]: -----BEGIN SSH HOST KEY FINGERPRINTS-----
Jun 20 00:09:48 fv-az72-309 cloud-init[1398]: 1024 SHA256:XGcH1MwJBdIF6Dewo/X0osa6enb06nIxLFk5/q6UAm8 root@fv-az72-309 (DSA)
Jun 20 00:09:48 fv-az72-309 cloud-init[1400]: 256 SHA256:6mlnK30thyw5xPDk04GqP6xQuDIEULkyXUpZlLeH/us root@fv-az72-309 (ECDSA)
Jun 20 00:09:48 fv-az72-309 cloud-init[1402]: 256 SHA256:CkHtFnUi6AlzMN/rV+bhb10BzsSd4ft4SZpQoIuutqU root@fv-az72-309 (ED25519)
Jun 20 00:09:48 fv-az72-309 cloud-init[1404]: 3072 SHA256:xgHytXr9AoLrM03Ufz16C0Uv9NChaKGn5zsbtoCu2Gg root@fv-az72-309 (RSA)
Jun 20 00:09:48 fv-az72-309 cloud-init[1405]: -----END SSH HOST KEY FINGERPRINTS-----
Jun 20 00:09:48 fv-az72-309 cloud-init[1406]: #############################################################
Jun 20 00:09:48 fv-az72-309 cloud-init[1393]: Cloud-init v. 22.2-0ubuntu1~20.04.2 running 'modules:final' at Mon, 20 Jun 2022 00:09:48 +0000. Up 16.62 seconds.
Jun 20 00:09:48 fv-az72-309 cloud-init[1393]: Cloud-init v. 22.2-0ubuntu1~20.04.2 finished at Mon, 20 Jun 2022 00:09:48 +0000. Datasource DataSourceAzure [seed=/var/lib/waagent]. Up 16.83 seconds
Jun 20 00:09:48 fv-az72-309 systemd[1]: Finished Execute cloud user/final scripts.
Jun 20 00:09:48 fv-az72-309 systemd[1]: Reached target Cloud-init target.
Jun 20 00:09:48 fv-az72-309 systemd[1]: Startup finished in 2.377s (kernel) + 14.528s (userspace) = 16.905s.
Jun 20 00:10:03 fv-az72-309 provisioner[1432]: √ Connected to GitHub
Jun 20 00:10:03 fv-az72-309 provisioner[1432]: Current runner version: '2.293.0'
Jun 20 00:10:03 fv-az72-309 provisioner[1432]: 2022-06-20 00:10:03Z: Listening for Jobs
Jun 20 00:10:04 fv-az72-309 provisioner[1432]: 2022-06-20 00:10:04Z: Running job: py38-ansible_4
Jun 20 00:10:05 fv-az72-309 chronyd[666]: Selected source PHC0
Jun 20 00:10:06 fv-az72-309 systemd[1]: systemd-fsckd.service: Succeeded.
Jun 20 00:10:08 fv-az72-309 sudo[1563]: runner : TTY=unknown ; PWD=/home/runner/work/molecule-kubevirt/molecule-kubevirt ; USER=root ; COMMAND=/usr/bin/apt-get update
Jun 20 00:10:08 fv-az72-309 sudo[1563]: pam_unix(sudo:session): session opened for user root by (uid=0)
Jun 20 00:10:10 fv-az72-309 dbus-daemon[658]: [system] Activating via systemd: service name='org.freedesktop.PackageKit' unit='packagekit.service' requested by ':1.13' (uid=0 pid=2142 comm="/usr/bin/gdbus call --system --dest org.freedeskto" label="unconfined")
Jun 20 00:10:10 fv-az72-309 systemd[1]: Starting PackageKit Daemon...
Jun 20 00:10:10 fv-az72-309 PackageKit[2145]: daemon start
Jun 20 00:10:10 fv-az72-309 dbus-daemon[658]: [system] Successfully activated service 'org.freedesktop.PackageKit'
Jun 20 00:10:10 fv-az72-309 systemd[1]: Started PackageKit Daemon.
Jun 20 00:10:13 fv-az72-309 systemd[1]: Starting Online ext4 Metadata Check for All Filesystems...
Jun 20 00:10:13 fv-az72-309 systemd[1]: e2scrub_all.service: Succeeded.
Jun 20 00:10:13 fv-az72-309 systemd[1]: Finished Online ext4 Metadata Check for All Filesystems.
Jun 20 00:10:16 fv-az72-309 sudo[1563]: pam_unix(sudo:session): session closed for user root
Jun 20 00:10:16 fv-az72-309 sudo[2229]: runner : TTY=unknown ; PWD=/home/runner/work/molecule-kubevirt/molecule-kubevirt ; USER=root ; COMMAND=/usr/bin/apt-key add -
Jun 20 00:10:16 fv-az72-309 sudo[2229]: pam_unix(sudo:session): session opened for user root by (uid=0)
Jun 20 00:10:17 fv-az72-309 sudo[2229]: pam_unix(sudo:session): session closed for user root
Jun 20 00:10:17 fv-az72-309 sudo[2695]: runner : TTY=unknown ; PWD=/home/runner/work/molecule-kubevirt/molecule-kubevirt ; USER=root ; COMMAND=/usr/bin/add-apt-repository deb [arch=amd64] https://download.docker.com/linux/ubuntu focal stable
Jun 20 00:10:17 fv-az72-309 sudo[2695]: pam_unix(sudo:session): session opened for user root by (uid=0)
Jun 20 00:10:17 fv-az72-309 systemd[1]: systemd-timedated.service: Succeeded.
Jun 20 00:10:22 fv-az72-309 sudo[2695]: pam_unix(sudo:session): session closed for user root
Jun 20 00:10:22 fv-az72-309 sudo[3435]: runner : TTY=unknown ; PWD=/home/runner/work/molecule-kubevirt/molecule-kubevirt ; USER=root ; COMMAND=/usr/bin/apt-get update
Jun 20 00:10:22 fv-az72-309 sudo[3435]: pam_unix(sudo:session): session opened for user root by (uid=0)
Jun 20 00:10:23 fv-az72-309 kernel: hv_balloon: Max. dynamic memory size: 7168 MB
Jun 20 00:10:24 fv-az72-309 sudo[3435]: pam_unix(sudo:session): session closed for user root
Jun 20 00:10:24 fv-az72-309 sudo[2227]: runner : TTY=unknown ; PWD=/home/runner/work/molecule-kubevirt/molecule-kubevirt ; USER=root ; COMMAND=/usr/bin/apt-get -y -o Dpkg::Options::=--force-confnew install docker-ce
Jun 20 00:10:24 fv-az72-309 sudo[2227]: pam_unix(sudo:session): session opened for user root by (uid=0)
Jun 20 00:10:35 fv-az72-309 systemd[1]: Stopping Docker Application Container Engine...
Jun 20 00:10:35 fv-az72-309 dockerd[913]: time="2022-06-20T00:10:35.116313511Z" level=info msg="Processing signal 'terminated'"
Jun 20 00:10:35 fv-az72-309 dockerd[913]: time="2022-06-20T00:10:35.117809018Z" level=info msg="stopping event stream following graceful shutdown" error="<nil>" module=libcontainerd namespace=moby
Jun 20 00:10:35 fv-az72-309 dockerd[913]: time="2022-06-20T00:10:35.117972530Z" level=info msg="Daemon shutdown complete"
Jun 20 00:10:35 fv-az72-309 systemd[1]: docker.service: Succeeded.
Jun 20 00:10:35 fv-az72-309 systemd[1]: Stopped Docker Application Container Engine.
Jun 20 00:10:35 fv-az72-309 systemd[1]: docker.socket: Succeeded.
Jun 20 00:10:35 fv-az72-309 systemd[1]: Closed Docker Socket for the API.
Jun 20 00:10:35 fv-az72-309 systemd[1]: Reloading.
Jun 20 00:10:35 fv-az72-309 systemd[1]: /etc/systemd/system/runner-provisioner.service:3: Invalid URL, ignoring: None
Jun 20 00:10:35 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network-online.target ignored (device units cannot be delayed).
Jun 20 00:10:35 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network.target ignored (device units cannot be delayed).
Jun 20 00:10:35 fv-az72-309 systemd[1]: Reloading.
Jun 20 00:10:35 fv-az72-309 systemd[1]: /etc/systemd/system/runner-provisioner.service:3: Invalid URL, ignoring: None
Jun 20 00:10:35 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network-online.target ignored (device units cannot be delayed).
Jun 20 00:10:35 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network.target ignored (device units cannot be delayed).
Jun 20 00:10:35 fv-az72-309 systemd[1]: Stopping containerd container runtime...
Jun 20 00:10:35 fv-az72-309 containerd[707]: time="2022-06-20T00:10:35.864411488Z" level=info msg="Stop CRI service"
Jun 20 00:10:35 fv-az72-309 containerd[707]: time="2022-06-20T00:10:35.875503484Z" level=info msg="Stop CRI service"
Jun 20 00:10:35 fv-az72-309 systemd[1]: containerd.service: Succeeded.
Jun 20 00:10:35 fv-az72-309 systemd[1]: Stopped containerd container runtime.
Jun 20 00:10:35 fv-az72-309 systemd[1]: Reloading.
Jun 20 00:10:36 fv-az72-309 systemd[1]: /etc/systemd/system/runner-provisioner.service:3: Invalid URL, ignoring: None
Jun 20 00:10:36 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network-online.target ignored (device units cannot be delayed).
Jun 20 00:10:36 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network.target ignored (device units cannot be delayed).
Jun 20 00:10:36 fv-az72-309 systemd[1]: Reloading.
Jun 20 00:10:36 fv-az72-309 systemd[1]: /etc/systemd/system/runner-provisioner.service:3: Invalid URL, ignoring: None
Jun 20 00:10:36 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network-online.target ignored (device units cannot be delayed).
Jun 20 00:10:36 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network.target ignored (device units cannot be delayed).
Jun 20 00:10:47 fv-az72-309 systemd[1]: Reloading.
Jun 20 00:10:47 fv-az72-309 systemd[1]: /etc/systemd/system/runner-provisioner.service:3: Invalid URL, ignoring: None
Jun 20 00:10:47 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network-online.target ignored (device units cannot be delayed).
Jun 20 00:10:47 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network.target ignored (device units cannot be delayed).
Jun 20 00:10:47 fv-az72-309 systemd[1]: Reloading.
Jun 20 00:10:47 fv-az72-309 systemd[1]: /etc/systemd/system/runner-provisioner.service:3: Invalid URL, ignoring: None
Jun 20 00:10:47 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network-online.target ignored (device units cannot be delayed).
Jun 20 00:10:47 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network.target ignored (device units cannot be delayed).
Jun 20 00:10:47 fv-az72-309 systemd[1]: Starting containerd container runtime...
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47Z" level=warning msg="containerd config version `1` has been deprecated and will be removed in containerd v2.0, please switch to version `2`, see https://github.com/containerd/containerd/blob/main/docs/PLUGINS.md#version-header"
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.796723350Z" level=info msg="starting containerd" revision=10c12954828e7c7c9b6e0ea9b0c02b01407d3ae1 version=1.6.6
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.814972692Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.815320319Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.817331178Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.817549095Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.817575097Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.817589598Z" level=warning msg="failed to load plugin io.containerd.snapshotter.v1.devmapper" error="devmapper not configured"
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.817599999Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.817623101Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.817757012Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.817918524Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.817942026Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.817960228Z" level=warning msg="could not use snapshotter devmapper in metadata plugin" error="devmapper not configured"
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.817971528Z" level=info msg="metadata content store policy set" policy=shared
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.818061336Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.818078537Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.818092038Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.818118740Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.818133741Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.818148042Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.818160643Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.818184145Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.818198046Z" level=info msg="loading plugin \"io.containerd.service.v1.leases-service\"..." type=io.containerd.service.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.818212247Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.818225048Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.818241550Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.818277953Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.818312855Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.818627680Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.818669784Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.818684285Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.818727188Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.818743789Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.818756790Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.818768791Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.818781592Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.818795094Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.818820496Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.818831896Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.818844797Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.818880700Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.818893401Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.818908602Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.818919303Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.818933304Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="no OpenTelemetry endpoint: skip plugin" type=io.containerd.tracing.processor.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.818944505Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.818966007Z" level=error msg="failed to initialize a tracing processor \"otlp\"" error="no OpenTelemetry endpoint: skip plugin"
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.819166623Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.819201426Z" level=info msg=serving... address=/run/containerd/containerd.sock
Jun 20 00:10:47 fv-az72-309 systemd[1]: Started containerd container runtime.
Jun 20 00:10:47 fv-az72-309 containerd[4465]: time="2022-06-20T00:10:47.820354117Z" level=info msg="containerd successfully booted in 0.024895s"
Jun 20 00:10:47 fv-az72-309 systemd[1]: Reloading.
Jun 20 00:10:48 fv-az72-309 systemd[1]: /etc/systemd/system/runner-provisioner.service:3: Invalid URL, ignoring: None
Jun 20 00:10:48 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network-online.target ignored (device units cannot be delayed).
Jun 20 00:10:48 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network.target ignored (device units cannot be delayed).
Jun 20 00:10:48 fv-az72-309 systemd[1]: Reloading.
Jun 20 00:10:48 fv-az72-309 systemd[1]: /etc/systemd/system/runner-provisioner.service:3: Invalid URL, ignoring: None
Jun 20 00:10:48 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network-online.target ignored (device units cannot be delayed).
Jun 20 00:10:48 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network.target ignored (device units cannot be delayed).
Jun 20 00:10:48 fv-az72-309 systemd[1]: Reloading.
Jun 20 00:10:48 fv-az72-309 systemd[1]: /etc/systemd/system/runner-provisioner.service:3: Invalid URL, ignoring: None
Jun 20 00:10:48 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network-online.target ignored (device units cannot be delayed).
Jun 20 00:10:48 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network.target ignored (device units cannot be delayed).
Jun 20 00:10:49 fv-az72-309 systemd[1]: Reloading.
Jun 20 00:10:49 fv-az72-309 systemd[1]: /etc/systemd/system/runner-provisioner.service:3: Invalid URL, ignoring: None
Jun 20 00:10:49 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network-online.target ignored (device units cannot be delayed).
Jun 20 00:10:49 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network.target ignored (device units cannot be delayed).
Jun 20 00:10:49 fv-az72-309 systemd[1]: Starting Docker Socket for the API.
Jun 20 00:10:49 fv-az72-309 systemd[1]: Listening on Docker Socket for the API.
Jun 20 00:10:49 fv-az72-309 systemd[1]: Starting Docker Application Container Engine...
Jun 20 00:10:49 fv-az72-309 dockerd[4625]: time="2022-06-20T00:10:49.420511696Z" level=info msg="Starting up"
Jun 20 00:10:49 fv-az72-309 dockerd[4625]: time="2022-06-20T00:10:49.421066140Z" level=info msg="detected 127.0.0.53 nameserver, assuming systemd-resolved, so using resolv.conf: /run/systemd/resolve/resolv.conf"
Jun 20 00:10:49 fv-az72-309 dockerd[4625]: time="2022-06-20T00:10:49.422105122Z" level=info msg="parsed scheme: \"unix\"" module=grpc
Jun 20 00:10:49 fv-az72-309 dockerd[4625]: time="2022-06-20T00:10:49.422132724Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc
Jun 20 00:10:49 fv-az72-309 dockerd[4625]: time="2022-06-20T00:10:49.422158026Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///run/containerd/containerd.sock <nil> 0 <nil>}] <nil> <nil>}" module=grpc
Jun 20 00:10:49 fv-az72-309 dockerd[4625]: time="2022-06-20T00:10:49.422169027Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc
Jun 20 00:10:49 fv-az72-309 dockerd[4625]: time="2022-06-20T00:10:49.425101059Z" level=info msg="parsed scheme: \"unix\"" module=grpc
Jun 20 00:10:49 fv-az72-309 dockerd[4625]: time="2022-06-20T00:10:49.425124860Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc
Jun 20 00:10:49 fv-az72-309 dockerd[4625]: time="2022-06-20T00:10:49.425160763Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///run/containerd/containerd.sock <nil> 0 <nil>}] <nil> <nil>}" module=grpc
Jun 20 00:10:49 fv-az72-309 dockerd[4625]: time="2022-06-20T00:10:49.425171164Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc
Jun 20 00:10:49 fv-az72-309 systemd[1]: var-lib-docker-overlay2-check\x2doverlayfs\x2dsupport3991811295-merged.mount: Succeeded.
Jun 20 00:10:49 fv-az72-309 dockerd[4625]: time="2022-06-20T00:10:49.592773901Z" level=info msg="[graphdriver] using prior storage driver: overlay2"
Jun 20 00:10:49 fv-az72-309 dockerd[4625]: time="2022-06-20T00:10:49.602167743Z" level=warning msg="Your kernel does not support CPU realtime scheduler"
Jun 20 00:10:49 fv-az72-309 dockerd[4625]: time="2022-06-20T00:10:49.602201646Z" level=warning msg="Your kernel does not support cgroup blkio weight"
Jun 20 00:10:49 fv-az72-309 dockerd[4625]: time="2022-06-20T00:10:49.602208646Z" level=warning msg="Your kernel does not support cgroup blkio weight_device"
Jun 20 00:10:49 fv-az72-309 dockerd[4625]: time="2022-06-20T00:10:49.602353958Z" level=info msg="Loading containers: start."
Jun 20 00:10:49 fv-az72-309 dockerd[4625]: time="2022-06-20T00:10:49.686397796Z" level=info msg="Default bridge (docker0) is assigned with an IP address 172.17.0.0/16. Daemon option --bip can be used to set a preferred IP address"
Jun 20 00:10:49 fv-az72-309 dockerd[4625]: time="2022-06-20T00:10:49.720531291Z" level=info msg="Loading containers: done."
Jun 20 00:10:49 fv-az72-309 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck1251787295-merged.mount: Succeeded.
Jun 20 00:10:49 fv-az72-309 dockerd[4625]: time="2022-06-20T00:10:49.739961926Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2
Jun 20 00:10:49 fv-az72-309 dockerd[4625]: time="2022-06-20T00:10:49.740165442Z" level=info msg="Docker daemon" commit=a89b842 graphdriver(s)=overlay2 version=20.10.17
Jun 20 00:10:49 fv-az72-309 dockerd[4625]: time="2022-06-20T00:10:49.740214946Z" level=info msg="Daemon has completed initialization"
Jun 20 00:10:49 fv-az72-309 systemd[1]: Started Docker Application Container Engine.
Jun 20 00:10:49 fv-az72-309 dockerd[4625]: time="2022-06-20T00:10:49.756247612Z" level=info msg="API listen on /run/docker.sock"
Jun 20 00:10:55 fv-az72-309 systemd[1]: Reloading.
Jun 20 00:10:55 fv-az72-309 systemd[1]: /etc/systemd/system/runner-provisioner.service:3: Invalid URL, ignoring: None
Jun 20 00:10:55 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network-online.target ignored (device units cannot be delayed).
Jun 20 00:10:55 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network.target ignored (device units cannot be delayed).
Jun 20 00:10:57 fv-az72-309 sudo[2227]: pam_unix(sudo:session): session closed for user root
Jun 20 00:10:58 fv-az72-309 sudo[5792]: runner : TTY=unknown ; PWD=/home/runner/work/molecule-kubevirt/molecule-kubevirt ; USER=root ; COMMAND=/usr/bin/install kubectl /usr/local/bin
Jun 20 00:10:58 fv-az72-309 sudo[5792]: pam_unix(sudo:session): session opened for user root by (uid=0)
Jun 20 00:10:58 fv-az72-309 sudo[5792]: pam_unix(sudo:session): session closed for user root
Jun 20 00:11:00 fv-az72-309 sudo[5801]: runner : TTY=unknown ; PWD=/home/runner/work/molecule-kubevirt/molecule-kubevirt ; USER=root ; COMMAND=/usr/bin/install minikube-linux-amd64 /usr/local/bin/minikube
Jun 20 00:11:00 fv-az72-309 sudo[5801]: pam_unix(sudo:session): session opened for user root by (uid=0)
Jun 20 00:11:00 fv-az72-309 sudo[5801]: pam_unix(sudo:session): session closed for user root
Jun 20 00:11:00 fv-az72-309 sudo[5855]: runner : TTY=unknown ; PWD=/home/runner/work/molecule-kubevirt/molecule-kubevirt ; USER=root ; COMMAND=/usr/bin/echo -n
Jun 20 00:11:00 fv-az72-309 sudo[5855]: pam_unix(sudo:session): session opened for user root by (uid=0)
Jun 20 00:11:00 fv-az72-309 sudo[5855]: pam_unix(sudo:session): session closed for user root
Jun 20 00:11:00 fv-az72-309 sudo[5857]: runner : TTY=unknown ; PWD=/home/runner/work/molecule-kubevirt/molecule-kubevirt ; USER=root ; COMMAND=/usr/bin/podman version --format {{.Version}}
Jun 20 00:11:00 fv-az72-309 sudo[5857]: pam_unix(sudo:session): session opened for user root by (uid=0)
Jun 20 00:11:05 fv-az72-309 podman[5858]: 2022-06-20 00:11:05.639907895 +0000 UTC m=+4.331075114 system refresh
Jun 20 00:11:05 fv-az72-309 systemd[1]: var-lib-containers-storage-overlay.mount: Succeeded.
Jun 20 00:11:05 fv-az72-309 sudo[5857]: pam_unix(sudo:session): session closed for user root
Jun 20 00:11:16 fv-az72-309 systemd[1]: var-lib-docker-overlay2-6c321a59084cff0062aa517d0b7c15b48b7ae5954a32f720afef9c3d9fac2981-merged.mount: Succeeded.
Jun 20 00:11:18 fv-az72-309 systemd[1]: var-lib-docker-overlay2-e97c88604001e5d750978a5f1491da128224cfbc0159d279014fd7a0c4fef3aa-merged.mount: Succeeded.
Jun 20 00:11:18 fv-az72-309 systemd[1]: var-lib-docker-overlay2-939c81a56812989ae0c022f9b9e4178df39ad3eba5204a23d01089957cf3d9df-merged.mount: Succeeded.
Jun 20 00:11:18 fv-az72-309 systemd[1]: var-lib-docker-overlay2-701f55828eb21355bdea257a5a7834ee61286cb8b6682b45cae571479c161431-merged.mount: Succeeded.
Jun 20 00:11:18 fv-az72-309 systemd[1]: var-lib-docker-overlay2-a6d415c4f816fe6aebb48dadcda8cf5c74a58cc1d79d327bb9476a50bf8f22ad-merged.mount: Succeeded.
Jun 20 00:11:18 fv-az72-309 systemd[1]: var-lib-docker-overlay2-eadead2972bf101e5c1a2f5c42210bce6d776d0fdb849bc18d7d7da54255a41b-merged.mount: Succeeded.
Jun 20 00:11:18 fv-az72-309 systemd[1]: var-lib-docker-overlay2-231e41ff55fdeb4fdff88135e4a347d309275fdaede65728ad2516ae746b4a27-merged.mount: Succeeded.
Jun 20 00:11:19 fv-az72-309 systemd[1]: var-lib-docker-overlay2-da081718bdf45f5d15924e71da9a20ff0196d55ed0d3a1cbe6f1c59a4ad34bdc-merged.mount: Succeeded.
Jun 20 00:11:19 fv-az72-309 systemd[1]: var-lib-docker-overlay2-d9ed3e8393decd85b3a4c6a80650266cf6dc79425a38460117d7beef209c7033-merged.mount: Succeeded.
Jun 20 00:11:19 fv-az72-309 systemd[1]: var-lib-docker-overlay2-bf06a58cbabae6bd25d1718d74800b847146023670afc5a357369262de6563e0-merged.mount: Succeeded.
Jun 20 00:11:20 fv-az72-309 systemd[1]: var-lib-docker-overlay2-c78d379f8f717c9f54d82c2aa6aed83318867e72fdab84871d96bcd0a45e9390-merged.mount: Succeeded.
Jun 20 00:11:20 fv-az72-309 systemd[1]: var-lib-docker-overlay2-e596a31f802c317d50325fc3310b26e28d296e9453ebf05ea002eff611b70fac-merged.mount: Succeeded.
Jun 20 00:11:21 fv-az72-309 systemd[1]: var-lib-docker-overlay2-b90c0d51a1b98be92f1287c573dec35947ecad93feeea17df0cce10e710d9611-merged.mount: Succeeded.
Jun 20 00:11:24 fv-az72-309 systemd[1]: var-lib-docker-overlay2-9095418bb28b6b95b432d5399a4f225f5dd45ba43665337140607c45b952c71c-merged.mount: Succeeded.
Jun 20 00:11:26 fv-az72-309 systemd[1]: var-lib-docker-overlay2-a7d717a8db54f8dfdb56022e705ae2c1e6f05a680ae2429ec72371bfa0947da2-merged.mount: Succeeded.
Jun 20 00:11:26 fv-az72-309 systemd[1]: var-lib-docker-overlay2-69bd8e93815b6fe6189e90f5faafe1b2da213d92a8094703c4dd3e7c4fafcfd0-merged.mount: Succeeded.
Jun 20 00:11:26 fv-az72-309 systemd[1]: var-lib-docker-overlay2-9e91112175779bd56d8187989420c6e42097e2f0112ccb444814ede5c11f9c02-merged.mount: Succeeded.
Jun 20 00:11:27 fv-az72-309 systemd[1]: var-lib-docker-overlay2-5e73ccc9c742e427f132a279a66ce342e1aa1b74a06448970fd4addd7231bbbb-merged.mount: Succeeded.
Jun 20 00:11:29 fv-az72-309 systemd[1]: var-lib-docker-overlay2-c1b9946c54237f328340c6103a5a14b22e30a4476905c0f1101dd14bd34cf2bf-merged.mount: Succeeded.
Jun 20 00:11:30 fv-az72-309 systemd[1]: var-lib-docker-overlay2-e01d78150b1ef4cd2658a853c74c50b538b2aa58f5f7a2c8dbcb7cf1fabba111-merged.mount: Succeeded.
Jun 20 00:11:31 fv-az72-309 systemd[1]: var-lib-docker-overlay2-27323b6802f90d3cb3e3787d540d1e2523e61fa9734fc84c0a8b4ddca2b4b24f-merged.mount: Succeeded.
Jun 20 00:11:32 fv-az72-309 systemd[1]: var-lib-docker-overlay2-cd56f4e681eaca678eaed0d0264377b89c8e18b170563f9ace857ddfc149fcf1-merged.mount: Succeeded.
Jun 20 00:11:32 fv-az72-309 systemd[1]: var-lib-docker-overlay2-55eb779a7ffb4780331b4a685ee90b85366c0b40d440e5eb5140e62419957dad-merged.mount: Succeeded.
Jun 20 00:11:32 fv-az72-309 systemd[1]: var-lib-docker-overlay2-cb3c3a07917ad57f1ae7ed4818c6f62ce6d05d41a93d27f6d61f3df338dd3fd4-merged.mount: Succeeded.
Jun 20 00:11:32 fv-az72-309 systemd[1]: var-lib-docker-overlay2-e07b79b8c2d6eb35285c01badba514e2b9e6e9a713f8c1a6aa16ae4e092676b7-merged.mount: Succeeded.
Jun 20 00:11:32 fv-az72-309 systemd[1]: var-lib-docker-overlay2-f7f35747b342c73deefc5b36265ac1b816be16990ba9a573025627e734fd04a2-merged.mount: Succeeded.
Jun 20 00:11:32 fv-az72-309 systemd[1]: var-lib-docker-overlay2-29aa876d126e1a3b2dcc87cabc9f2ec9139946d6dd1c21c37b0ab7c1b8d09bd1-merged.mount: Succeeded.
Jun 20 00:11:32 fv-az72-309 systemd[1]: var-lib-docker-overlay2-d41d1a5a7740083f5b8013b2512160282a4c4dbb550d8ac9d8a9770684095f46-merged.mount: Succeeded.
Jun 20 00:11:32 fv-az72-309 systemd[1]: var-lib-docker-overlay2-8bd0ccec22895d82675b97cb4ef1579f985e88edd3a8bbe45bc6c741c61535cd-merged.mount: Succeeded.
Jun 20 00:11:32 fv-az72-309 systemd[1]: var-lib-docker-overlay2-a98382247c8d1a0e8732ccc7f4166fec1ce6f32ffb1cdc9c084e8e83b452f5b2-merged.mount: Succeeded.
Jun 20 00:11:32 fv-az72-309 systemd[1]: var-lib-docker-overlay2-64dbe618f70ad0b5fe1b3f4ac82c405c50e20966745d4d47c8ee08f4017e8ad9-merged.mount: Succeeded.
Jun 20 00:11:32 fv-az72-309 systemd[1]: var-lib-docker-overlay2-ee7d050153ac0527212321510ca30bb925b626d30cdb8a7a1023f3671c4de2e2-merged.mount: Succeeded.
Jun 20 00:11:32 fv-az72-309 systemd[1]: var-lib-docker-overlay2-9f9281029c0d0203621df207e6e7c833e8d51d56bc622dcb98b7468e76a3634b-merged.mount: Succeeded.
Jun 20 00:11:32 fv-az72-309 systemd[1]: var-lib-docker-overlay2-316a1b8d8c82b4aecfb6dabafe3f9c1b9843fc8b9638ccfe320962bbdcb6175b-merged.mount: Succeeded.
Jun 20 00:11:32 fv-az72-309 systemd[1]: var-lib-docker-overlay2-f36743a9ce57dd9940ca50b10bd2ee08a0a7fb38ef9b73508df7b2fc8e84d1dc-merged.mount: Succeeded.
Jun 20 00:11:32 fv-az72-309 systemd[1]: var-lib-docker-overlay2-3ed3613e2ba316942d24faeca9b157a20297652ed61f3d0bf86cc0eb8fb64da8-merged.mount: Succeeded.
Jun 20 00:11:32 fv-az72-309 systemd[1]: var-lib-docker-overlay2-730175aadb1969078d811fcb5011a8b7390b6a6c30b6d127148694f4368bdf10-merged.mount: Succeeded.
Jun 20 00:11:33 fv-az72-309 systemd[1]: var-lib-docker-overlay2-3b2bee075d790f9351547c0180f73dfc6b0111299b26abe53b09dda6dd4d7277-merged.mount: Succeeded.
Jun 20 00:11:33 fv-az72-309 systemd[1]: var-lib-docker-overlay2-afcc22a6f583683733aaf876a399a3f29506768725cad8c3efea2ce479349e8f-merged.mount: Succeeded.
Jun 20 00:11:33 fv-az72-309 systemd[1]: var-lib-docker-overlay2-4fa2c6f702477b3992762a5b58303381268ba5b93cd8778df3f48f6c8de91eb6-merged.mount: Succeeded.
Jun 20 00:11:33 fv-az72-309 systemd[1]: var-lib-docker-overlay2-0f8a0319a399c0b3677db977f737db88127a034ca626f536a48f6e3482a45960-merged.mount: Succeeded.
Jun 20 00:11:33 fv-az72-309 systemd[1]: var-lib-docker-overlay2-d4194db259036492eb1b65366c0214ee4cf3f3531c528d0c8207ed6b63435ae0-merged.mount: Succeeded.
Jun 20 00:11:33 fv-az72-309 systemd[1]: var-lib-docker-overlay2-6b7009d4226b6c3f16c2b6c871514f82fe4cadc1442a53b24399db6300fe548c-merged.mount: Succeeded.
Jun 20 00:11:33 fv-az72-309 systemd[1]: var-lib-docker-overlay2-e9cc4f7e92378ab897810d8312cfe661382226bd5e03aa4b82fef3cc32c92422-merged.mount: Succeeded.
Jun 20 00:11:33 fv-az72-309 systemd-networkd[532]: rtnl: received neighbor for link '4' we don't know about, ignoring.
Jun 20 00:11:33 fv-az72-309 systemd-networkd[532]: rtnl: received neighbor for link '4' we don't know about, ignoring.
Jun 20 00:11:33 fv-az72-309 systemd-networkd[532]: rtnl: received neighbor for link '4' we don't know about, ignoring.
Jun 20 00:11:33 fv-az72-309 systemd-networkd[532]: rtnl: received neighbor for link '4' we don't know about, ignoring.
Jun 20 00:11:33 fv-az72-309 networkd-dispatcher[674]: WARNING:Unknown index 4 seen, reloading interface list
Jun 20 00:11:33 fv-az72-309 systemd-udevd[6451]: Using default interface naming scheme 'v245'.
Jun 20 00:11:33 fv-az72-309 systemd-udevd[6451]: ethtool: autonegotiation is unset or enabled, the speed and duplex are not writable.
Jun 20 00:11:33 fv-az72-309 systemd-networkd[532]: br-d1e9d479f443: Link UP
Jun 20 00:11:33 fv-az72-309 dockerd[4625]: time="2022-06-20T00:11:33.990813528Z" level=warning msg="reference for unknown type: " digest="sha256:02c921df998f95e849058af14de7045efc3954d90320967418a0d1f182bbc0b2" remote="gcr.io/k8s-minikube/kicbase@sha256:02c921df998f95e849058af14de7045efc3954d90320967418a0d1f182bbc0b2"
Jun 20 00:11:34 fv-az72-309 systemd[1]: var-lib-docker-overlay2-3ca3a993ed9e2d45f8bfa6b6ea429467eb15347912f720aee92e4256fbf6e461\x2dinit-merged.mount: Succeeded.
Jun 20 00:11:34 fv-az72-309 systemd[1]: var-lib-docker-overlay2-3ca3a993ed9e2d45f8bfa6b6ea429467eb15347912f720aee92e4256fbf6e461-merged.mount: Succeeded.
Jun 20 00:11:34 fv-az72-309 systemd-udevd[6464]: ethtool: autonegotiation is unset or enabled, the speed and duplex are not writable.
Jun 20 00:11:34 fv-az72-309 kernel: docker0: port 1(vethf1b0b28) entered blocking state
Jun 20 00:11:34 fv-az72-309 kernel: docker0: port 1(vethf1b0b28) entered disabled state
Jun 20 00:11:34 fv-az72-309 kernel: device vethf1b0b28 entered promiscuous mode
Jun 20 00:11:34 fv-az72-309 systemd-udevd[6451]: ethtool: autonegotiation is unset or enabled, the speed and duplex are not writable.
Jun 20 00:11:34 fv-az72-309 systemd-udevd[6451]: vethf1b0b28: Could not generate persistent MAC: No data available
Jun 20 00:11:34 fv-az72-309 systemd-udevd[6464]: Using default interface naming scheme 'v245'.
Jun 20 00:11:34 fv-az72-309 systemd-udevd[6464]: veth63d8682: Could not generate persistent MAC: No data available
Jun 20 00:11:34 fv-az72-309 networkd-dispatcher[674]: WARNING:Unknown index 5 seen, reloading interface list
Jun 20 00:11:34 fv-az72-309 systemd-networkd[532]: vethf1b0b28: Link UP
Jun 20 00:11:34 fv-az72-309 kernel: docker0: port 1(vethf1b0b28) entered blocking state
Jun 20 00:11:34 fv-az72-309 kernel: docker0: port 1(vethf1b0b28) entered forwarding state
Jun 20 00:11:34 fv-az72-309 kernel: docker0: port 1(vethf1b0b28) entered disabled state
Jun 20 00:11:34 fv-az72-309 containerd[4465]: time="2022-06-20T00:11:34.492887573Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1
Jun 20 00:11:34 fv-az72-309 containerd[4465]: time="2022-06-20T00:11:34.492956079Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1
Jun 20 00:11:34 fv-az72-309 containerd[4465]: time="2022-06-20T00:11:34.492969780Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jun 20 00:11:34 fv-az72-309 containerd[4465]: time="2022-06-20T00:11:34.493149594Z" level=info msg="starting signal loop" namespace=moby path=/run/containerd/io.containerd.runtime.v2.task/moby/1a2717258ff0923e177c915368c689d0457b73667aed09174aa9b155d7804bb5 pid=6505 runtime=io.containerd.runc.v2
Jun 20 00:11:34 fv-az72-309 systemd[1]: run-docker-runtime\x2drunc-moby-1a2717258ff0923e177c915368c689d0457b73667aed09174aa9b155d7804bb5-runc.074VXr.mount: Succeeded.
Jun 20 00:11:34 fv-az72-309 kernel: eth0: renamed from veth63d8682
Jun 20 00:11:34 fv-az72-309 systemd-networkd[532]: vethf1b0b28: Gained carrier
Jun 20 00:11:34 fv-az72-309 systemd-networkd[532]: docker0: Gained carrier
Jun 20 00:11:34 fv-az72-309 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): vethf1b0b28: link becomes ready
Jun 20 00:11:34 fv-az72-309 kernel: docker0: port 1(vethf1b0b28) entered blocking state
Jun 20 00:11:34 fv-az72-309 kernel: docker0: port 1(vethf1b0b28) entered forwarding state
Jun 20 00:11:34 fv-az72-309 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): docker0: link becomes ready
Jun 20 00:11:34 fv-az72-309 dockerd[4625]: time="2022-06-20T00:11:34.722254430Z" level=info msg="ignoring event" container=1a2717258ff0923e177c915368c689d0457b73667aed09174aa9b155d7804bb5 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Jun 20 00:11:34 fv-az72-309 containerd[4465]: time="2022-06-20T00:11:34.722628660Z" level=info msg="shim disconnected" id=1a2717258ff0923e177c915368c689d0457b73667aed09174aa9b155d7804bb5
Jun 20 00:11:34 fv-az72-309 containerd[4465]: time="2022-06-20T00:11:34.722683364Z" level=warning msg="cleaning up after shim disconnected" id=1a2717258ff0923e177c915368c689d0457b73667aed09174aa9b155d7804bb5 namespace=moby
Jun 20 00:11:34 fv-az72-309 containerd[4465]: time="2022-06-20T00:11:34.722697665Z" level=info msg="cleaning up dead shim"
Jun 20 00:11:34 fv-az72-309 containerd[4465]: time="2022-06-20T00:11:34.733387312Z" level=warning msg="cleanup warnings time=\"2022-06-20T00:11:34Z\" level=info msg=\"starting signal loop\" namespace=moby pid=6566 runtime=io.containerd.runc.v2\n"
Jun 20 00:11:34 fv-az72-309 systemd-networkd[532]: vethf1b0b28: Lost carrier
Jun 20 00:11:34 fv-az72-309 kernel: docker0: port 1(vethf1b0b28) entered disabled state
Jun 20 00:11:34 fv-az72-309 kernel: veth63d8682: renamed from eth0
Jun 20 00:11:34 fv-az72-309 systemd-udevd[6496]: ethtool: autonegotiation is unset or enabled, the speed and duplex are not writable.
Jun 20 00:11:34 fv-az72-309 networkd-dispatcher[674]: WARNING:Unknown index 5 seen, reloading interface list
Jun 20 00:11:34 fv-az72-309 systemd-udevd[6496]: Using default interface naming scheme 'v245'.
Jun 20 00:11:34 fv-az72-309 systemd-networkd[532]: vethf1b0b28: Link DOWN
Jun 20 00:11:34 fv-az72-309 kernel: docker0: port 1(vethf1b0b28) entered disabled state
Jun 20 00:11:34 fv-az72-309 kernel: device vethf1b0b28 left promiscuous mode
Jun 20 00:11:34 fv-az72-309 kernel: docker0: port 1(vethf1b0b28) entered disabled state
Jun 20 00:11:34 fv-az72-309 systemd-networkd[532]: rtnl: received neighbor for link '6' we don't know about, ignoring.
Jun 20 00:11:34 fv-az72-309 systemd-networkd[532]: rtnl: received neighbor for link '6' we don't know about, ignoring.
Jun 20 00:11:34 fv-az72-309 systemd-networkd[532]: veth63d8682: Failed to wait for the interface to be initialized: No such device
Jun 20 00:11:34 fv-az72-309 systemd-udevd[6496]: veth63d8682: Failed to get link config: No such device
Jun 20 00:11:34 fv-az72-309 networkd-dispatcher[6584]: Interface "vethf1b0b28" not found.
Jun 20 00:11:34 fv-az72-309 systemd[1]: networkd-dispatcher.service: Got notification message from PID 6584, but reception only permitted for main PID 674
Jun 20 00:11:34 fv-az72-309 networkd-dispatcher[674]: ERROR:Failed to get interface "vethf1b0b28" status: Command '['/usr/bin/networkctl', 'status', '--no-pager', '--no-legend', '--', 'vethf1b0b28']' returned non-zero exit status 1.
Jun 20 00:11:34 fv-az72-309 systemd-networkd[532]: docker0: Lost carrier
Jun 20 00:11:34 fv-az72-309 systemd[1]: run-docker-netns-a6977305fb52.mount: Succeeded.
Jun 20 00:11:34 fv-az72-309 systemd[1]: var-lib-docker-overlay2-3ca3a993ed9e2d45f8bfa6b6ea429467eb15347912f720aee92e4256fbf6e461-merged.mount: Succeeded.
Jun 20 00:11:34 fv-az72-309 systemd[1]: var-lib-docker-overlay2-c5545e751edee2d4c5f1da7b235db9235e2b92dbdc11130852f9deee263df41b\x2dinit-merged.mount: Succeeded.
Jun 20 00:11:34 fv-az72-309 systemd[1]: var-lib-docker-overlay2-c5545e751edee2d4c5f1da7b235db9235e2b92dbdc11130852f9deee263df41b-merged.mount: Succeeded.
Jun 20 00:11:34 fv-az72-309 kernel: docker0: port 1(vethfb268c1) entered blocking state
Jun 20 00:11:34 fv-az72-309 kernel: docker0: port 1(vethfb268c1) entered disabled state
Jun 20 00:11:34 fv-az72-309 kernel: device vethfb268c1 entered promiscuous mode
Jun 20 00:11:34 fv-az72-309 kernel: docker0: port 1(vethfb268c1) entered blocking state
Jun 20 00:11:34 fv-az72-309 kernel: docker0: port 1(vethfb268c1) entered forwarding state
Jun 20 00:11:34 fv-az72-309 kernel: docker0: port 1(vethfb268c1) entered disabled state
Jun 20 00:11:34 fv-az72-309 systemd-udevd[6496]: ethtool: autonegotiation is unset or enabled, the speed and duplex are not writable.
Jun 20 00:11:34 fv-az72-309 systemd-udevd[6496]: vethecca8aa: Could not generate persistent MAC: No data available
Jun 20 00:11:34 fv-az72-309 systemd-udevd[6451]: ethtool: autonegotiation is unset or enabled, the speed and duplex are not writable.
Jun 20 00:11:34 fv-az72-309 systemd-udevd[6451]: vethfb268c1: Could not generate persistent MAC: No data available
Jun 20 00:11:34 fv-az72-309 systemd-networkd[532]: vethfb268c1: Link UP
Jun 20 00:11:34 fv-az72-309 networkd-dispatcher[674]: WARNING:Unknown index 7 seen, reloading interface list
Jun 20 00:11:34 fv-az72-309 containerd[4465]: time="2022-06-20T00:11:34.955389186Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1
Jun 20 00:11:34 fv-az72-309 containerd[4465]: time="2022-06-20T00:11:34.955465192Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1
Jun 20 00:11:34 fv-az72-309 containerd[4465]: time="2022-06-20T00:11:34.955478493Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jun 20 00:11:34 fv-az72-309 containerd[4465]: time="2022-06-20T00:11:34.955850122Z" level=info msg="starting signal loop" namespace=moby path=/run/containerd/io.containerd.runtime.v2.task/moby/f665726575a2f573f76c241a1713ecbad3279d6357c95f1a74ce0ad837542379 pid=6615 runtime=io.containerd.runc.v2
Jun 20 00:11:35 fv-az72-309 kernel: eth0: renamed from vethecca8aa
Jun 20 00:11:35 fv-az72-309 systemd-networkd[532]: vethfb268c1: Gained carrier
Jun 20 00:11:35 fv-az72-309 systemd-networkd[532]: docker0: Gained carrier
Jun 20 00:11:35 fv-az72-309 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): vethfb268c1: link becomes ready
Jun 20 00:11:35 fv-az72-309 kernel: docker0: port 1(vethfb268c1) entered blocking state
Jun 20 00:11:35 fv-az72-309 kernel: docker0: port 1(vethfb268c1) entered forwarding state
Jun 20 00:11:36 fv-az72-309 systemd-networkd[532]: docker0: Gained IPv6LL
Jun 20 00:11:36 fv-az72-309 systemd-networkd[532]: vethfb268c1: Gained IPv6LL
Jun 20 00:11:38 fv-az72-309 dockerd[4625]: time="2022-06-20T00:11:38.494853075Z" level=info msg="ignoring event" container=f665726575a2f573f76c241a1713ecbad3279d6357c95f1a74ce0ad837542379 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Jun 20 00:11:38 fv-az72-309 containerd[4465]: time="2022-06-20T00:11:38.495523028Z" level=info msg="shim disconnected" id=f665726575a2f573f76c241a1713ecbad3279d6357c95f1a74ce0ad837542379
Jun 20 00:11:38 fv-az72-309 containerd[4465]: time="2022-06-20T00:11:38.495579832Z" level=warning msg="cleaning up after shim disconnected" id=f665726575a2f573f76c241a1713ecbad3279d6357c95f1a74ce0ad837542379 namespace=moby
Jun 20 00:11:38 fv-az72-309 containerd[4465]: time="2022-06-20T00:11:38.495589433Z" level=info msg="cleaning up dead shim"
Jun 20 00:11:38 fv-az72-309 containerd[4465]: time="2022-06-20T00:11:38.505608326Z" level=warning msg="cleanup warnings time=\"2022-06-20T00:11:38Z\" level=info msg=\"starting signal loop\" namespace=moby pid=6674 runtime=io.containerd.runc.v2\n"
Jun 20 00:11:38 fv-az72-309 systemd-networkd[532]: vethfb268c1: Lost carrier
Jun 20 00:11:38 fv-az72-309 kernel: docker0: port 1(vethfb268c1) entered disabled state
Jun 20 00:11:38 fv-az72-309 kernel: vethecca8aa: renamed from eth0
Jun 20 00:11:38 fv-az72-309 networkd-dispatcher[674]: WARNING:Unknown index 7 seen, reloading interface list
Jun 20 00:11:38 fv-az72-309 systemd-udevd[6686]: ethtool: autonegotiation is unset or enabled, the speed and duplex are not writable.
Jun 20 00:11:38 fv-az72-309 systemd-udevd[6686]: Using default interface naming scheme 'v245'.
Jun 20 00:11:38 fv-az72-309 systemd-udevd[6686]: ethtool: autonegotiation is unset or enabled, the speed and duplex are not writable.
Jun 20 00:11:38 fv-az72-309 kernel: docker0: port 1(vethfb268c1) entered disabled state
Jun 20 00:11:38 fv-az72-309 kernel: device vethfb268c1 left promiscuous mode
Jun 20 00:11:38 fv-az72-309 kernel: docker0: port 1(vethfb268c1) entered disabled state
Jun 20 00:11:38 fv-az72-309 systemd-networkd[532]: vethfb268c1: Link DOWN
Jun 20 00:11:38 fv-az72-309 systemd-networkd[532]: rtnl: received neighbor for link '8' we don't know about, ignoring.
Jun 20 00:11:38 fv-az72-309 systemd-networkd[532]: rtnl: received neighbor for link '8' we don't know about, ignoring.
Jun 20 00:11:38 fv-az72-309 networkd-dispatcher[6691]: Interface "vethfb268c1" not found.
Jun 20 00:11:38 fv-az72-309 systemd[1]: networkd-dispatcher.service: Got notification message from PID 6691, but reception only permitted for main PID 674
Jun 20 00:11:38 fv-az72-309 networkd-dispatcher[674]: ERROR:Failed to get interface "vethfb268c1" status: Command '['/usr/bin/networkctl', 'status', '--no-pager', '--no-legend', '--', 'vethfb268c1']' returned non-zero exit status 1.
Jun 20 00:11:39 fv-az72-309 systemd[1]: run-docker-netns-05eaf9a0ca49.mount: Succeeded.
Jun 20 00:11:39 fv-az72-309 systemd[1]: var-lib-docker-overlay2-c5545e751edee2d4c5f1da7b235db9235e2b92dbdc11130852f9deee263df41b-merged.mount: Succeeded.
Jun 20 00:11:39 fv-az72-309 systemd-networkd[532]: docker0: Lost carrier
Jun 20 00:11:42 fv-az72-309 systemd[1]: var-lib-docker-overlay2-27b21e55169967d0d88fe85abc53300b79e020a26c271308cc67ef3a17972f16\x2dinit-merged.mount: Succeeded.
Jun 20 00:11:42 fv-az72-309 systemd[1]: var-lib-docker-overlay2-27b21e55169967d0d88fe85abc53300b79e020a26c271308cc67ef3a17972f16-merged.mount: Succeeded.
Jun 20 00:11:42 fv-az72-309 systemd-udevd[6734]: ethtool: autonegotiation is unset or enabled, the speed and duplex are not writable.
Jun 20 00:11:42 fv-az72-309 systemd-udevd[6734]: Using default interface naming scheme 'v245'.
Jun 20 00:11:42 fv-az72-309 systemd-udevd[6734]: veth37c7cf5: Could not generate persistent MAC: No data available
Jun 20 00:11:42 fv-az72-309 networkd-dispatcher[674]: WARNING:Unknown index 9 seen, reloading interface list
Jun 20 00:11:42 fv-az72-309 systemd-networkd[532]: vethc0cce7e: Link UP
Jun 20 00:11:42 fv-az72-309 kernel: br-d1e9d479f443: port 1(vethc0cce7e) entered blocking state
Jun 20 00:11:42 fv-az72-309 kernel: br-d1e9d479f443: port 1(vethc0cce7e) entered disabled state
Jun 20 00:11:42 fv-az72-309 kernel: device vethc0cce7e entered promiscuous mode
Jun 20 00:11:42 fv-az72-309 systemd-udevd[6735]: ethtool: autonegotiation is unset or enabled, the speed and duplex are not writable.
Jun 20 00:11:42 fv-az72-309 systemd-udevd[6735]: Using default interface naming scheme 'v245'.
Jun 20 00:11:42 fv-az72-309 systemd-udevd[6735]: vethc0cce7e: Could not generate persistent MAC: No data available
Jun 20 00:11:42 fv-az72-309 dockerd[4625]: time="2022-06-20T00:11:42.578170727Z" level=info msg="No non-localhost DNS nameservers are left in resolv.conf. Using default external servers: [nameserver 8.8.8.8 nameserver 8.8.4.4]"
Jun 20 00:11:42 fv-az72-309 dockerd[4625]: time="2022-06-20T00:11:42.578198129Z" level=info msg="IPv6 enabled; Adding default IPv6 external servers: [nameserver 2001:4860:4860::8888 nameserver 2001:4860:4860::8844]"
Jun 20 00:11:42 fv-az72-309 dockerd[4625]: time="2022-06-20T00:11:42.669328106Z" level=warning msg="path in container /dev/fuse already exists in privileged mode" container=a20c82aba6259bc7ba2452946286e99f59b49b1905483f4512c4f453a3631e0b
Jun 20 00:11:42 fv-az72-309 containerd[4465]: time="2022-06-20T00:11:42.686498758Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1
Jun 20 00:11:42 fv-az72-309 containerd[4465]: time="2022-06-20T00:11:42.686550262Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1
Jun 20 00:11:42 fv-az72-309 containerd[4465]: time="2022-06-20T00:11:42.686564864Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jun 20 00:11:42 fv-az72-309 containerd[4465]: time="2022-06-20T00:11:42.686740377Z" level=info msg="starting signal loop" namespace=moby path=/run/containerd/io.containerd.runtime.v2.task/moby/a20c82aba6259bc7ba2452946286e99f59b49b1905483f4512c4f453a3631e0b pid=6821 runtime=io.containerd.runc.v2
Jun 20 00:11:42 fv-az72-309 kernel: eth0: renamed from veth37c7cf5
Jun 20 00:11:42 fv-az72-309 systemd-networkd[532]: vethc0cce7e: Gained carrier
Jun 20 00:11:42 fv-az72-309 systemd-networkd[532]: br-d1e9d479f443: Gained carrier
Jun 20 00:11:42 fv-az72-309 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): vethc0cce7e: link becomes ready
Jun 20 00:11:42 fv-az72-309 kernel: br-d1e9d479f443: port 1(vethc0cce7e) entered blocking state
Jun 20 00:11:42 fv-az72-309 kernel: br-d1e9d479f443: port 1(vethc0cce7e) entered forwarding state
Jun 20 00:11:42 fv-az72-309 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): br-d1e9d479f443: link becomes ready
Jun 20 00:11:43 fv-az72-309 systemd[1]: run-docker-runtime\x2drunc-moby-a20c82aba6259bc7ba2452946286e99f59b49b1905483f4512c4f453a3631e0b-runc.r3Nxcj.mount: Succeeded.
Jun 20 00:11:43 fv-az72-309 systemd[1]: var-lib-docker-overlay2-27b21e55169967d0d88fe85abc53300b79e020a26c271308cc67ef3a17972f16-merged-usr-lib-modules.mount: Succeeded.
Jun 20 00:11:43 fv-az72-309 systemd[1]: var-lib-docker-overlay2-27b21e55169967d0d88fe85abc53300b79e020a26c271308cc67ef3a17972f16-merged-etc-hosts.mount: Succeeded.
Jun 20 00:11:43 fv-az72-309 systemd[1]: var-lib-docker-overlay2-27b21e55169967d0d88fe85abc53300b79e020a26c271308cc67ef3a17972f16-merged-etc-hostname.mount: Succeeded.
Jun 20 00:11:43 fv-az72-309 systemd[1]: var-lib-docker-overlay2-27b21e55169967d0d88fe85abc53300b79e020a26c271308cc67ef3a17972f16-merged-etc-resolv.conf.mount: Succeeded.
Jun 20 00:11:43 fv-az72-309 systemd[1]: var-lib-docker-overlay2-27b21e55169967d0d88fe85abc53300b79e020a26c271308cc67ef3a17972f16-merged-var.mount: Succeeded.
Jun 20 00:11:43 fv-az72-309 systemd-journald[180]: Received client request to flush runtime journal.
Jun 20 00:11:43 fv-az72-309 systemd[1]: run-docker-runtime\x2drunc-moby-a20c82aba6259bc7ba2452946286e99f59b49b1905483f4512c4f453a3631e0b-runc.W6g6fL.mount: Succeeded.
Jun 20 00:11:44 fv-az72-309 systemd-networkd[532]: br-d1e9d479f443: Gained IPv6LL
Jun 20 00:11:44 fv-az72-309 systemd-networkd[532]: vethc0cce7e: Gained IPv6LL
Jun 20 00:12:02 fv-az72-309 sudo[9090]: runner : TTY=unknown ; PWD=/home/runner/work/molecule-kubevirt/molecule-kubevirt ; USER=root ; COMMAND=/usr/bin/apt install -y qemu qemu-kvm libvirt-daemon libvirt-clients bridge-utils virt-manager libvirt-daemon-system
Jun 20 00:12:02 fv-az72-309 sudo[9090]: pam_unix(sudo:session): session opened for user root by (uid=0)
Jun 20 00:12:14 fv-az72-309 kernel: IPVS: Registered protocols (TCP, UDP, SCTP, AH, ESP)
Jun 20 00:12:14 fv-az72-309 kernel: IPVS: Connection hash table configured (size=4096, memory=64Kbytes)
Jun 20 00:12:14 fv-az72-309 kernel: IPVS: ipvs loaded.
Jun 20 00:12:14 fv-az72-309 kernel: IPVS: [rr] scheduler registered.
Jun 20 00:12:14 fv-az72-309 kernel: IPVS: [wrr] scheduler registered.
Jun 20 00:12:14 fv-az72-309 kernel: IPVS: [sh] scheduler registered.
Jun 20 00:12:15 fv-az72-309 kernel: docker0: port 1(vethf3c0380) entered blocking state
Jun 20 00:12:15 fv-az72-309 kernel: docker0: port 1(vethf3c0380) entered disabled state
Jun 20 00:12:15 fv-az72-309 kernel: device vethf3c0380 entered promiscuous mode
Jun 20 00:12:15 fv-az72-309 kernel: eth0: renamed from veth76d993e
Jun 20 00:12:15 fv-az72-309 kernel: docker0: port 1(vethf3c0380) entered blocking state
Jun 20 00:12:15 fv-az72-309 kernel: docker0: port 1(vethf3c0380) entered forwarding state
Jun 20 00:12:15 fv-az72-309 dbus-daemon[658]: [system] Reloaded configuration
Jun 20 00:12:15 fv-az72-309 dbus-daemon[658]: [system] Reloaded configuration
Jun 20 00:12:15 fv-az72-309 dbus-daemon[658]: [system] Reloaded configuration
Jun 20 00:12:15 fv-az72-309 dbus-daemon[658]: [system] Reloaded configuration
Jun 20 00:12:15 fv-az72-309 dbus-daemon[658]: [system] Reloaded configuration
Jun 20 00:12:15 fv-az72-309 dbus-daemon[658]: [system] Reloaded configuration
Jun 20 00:12:15 fv-az72-309 dbus-daemon[658]: [system] Reloaded configuration
Jun 20 00:12:15 fv-az72-309 dbus-daemon[658]: [system] Reloaded configuration
Jun 20 00:12:20 fv-az72-309 systemd[1]: Reloading.
Jun 20 00:12:21 fv-az72-309 systemd[1]: /etc/systemd/system/runner-provisioner.service:3: Invalid URL, ignoring: None
Jun 20 00:12:21 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network-online.target ignored (device units cannot be delayed).
Jun 20 00:12:21 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network.target ignored (device units cannot be delayed).
Jun 20 00:12:21 fv-az72-309 groupadd[10938]: group added to /etc/group: name=rdma, GID=127
Jun 20 00:12:21 fv-az72-309 groupadd[10938]: group added to /etc/gshadow: name=rdma
Jun 20 00:12:21 fv-az72-309 groupadd[10938]: new group: name=rdma, GID=127
Jun 20 00:12:22 fv-az72-309 systemd[1]: Reloading.
Jun 20 00:12:22 fv-az72-309 systemd[1]: /etc/systemd/system/runner-provisioner.service:3: Invalid URL, ignoring: None
Jun 20 00:12:22 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network-online.target ignored (device units cannot be delayed).
Jun 20 00:12:22 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network.target ignored (device units cannot be delayed).
Jun 20 00:12:22 fv-az72-309 systemd[1]: Reloading.
Jun 20 00:12:22 fv-az72-309 systemd[1]: /etc/systemd/system/runner-provisioner.service:3: Invalid URL, ignoring: None
Jun 20 00:12:22 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network-online.target ignored (device units cannot be delayed).
Jun 20 00:12:22 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network.target ignored (device units cannot be delayed).
Jun 20 00:12:22 fv-az72-309 systemd[1]: Starting QEMU KVM preparation - module, ksm, hugepages...
Jun 20 00:12:22 fv-az72-309 systemd[1]: Finished QEMU KVM preparation - module, ksm, hugepages.
Jun 20 00:12:22 fv-az72-309 systemd[1]: Reloading.
Jun 20 00:12:23 fv-az72-309 systemd[1]: /etc/systemd/system/runner-provisioner.service:3: Invalid URL, ignoring: None
Jun 20 00:12:23 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network-online.target ignored (device units cannot be delayed).
Jun 20 00:12:23 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network.target ignored (device units cannot be delayed).
Jun 20 00:12:23 fv-az72-309 groupadd[11140]: group added to /etc/group: name=libvirt, GID=128
Jun 20 00:12:23 fv-az72-309 groupadd[11140]: group added to /etc/gshadow: name=libvirt
Jun 20 00:12:23 fv-az72-309 groupadd[11140]: new group: name=libvirt, GID=128
Jun 20 00:12:23 fv-az72-309 useradd[11150]: new user: name=libvirt-qemu, UID=64055, GID=108, home=/var/lib/libvirt, shell=/usr/sbin/nologin, from=none
Jun 20 00:12:23 fv-az72-309 chage[11158]: changed password expiry for libvirt-qemu
Jun 20 00:12:23 fv-az72-309 chfn[11162]: changed user 'libvirt-qemu' information
Jun 20 00:12:24 fv-az72-309 groupadd[11171]: group added to /etc/group: name=libvirt-qemu, GID=64055
Jun 20 00:12:24 fv-az72-309 groupadd[11171]: group added to /etc/gshadow: name=libvirt-qemu
Jun 20 00:12:24 fv-az72-309 groupadd[11171]: new group: name=libvirt-qemu, GID=64055
Jun 20 00:12:24 fv-az72-309 gpasswd[11178]: user libvirt-qemu added by root to group libvirt-qemu
Jun 20 00:12:24 fv-az72-309 gpasswd[11188]: user runneradmin added by root to group libvirt
Jun 20 00:12:24 fv-az72-309 groupadd[11196]: group added to /etc/group: name=libvirt-dnsmasq, GID=129
Jun 20 00:12:24 fv-az72-309 groupadd[11196]: group added to /etc/gshadow: name=libvirt-dnsmasq
Jun 20 00:12:24 fv-az72-309 groupadd[11196]: new group: name=libvirt-dnsmasq, GID=129
Jun 20 00:12:24 fv-az72-309 useradd[11204]: new user: name=libvirt-dnsmasq, UID=118, GID=129, home=/var/lib/libvirt/dnsmasq, shell=/usr/sbin/nologin, from=none
Jun 20 00:12:24 fv-az72-309 chage[11212]: changed password expiry for libvirt-dnsmasq
Jun 20 00:12:24 fv-az72-309 chfn[11216]: changed user 'libvirt-dnsmasq' information
Jun 20 00:12:24 fv-az72-309 audit[11337]: AVC apparmor="STATUS" operation="profile_load" profile="unconfined" name="virt-aa-helper" pid=11337 comm="apparmor_parser"
Jun 20 00:12:24 fv-az72-309 kernel: audit: type=1400 audit(1655683944.647:34): apparmor="STATUS" operation="profile_load" profile="unconfined" name="virt-aa-helper" pid=11337 comm="apparmor_parser"
Jun 20 00:12:24 fv-az72-309 audit[11343]: AVC apparmor="STATUS" operation="profile_load" profile="unconfined" name="libvirtd" pid=11343 comm="apparmor_parser"
Jun 20 00:12:24 fv-az72-309 kernel: audit: type=1400 audit(1655683944.719:35): apparmor="STATUS" operation="profile_load" profile="unconfined" name="libvirtd" pid=11343 comm="apparmor_parser"
Jun 20 00:12:24 fv-az72-309 audit[11343]: AVC apparmor="STATUS" operation="profile_load" profile="unconfined" name="libvirtd//qemu_bridge_helper" pid=11343 comm="apparmor_parser"
Jun 20 00:12:24 fv-az72-309 kernel: audit: type=1400 audit(1655683944.723:36): apparmor="STATUS" operation="profile_load" profile="unconfined" name="libvirtd//qemu_bridge_helper" pid=11343 comm="apparmor_parser"
Jun 20 00:12:24 fv-az72-309 systemd[1]: Reloading.
Jun 20 00:12:25 fv-az72-309 systemd[1]: /etc/systemd/system/runner-provisioner.service:3: Invalid URL, ignoring: None
Jun 20 00:12:25 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network-online.target ignored (device units cannot be delayed).
Jun 20 00:12:25 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network.target ignored (device units cannot be delayed).
Jun 20 00:12:25 fv-az72-309 systemd[1]: Reloading.
Jun 20 00:12:25 fv-az72-309 systemd[1]: /etc/systemd/system/runner-provisioner.service:3: Invalid URL, ignoring: None
Jun 20 00:12:25 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network-online.target ignored (device units cannot be delayed).
Jun 20 00:12:25 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network.target ignored (device units cannot be delayed).
Jun 20 00:12:25 fv-az72-309 systemd[1]: Reloading.
Jun 20 00:12:25 fv-az72-309 systemd[1]: /etc/systemd/system/runner-provisioner.service:3: Invalid URL, ignoring: None
Jun 20 00:12:25 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network-online.target ignored (device units cannot be delayed).
Jun 20 00:12:25 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network.target ignored (device units cannot be delayed).
Jun 20 00:12:26 fv-az72-309 systemd[1]: Reloading.
Jun 20 00:12:26 fv-az72-309 systemd[1]: /etc/systemd/system/runner-provisioner.service:3: Invalid URL, ignoring: None
Jun 20 00:12:26 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network-online.target ignored (device units cannot be delayed).
Jun 20 00:12:26 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network.target ignored (device units cannot be delayed).
Jun 20 00:12:26 fv-az72-309 systemd[1]: Reloading.
Jun 20 00:12:26 fv-az72-309 systemd[1]: /etc/systemd/system/runner-provisioner.service:3: Invalid URL, ignoring: None
Jun 20 00:12:26 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network-online.target ignored (device units cannot be delayed).
Jun 20 00:12:26 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network.target ignored (device units cannot be delayed).
Jun 20 00:12:26 fv-az72-309 systemd[1]: Reloading.
Jun 20 00:12:27 fv-az72-309 systemd[1]: /etc/systemd/system/runner-provisioner.service:3: Invalid URL, ignoring: None
Jun 20 00:12:27 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network-online.target ignored (device units cannot be delayed).
Jun 20 00:12:27 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network.target ignored (device units cannot be delayed).
Jun 20 00:12:27 fv-az72-309 systemd[1]: Starting Libvirt local socket.
Jun 20 00:12:27 fv-az72-309 systemd[1]: Listening on Libvirt local socket.
Jun 20 00:12:27 fv-az72-309 systemd[1]: Listening on Libvirt local read-only socket.
Jun 20 00:12:27 fv-az72-309 systemd[1]: Created slice Virtual Machine and Container Slice.
Jun 20 00:12:27 fv-az72-309 systemd[1]: Listening on Libvirt admin socket.
Jun 20 00:12:27 fv-az72-309 systemd[1]: Listening on Virtual machine lock manager socket.
Jun 20 00:12:27 fv-az72-309 systemd[1]: Listening on Virtual machine log manager socket.
Jun 20 00:12:27 fv-az72-309 systemd[1]: Condition check resulted in Virtual Machine and Container Storage (Compatibility) being skipped.
Jun 20 00:12:27 fv-az72-309 systemd[1]: Starting Virtual Machine and Container Registration Service...
Jun 20 00:12:27 fv-az72-309 systemd[1]: Started Virtual Machine and Container Registration Service.
Jun 20 00:12:27 fv-az72-309 systemd[1]: Starting Virtualization daemon...
Jun 20 00:12:27 fv-az72-309 systemd[1]: Started Virtualization daemon.
Jun 20 00:12:27 fv-az72-309 systemd-networkd[532]: rtnl: received neighbor for link '11' we don't know about, ignoring.
Jun 20 00:12:27 fv-az72-309 systemd-networkd[532]: rtnl: received neighbor for link '11' we don't know about, ignoring.
Jun 20 00:12:27 fv-az72-309 networkd-dispatcher[674]: WARNING:Unknown index 11 seen, reloading interface list
Jun 20 00:12:27 fv-az72-309 systemd-networkd[532]: rtnl: received neighbor for link '11' we don't know about, ignoring.
Jun 20 00:12:27 fv-az72-309 systemd-networkd[532]: rtnl: received neighbor for link '11' we don't know about, ignoring.
Jun 20 00:12:27 fv-az72-309 systemd-networkd[532]: virbr0-nic: Link UP
Jun 20 00:12:27 fv-az72-309 systemd-networkd[532]: virbr0-nic: Gained carrier
Jun 20 00:12:27 fv-az72-309 kernel: virbr0: port 1(virbr0-nic) entered blocking state
Jun 20 00:12:27 fv-az72-309 kernel: virbr0: port 1(virbr0-nic) entered disabled state
Jun 20 00:12:27 fv-az72-309 kernel: device virbr0-nic entered promiscuous mode
Jun 20 00:12:27 fv-az72-309 systemd-udevd[11620]: Using default interface naming scheme 'v245'.
Jun 20 00:12:27 fv-az72-309 systemd-udevd[11620]: ethtool: autonegotiation is unset or enabled, the speed and duplex are not writable.
Jun 20 00:12:27 fv-az72-309 systemd-udevd[11619]: Using default interface naming scheme 'v245'.
Jun 20 00:12:27 fv-az72-309 systemd-udevd[11619]: ethtool: autonegotiation is unset or enabled, the speed and duplex are not writable.
Jun 20 00:12:27 fv-az72-309 systemd[1]: Reloading.
Jun 20 00:12:27 fv-az72-309 systemd[1]: /etc/systemd/system/runner-provisioner.service:3: Invalid URL, ignoring: None
Jun 20 00:12:27 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network-online.target ignored (device units cannot be delayed).
Jun 20 00:12:27 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network.target ignored (device units cannot be delayed).
Jun 20 00:12:27 fv-az72-309 systemd-networkd[532]: virbr0: Link UP
Jun 20 00:12:27 fv-az72-309 kernel: virbr0: port 1(virbr0-nic) entered blocking state
Jun 20 00:12:27 fv-az72-309 kernel: virbr0: port 1(virbr0-nic) entered listening state
Jun 20 00:12:27 fv-az72-309 dnsmasq[11724]: started, version 2.80 cachesize 150
Jun 20 00:12:27 fv-az72-309 dnsmasq[11724]: compile time options: IPv6 GNU-getopt DBus i18n IDN DHCP DHCPv6 no-Lua TFTP conntrack ipset auth nettlehash DNSSEC loop-detect inotify dumpfile
Jun 20 00:12:27 fv-az72-309 dnsmasq-dhcp[11724]: DHCP, IP range 192.168.122.2 -- 192.168.122.254, lease time 1h
Jun 20 00:12:27 fv-az72-309 dnsmasq-dhcp[11724]: DHCP, sockets bound exclusively to interface virbr0
Jun 20 00:12:27 fv-az72-309 dnsmasq[11724]: reading /etc/resolv.conf
Jun 20 00:12:27 fv-az72-309 dnsmasq[11724]: using nameserver 127.0.0.53#53
Jun 20 00:12:27 fv-az72-309 kernel: virbr0: port 1(virbr0-nic) entered disabled state
Jun 20 00:12:27 fv-az72-309 systemd-networkd[532]: virbr0-nic: Link DOWN
Jun 20 00:12:27 fv-az72-309 systemd-networkd[532]: virbr0-nic: Lost carrier
Jun 20 00:12:27 fv-az72-309 dnsmasq[11724]: read /etc/hosts - 8 addresses
Jun 20 00:12:27 fv-az72-309 dnsmasq[11724]: read /var/lib/libvirt/dnsmasq/default.addnhosts - 0 addresses
Jun 20 00:12:27 fv-az72-309 dnsmasq-dhcp[11724]: read /var/lib/libvirt/dnsmasq/default.hostsfile
Jun 20 00:12:28 fv-az72-309 systemd[1]: Reloading.
Jun 20 00:12:28 fv-az72-309 systemd[1]: /etc/systemd/system/runner-provisioner.service:3: Invalid URL, ignoring: None
Jun 20 00:12:28 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network-online.target ignored (device units cannot be delayed).
Jun 20 00:12:28 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network.target ignored (device units cannot be delayed).
Jun 20 00:12:28 fv-az72-309 systemd[1]: Reloading.
Jun 20 00:12:28 fv-az72-309 systemd[1]: /etc/systemd/system/runner-provisioner.service:3: Invalid URL, ignoring: None
Jun 20 00:12:29 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network-online.target ignored (device units cannot be delayed).
Jun 20 00:12:29 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network.target ignored (device units cannot be delayed).
Jun 20 00:12:29 fv-az72-309 systemd[1]: Reloading.
Jun 20 00:12:29 fv-az72-309 systemd[1]: /etc/systemd/system/runner-provisioner.service:3: Invalid URL, ignoring: None
Jun 20 00:12:29 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network-online.target ignored (device units cannot be delayed).
Jun 20 00:12:29 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network.target ignored (device units cannot be delayed).
Jun 20 00:12:29 fv-az72-309 systemd[1]: Reloading.
Jun 20 00:12:30 fv-az72-309 systemd[1]: /etc/systemd/system/runner-provisioner.service:3: Invalid URL, ignoring: None
Jun 20 00:12:30 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network-online.target ignored (device units cannot be delayed).
Jun 20 00:12:30 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network.target ignored (device units cannot be delayed).
Jun 20 00:12:30 fv-az72-309 systemd[1]: Reloading.
Jun 20 00:12:30 fv-az72-309 systemd[1]: /etc/systemd/system/runner-provisioner.service:3: Invalid URL, ignoring: None
Jun 20 00:12:30 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network-online.target ignored (device units cannot be delayed).
Jun 20 00:12:30 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network.target ignored (device units cannot be delayed).
Jun 20 00:12:30 fv-az72-309 systemd[1]: Reloading.
Jun 20 00:12:30 fv-az72-309 systemd[1]: /etc/systemd/system/runner-provisioner.service:3: Invalid URL, ignoring: None
Jun 20 00:12:30 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network-online.target ignored (device units cannot be delayed).
Jun 20 00:12:30 fv-az72-309 systemd[1]: dev-disk-cloud-azure_resource\x2dpart1.device: Requested dependency After=network.target ignored (device units cannot be delayed).
Jun 20 00:12:31 fv-az72-309 systemd[1]: Reached target Libvirt guests shutdown.
Jun 20 00:12:31 fv-az72-309 systemd[1]: Starting Suspend/Resume Running libvirt Guests...
Jun 20 00:12:31 fv-az72-309 systemd[1]: Listening on Virtual machine lock manager admin socket.
Jun 20 00:12:31 fv-az72-309 systemd[1]: Listening on Virtual machine log manager socket.
Jun 20 00:12:31 fv-az72-309 systemd[1]: Finished Suspend/Resume Running libvirt Guests.
Jun 20 00:12:36 fv-az72-309 dbus-daemon[658]: [system] Reloaded configuration
Jun 20 00:12:39 fv-az72-309 sudo[9090]: pam_unix(sudo:session): session closed for user root
Jun 20 00:12:39 fv-az72-309 sudo[12639]: runner : TTY=unknown ; PWD=/home/runner/work/molecule-kubevirt/molecule-kubevirt ; USER=root ; COMMAND=/usr/bin/systemctl restart libvirtd
Jun 20 00:12:39 fv-az72-309 sudo[12639]: pam_unix(sudo:session): session opened for user root by (uid=0)
Jun 20 00:12:39 fv-az72-309 systemd[1]: Stopping Virtualization daemon...
Jun 20 00:12:39 fv-az72-309 systemd[1]: libvirtd.service: Succeeded.
Jun 20 00:12:39 fv-az72-309 systemd[1]: Stopped Virtualization daemon.
Jun 20 00:12:39 fv-az72-309 systemd[1]: libvirtd.service: Found left-over process 11724 (dnsmasq) in control group while starting unit. Ignoring.
Jun 20 00:12:39 fv-az72-309 systemd[1]: This usually indicates unclean termination of a previous run, or service implementation deficiencies.
Jun 20 00:12:39 fv-az72-309 systemd[1]: libvirtd.service: Found left-over process 11725 (dnsmasq) in control group while starting unit. Ignoring.
Jun 20 00:12:39 fv-az72-309 systemd[1]: This usually indicates unclean termination of a previous run, or service implementation deficiencies.
Jun 20 00:12:39 fv-az72-309 systemd[1]: Starting Virtualization daemon...
Jun 20 00:12:39 fv-az72-309 systemd[1]: Started Virtualization daemon.
Jun 20 00:12:39 fv-az72-309 sudo[12639]: pam_unix(sudo:session): session closed for user root
Jun 20 00:12:39 fv-az72-309 sudo[12668]: runner : TTY=unknown ; PWD=/home/runner/work/molecule-kubevirt/molecule-kubevirt ; USER=root ; COMMAND=/usr/bin/dd of=/etc/apparmor.d/usr.sbin.libvirtd
Jun 20 00:12:39 fv-az72-309 sudo[12668]: pam_unix(sudo:session): session opened for user root by (uid=0)
Jun 20 00:12:39 fv-az72-309 sudo[12668]: pam_unix(sudo:session): session closed for user root
Jun 20 00:12:39 fv-az72-309 sudo[12672]: runner : TTY=unknown ; PWD=/home/runner/work/molecule-kubevirt/molecule-kubevirt ; USER=root ; COMMAND=/usr/bin/systemctl reload apparmor.service
Jun 20 00:12:39 fv-az72-309 sudo[12672]: pam_unix(sudo:session): session opened for user root by (uid=0)
Jun 20 00:12:39 fv-az72-309 systemd[1]: Reloading Load AppArmor profiles.
Jun 20 00:12:39 fv-az72-309 apparmor.systemd[12676]: Restarting AppArmor
Jun 20 00:12:39 fv-az72-309 apparmor.systemd[12676]: Reloading AppArmor profiles
Jun 20 00:12:39 fv-az72-309 audit[12686]: AVC apparmor="STATUS" operation="profile_replace" info="same as current profile, skipping" profile="unconfined" name="/usr/lib/NetworkManager/nm-dhcp-client.action" pid=12686 comm="apparmor_parser"
Jun 20 00:12:39 fv-az72-309 audit[12686]: AVC apparmor="STATUS" operation="profile_replace" info="same as current profile, skipping" profile="unconfined" name="/usr/lib/NetworkManager/nm-dhcp-helper" pid=12686 comm="apparmor_parser"
Jun 20 00:12:39 fv-az72-309 audit[12686]: AVC apparmor="S
View raw

(Sorry about that, but we can’t show files that are this big right now.)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment