Feb 9 09:03:09.958692 kernel: Linux version 5.15.148-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 11.3.1_p20221209 p3) 11.3.1 20221209, GNU ld (Gentoo 2.39 p5) 2.39.0) #1 SMP Thu Feb 8 21:14:17 -00 2024 Feb 9 09:03:09.958716 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=ae7db544026ede4699ee2036449b75950d3fb7929b25a6731d0ad396f1aa37c9 Feb 9 09:03:09.958729 kernel: BIOS-provided physical RAM map: Feb 9 09:03:09.958736 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable Feb 9 09:03:09.958742 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved Feb 9 09:03:09.958749 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Feb 9 09:03:09.958757 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000007ffd7fff] usable Feb 9 09:03:09.958764 kernel: BIOS-e820: [mem 0x000000007ffd8000-0x000000007fffffff] reserved Feb 9 09:03:09.958773 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Feb 9 09:03:09.958780 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Feb 9 09:03:09.958787 kernel: NX (Execute Disable) protection: active Feb 9 09:03:09.958794 kernel: SMBIOS 2.8 present. Feb 9 09:03:09.958801 kernel: DMI: DigitalOcean Droplet/Droplet, BIOS 20171212 12/12/2017 Feb 9 09:03:09.958808 kernel: Hypervisor detected: KVM Feb 9 09:03:09.958817 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Feb 9 09:03:09.958827 kernel: kvm-clock: cpu 0, msr 3cfaa001, primary cpu clock Feb 9 09:03:09.958835 kernel: kvm-clock: using sched offset of 4487852989 cycles Feb 9 09:03:09.958844 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Feb 9 09:03:09.958852 kernel: tsc: Detected 2294.608 MHz processor Feb 9 09:03:09.958860 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Feb 9 09:03:09.958868 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Feb 9 09:03:09.958876 kernel: last_pfn = 0x7ffd8 max_arch_pfn = 0x400000000 Feb 9 09:03:09.958884 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Feb 9 09:03:09.958894 kernel: ACPI: Early table checksum verification disabled Feb 9 09:03:09.958902 kernel: ACPI: RSDP 0x00000000000F5A50 000014 (v00 BOCHS ) Feb 9 09:03:09.958910 kernel: ACPI: RSDT 0x000000007FFE1986 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Feb 9 09:03:09.958917 kernel: ACPI: FACP 0x000000007FFE176A 000074 (v01 BOCHS BXPC 00000001 BXPC 00000001) Feb 9 09:03:09.958925 kernel: ACPI: DSDT 0x000000007FFE0040 00172A (v01 BOCHS BXPC 00000001 BXPC 00000001) Feb 9 09:03:09.958933 kernel: ACPI: FACS 0x000000007FFE0000 000040 Feb 9 09:03:09.958940 kernel: ACPI: APIC 0x000000007FFE17DE 000080 (v01 BOCHS BXPC 00000001 BXPC 00000001) Feb 9 09:03:09.958948 kernel: ACPI: HPET 0x000000007FFE185E 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Feb 9 09:03:09.958956 kernel: ACPI: SRAT 0x000000007FFE1896 0000C8 (v01 BOCHS BXPC 00000001 BXPC 00000001) Feb 9 09:03:09.958966 kernel: ACPI: WAET 0x000000007FFE195E 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Feb 9 09:03:09.958974 kernel: ACPI: Reserving FACP table memory at [mem 0x7ffe176a-0x7ffe17dd] Feb 9 09:03:09.958982 kernel: ACPI: Reserving DSDT table memory at [mem 0x7ffe0040-0x7ffe1769] Feb 9 09:03:09.958989 kernel: ACPI: Reserving FACS table memory at [mem 0x7ffe0000-0x7ffe003f] Feb 9 09:03:09.958997 kernel: ACPI: Reserving APIC table memory at [mem 0x7ffe17de-0x7ffe185d] Feb 9 09:03:09.959005 kernel: ACPI: Reserving HPET table memory at [mem 0x7ffe185e-0x7ffe1895] Feb 9 09:03:09.959013 kernel: ACPI: Reserving SRAT table memory at [mem 0x7ffe1896-0x7ffe195d] Feb 9 09:03:09.959020 kernel: ACPI: Reserving WAET table memory at [mem 0x7ffe195e-0x7ffe1985] Feb 9 09:03:09.959035 kernel: SRAT: PXM 0 -> APIC 0x00 -> Node 0 Feb 9 09:03:09.959043 kernel: SRAT: PXM 0 -> APIC 0x01 -> Node 0 Feb 9 09:03:09.959051 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x0009ffff] Feb 9 09:03:09.959059 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00100000-0x7fffffff] Feb 9 09:03:09.959068 kernel: NUMA: Node 0 [mem 0x00000000-0x0009ffff] + [mem 0x00100000-0x7ffd7fff] -> [mem 0x00000000-0x7ffd7fff] Feb 9 09:03:09.959076 kernel: NODE_DATA(0) allocated [mem 0x7ffd2000-0x7ffd7fff] Feb 9 09:03:09.959088 kernel: Zone ranges: Feb 9 09:03:09.959096 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Feb 9 09:03:09.959104 kernel: DMA32 [mem 0x0000000001000000-0x000000007ffd7fff] Feb 9 09:03:09.959113 kernel: Normal empty Feb 9 09:03:09.959121 kernel: Movable zone start for each node Feb 9 09:03:09.959129 kernel: Early memory node ranges Feb 9 09:03:09.959137 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Feb 9 09:03:09.959146 kernel: node 0: [mem 0x0000000000100000-0x000000007ffd7fff] Feb 9 09:03:09.959154 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000007ffd7fff] Feb 9 09:03:09.959165 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Feb 9 09:03:09.959174 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Feb 9 09:03:09.959182 kernel: On node 0, zone DMA32: 40 pages in unavailable ranges Feb 9 09:03:09.959191 kernel: ACPI: PM-Timer IO Port: 0x608 Feb 9 09:03:09.959199 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Feb 9 09:03:09.959207 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Feb 9 09:03:09.959215 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Feb 9 09:03:09.959224 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Feb 9 09:03:09.959232 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Feb 9 09:03:09.959243 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Feb 9 09:03:09.959251 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Feb 9 09:03:09.959260 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Feb 9 09:03:09.959268 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Feb 9 09:03:09.959276 kernel: TSC deadline timer available Feb 9 09:03:09.959288 kernel: smpboot: Allowing 2 CPUs, 0 hotplug CPUs Feb 9 09:03:09.959315 kernel: [mem 0x80000000-0xfeffbfff] available for PCI devices Feb 9 09:03:09.959327 kernel: Booting paravirtualized kernel on KVM Feb 9 09:03:09.959341 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Feb 9 09:03:09.965419 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:512 nr_cpu_ids:2 nr_node_ids:1 Feb 9 09:03:09.965440 kernel: percpu: Embedded 55 pages/cpu s185624 r8192 d31464 u1048576 Feb 9 09:03:09.965451 kernel: pcpu-alloc: s185624 r8192 d31464 u1048576 alloc=1*2097152 Feb 9 09:03:09.965462 kernel: pcpu-alloc: [0] 0 1 Feb 9 09:03:09.965472 kernel: kvm-guest: stealtime: cpu 0, msr 7dc1c0c0 Feb 9 09:03:09.965482 kernel: kvm-guest: PV spinlocks disabled, no host support Feb 9 09:03:09.965492 kernel: Built 1 zonelists, mobility grouping on. Total pages: 515800 Feb 9 09:03:09.965503 kernel: Policy zone: DMA32 Feb 9 09:03:09.965515 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=ae7db544026ede4699ee2036449b75950d3fb7929b25a6731d0ad396f1aa37c9 Feb 9 09:03:09.965530 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Feb 9 09:03:09.965539 kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Feb 9 09:03:09.965548 kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear) Feb 9 09:03:09.965556 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Feb 9 09:03:09.965566 kernel: Memory: 1975320K/2096600K available (12294K kernel code, 2275K rwdata, 13700K rodata, 45496K init, 4048K bss, 121020K reserved, 0K cma-reserved) Feb 9 09:03:09.965574 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Feb 9 09:03:09.965583 kernel: Kernel/User page tables isolation: enabled Feb 9 09:03:09.965592 kernel: ftrace: allocating 34475 entries in 135 pages Feb 9 09:03:09.965604 kernel: ftrace: allocated 135 pages with 4 groups Feb 9 09:03:09.965612 kernel: rcu: Hierarchical RCU implementation. Feb 9 09:03:09.965622 kernel: rcu: RCU event tracing is enabled. Feb 9 09:03:09.965631 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Feb 9 09:03:09.965640 kernel: Rude variant of Tasks RCU enabled. Feb 9 09:03:09.965648 kernel: Tracing variant of Tasks RCU enabled. Feb 9 09:03:09.965666 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Feb 9 09:03:09.965681 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Feb 9 09:03:09.965695 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 Feb 9 09:03:09.965713 kernel: random: crng init done Feb 9 09:03:09.965734 kernel: Console: colour VGA+ 80x25 Feb 9 09:03:09.965754 kernel: printk: console [tty0] enabled Feb 9 09:03:09.965775 kernel: printk: console [ttyS0] enabled Feb 9 09:03:09.965791 kernel: ACPI: Core revision 20210730 Feb 9 09:03:09.965812 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Feb 9 09:03:09.965833 kernel: APIC: Switch to symmetric I/O mode setup Feb 9 09:03:09.965855 kernel: x2apic enabled Feb 9 09:03:09.965876 kernel: Switched APIC routing to physical x2apic. Feb 9 09:03:09.965903 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Feb 9 09:03:09.965919 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x21134f58f0d, max_idle_ns: 440795217993 ns Feb 9 09:03:09.965939 kernel: Calibrating delay loop (skipped) preset value.. 4589.21 BogoMIPS (lpj=2294608) Feb 9 09:03:09.965961 kernel: Last level iTLB entries: 4KB 0, 2MB 0, 4MB 0 Feb 9 09:03:09.965981 kernel: Last level dTLB entries: 4KB 0, 2MB 0, 4MB 0, 1GB 0 Feb 9 09:03:09.965998 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Feb 9 09:03:09.966015 kernel: Spectre V2 : Mitigation: Retpolines Feb 9 09:03:09.966032 kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch Feb 9 09:03:09.966049 kernel: Spectre V2 : Spectre v2 / SpectreRSB : Filling RSB on VMEXIT Feb 9 09:03:09.966070 kernel: Spectre V2 : Enabling Restricted Speculation for firmware calls Feb 9 09:03:09.966099 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Feb 9 09:03:09.966118 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl and seccomp Feb 9 09:03:09.966150 kernel: MDS: Mitigation: Clear CPU buffers Feb 9 09:03:09.966173 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode Feb 9 09:03:09.966196 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Feb 9 09:03:09.966218 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Feb 9 09:03:09.966236 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Feb 9 09:03:09.966254 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Feb 9 09:03:09.966273 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'standard' format. Feb 9 09:03:09.966294 kernel: Freeing SMP alternatives memory: 32K Feb 9 09:03:09.966315 kernel: pid_max: default: 32768 minimum: 301 Feb 9 09:03:09.966338 kernel: LSM: Security Framework initializing Feb 9 09:03:09.966380 kernel: SELinux: Initializing. Feb 9 09:03:09.966398 kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Feb 9 09:03:09.966416 kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Feb 9 09:03:09.966437 kernel: smpboot: CPU0: Intel DO-Regular (family: 0x6, model: 0x3f, stepping: 0x2) Feb 9 09:03:09.966455 kernel: Performance Events: unsupported p6 CPU model 63 no PMU driver, software events only. Feb 9 09:03:09.966474 kernel: signal: max sigframe size: 1776 Feb 9 09:03:09.966492 kernel: rcu: Hierarchical SRCU implementation. Feb 9 09:03:09.966511 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Feb 9 09:03:09.966541 kernel: smp: Bringing up secondary CPUs ... Feb 9 09:03:09.966554 kernel: x86: Booting SMP configuration: Feb 9 09:03:09.966566 kernel: .... node #0, CPUs: #1 Feb 9 09:03:09.966579 kernel: kvm-clock: cpu 1, msr 3cfaa041, secondary cpu clock Feb 9 09:03:09.966593 kernel: kvm-guest: stealtime: cpu 1, msr 7dd1c0c0 Feb 9 09:03:09.966611 kernel: smp: Brought up 1 node, 2 CPUs Feb 9 09:03:09.966623 kernel: smpboot: Max logical packages: 1 Feb 9 09:03:09.966635 kernel: smpboot: Total of 2 processors activated (9178.43 BogoMIPS) Feb 9 09:03:09.966649 kernel: devtmpfs: initialized Feb 9 09:03:09.966661 kernel: x86/mm: Memory block size: 128MB Feb 9 09:03:09.966673 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Feb 9 09:03:09.966687 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Feb 9 09:03:09.966701 kernel: pinctrl core: initialized pinctrl subsystem Feb 9 09:03:09.966715 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Feb 9 09:03:09.966739 kernel: audit: initializing netlink subsys (disabled) Feb 9 09:03:09.966757 kernel: audit: type=2000 audit(1707469389.154:1): state=initialized audit_enabled=0 res=1 Feb 9 09:03:09.966775 kernel: thermal_sys: Registered thermal governor 'step_wise' Feb 9 09:03:09.966793 kernel: thermal_sys: Registered thermal governor 'user_space' Feb 9 09:03:09.966811 kernel: cpuidle: using governor menu Feb 9 09:03:09.966829 kernel: ACPI: bus type PCI registered Feb 9 09:03:09.966847 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Feb 9 09:03:09.966865 kernel: dca service started, version 1.12.1 Feb 9 09:03:09.966883 kernel: PCI: Using configuration type 1 for base access Feb 9 09:03:09.966904 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Feb 9 09:03:09.966922 kernel: HugeTLB registered 2.00 MiB page size, pre-allocated 0 pages Feb 9 09:03:09.966941 kernel: ACPI: Added _OSI(Module Device) Feb 9 09:03:09.966969 kernel: ACPI: Added _OSI(Processor Device) Feb 9 09:03:09.966987 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Feb 9 09:03:09.967009 kernel: ACPI: Added _OSI(Processor Aggregator Device) Feb 9 09:03:09.967027 kernel: ACPI: Added _OSI(Linux-Dell-Video) Feb 9 09:03:09.967045 kernel: ACPI: Added _OSI(Linux-Lenovo-NV-HDMI-Audio) Feb 9 09:03:09.967064 kernel: ACPI: Added _OSI(Linux-HPI-Hybrid-Graphics) Feb 9 09:03:09.967092 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Feb 9 09:03:09.967106 kernel: ACPI: Interpreter enabled Feb 9 09:03:09.967119 kernel: ACPI: PM: (supports S0 S5) Feb 9 09:03:09.967136 kernel: ACPI: Using IOAPIC for interrupt routing Feb 9 09:03:09.967155 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Feb 9 09:03:09.967173 kernel: ACPI: Enabled 2 GPEs in block 00 to 0F Feb 9 09:03:09.967191 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Feb 9 09:03:09.967537 kernel: acpi PNP0A03:00: _OSC: OS supports [ASPM ClockPM Segments MSI HPX-Type3] Feb 9 09:03:09.967722 kernel: acpi PNP0A03:00: fail to add MMCONFIG information, can't access extended PCI configuration space under this bridge. Feb 9 09:03:09.967752 kernel: acpiphp: Slot [3] registered Feb 9 09:03:09.967775 kernel: acpiphp: Slot [4] registered Feb 9 09:03:09.967798 kernel: acpiphp: Slot [5] registered Feb 9 09:03:09.967820 kernel: acpiphp: Slot [6] registered Feb 9 09:03:09.967842 kernel: acpiphp: Slot [7] registered Feb 9 09:03:09.967863 kernel: acpiphp: Slot [8] registered Feb 9 09:03:09.967881 kernel: acpiphp: Slot [9] registered Feb 9 09:03:09.967904 kernel: acpiphp: Slot [10] registered Feb 9 09:03:09.967934 kernel: acpiphp: Slot [11] registered Feb 9 09:03:09.967948 kernel: acpiphp: Slot [12] registered Feb 9 09:03:09.967963 kernel: acpiphp: Slot [13] registered Feb 9 09:03:09.967975 kernel: acpiphp: Slot [14] registered Feb 9 09:03:09.967988 kernel: acpiphp: Slot [15] registered Feb 9 09:03:09.968005 kernel: acpiphp: Slot [16] registered Feb 9 09:03:09.968023 kernel: acpiphp: Slot [17] registered Feb 9 09:03:09.968042 kernel: acpiphp: Slot [18] registered Feb 9 09:03:09.968060 kernel: acpiphp: Slot [19] registered Feb 9 09:03:09.968082 kernel: acpiphp: Slot [20] registered Feb 9 09:03:09.968100 kernel: acpiphp: Slot [21] registered Feb 9 09:03:09.968118 kernel: acpiphp: Slot [22] registered Feb 9 09:03:09.968136 kernel: acpiphp: Slot [23] registered Feb 9 09:03:09.968153 kernel: acpiphp: Slot [24] registered Feb 9 09:03:09.968171 kernel: acpiphp: Slot [25] registered Feb 9 09:03:09.968189 kernel: acpiphp: Slot [26] registered Feb 9 09:03:09.968207 kernel: acpiphp: Slot [27] registered Feb 9 09:03:09.968225 kernel: acpiphp: Slot [28] registered Feb 9 09:03:09.968251 kernel: acpiphp: Slot [29] registered Feb 9 09:03:09.968275 kernel: acpiphp: Slot [30] registered Feb 9 09:03:09.968295 kernel: acpiphp: Slot [31] registered Feb 9 09:03:09.968313 kernel: PCI host bridge to bus 0000:00 Feb 9 09:03:09.968565 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Feb 9 09:03:09.968661 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Feb 9 09:03:09.968745 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Feb 9 09:03:09.968828 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xfebfffff window] Feb 9 09:03:09.968916 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x17fffffff window] Feb 9 09:03:09.968996 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Feb 9 09:03:09.969112 kernel: pci 0000:00:00.0: [8086:1237] type 00 class 0x060000 Feb 9 09:03:09.969215 kernel: pci 0000:00:01.0: [8086:7000] type 00 class 0x060100 Feb 9 09:03:09.969316 kernel: pci 0000:00:01.1: [8086:7010] type 00 class 0x010180 Feb 9 09:03:09.969426 kernel: pci 0000:00:01.1: reg 0x20: [io 0xc1e0-0xc1ef] Feb 9 09:03:09.969522 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x10: [io 0x01f0-0x01f7] Feb 9 09:03:09.969653 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x14: [io 0x03f6] Feb 9 09:03:09.969747 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x18: [io 0x0170-0x0177] Feb 9 09:03:09.969837 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x1c: [io 0x0376] Feb 9 09:03:09.969948 kernel: pci 0000:00:01.2: [8086:7020] type 00 class 0x0c0300 Feb 9 09:03:09.970041 kernel: pci 0000:00:01.2: reg 0x20: [io 0xc180-0xc19f] Feb 9 09:03:09.970155 kernel: pci 0000:00:01.3: [8086:7113] type 00 class 0x068000 Feb 9 09:03:09.970295 kernel: pci 0000:00:01.3: quirk: [io 0x0600-0x063f] claimed by PIIX4 ACPI Feb 9 09:03:09.970442 kernel: pci 0000:00:01.3: quirk: [io 0x0700-0x070f] claimed by PIIX4 SMB Feb 9 09:03:09.970585 kernel: pci 0000:00:02.0: [1af4:1050] type 00 class 0x030000 Feb 9 09:03:09.970730 kernel: pci 0000:00:02.0: reg 0x10: [mem 0xfe000000-0xfe7fffff pref] Feb 9 09:03:09.970878 kernel: pci 0000:00:02.0: reg 0x18: [mem 0xfe800000-0xfe803fff 64bit pref] Feb 9 09:03:09.971026 kernel: pci 0000:00:02.0: reg 0x20: [mem 0xfebf0000-0xfebf0fff] Feb 9 09:03:09.971165 kernel: pci 0000:00:02.0: reg 0x30: [mem 0xfebe0000-0xfebeffff pref] Feb 9 09:03:09.971298 kernel: pci 0000:00:02.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Feb 9 09:03:09.971487 kernel: pci 0000:00:03.0: [1af4:1000] type 00 class 0x020000 Feb 9 09:03:09.971588 kernel: pci 0000:00:03.0: reg 0x10: [io 0xc1a0-0xc1bf] Feb 9 09:03:09.971684 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xfebf1000-0xfebf1fff] Feb 9 09:03:09.971777 kernel: pci 0000:00:03.0: reg 0x20: [mem 0xfe804000-0xfe807fff 64bit pref] Feb 9 09:03:09.971901 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 Feb 9 09:03:09.972020 kernel: pci 0000:00:04.0: reg 0x10: [io 0xc1c0-0xc1df] Feb 9 09:03:09.972116 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xfebf2000-0xfebf2fff] Feb 9 09:03:09.972210 kernel: pci 0000:00:04.0: reg 0x20: [mem 0xfe808000-0xfe80bfff 64bit pref] Feb 9 09:03:09.972313 kernel: pci 0000:00:05.0: [1af4:1004] type 00 class 0x010000 Feb 9 09:03:09.972437 kernel: pci 0000:00:05.0: reg 0x10: [io 0xc100-0xc13f] Feb 9 09:03:09.972534 kernel: pci 0000:00:05.0: reg 0x14: [mem 0xfebf3000-0xfebf3fff] Feb 9 09:03:09.972631 kernel: pci 0000:00:05.0: reg 0x20: [mem 0xfe80c000-0xfe80ffff 64bit pref] Feb 9 09:03:09.972740 kernel: pci 0000:00:06.0: [1af4:1001] type 00 class 0x010000 Feb 9 09:03:09.972835 kernel: pci 0000:00:06.0: reg 0x10: [io 0xc000-0xc07f] Feb 9 09:03:09.972948 kernel: pci 0000:00:06.0: reg 0x14: [mem 0xfebf4000-0xfebf4fff] Feb 9 09:03:09.973042 kernel: pci 0000:00:06.0: reg 0x20: [mem 0xfe810000-0xfe813fff 64bit pref] Feb 9 09:03:09.973158 kernel: pci 0000:00:07.0: [1af4:1001] type 00 class 0x010000 Feb 9 09:03:09.973305 kernel: pci 0000:00:07.0: reg 0x10: [io 0xc080-0xc0ff] Feb 9 09:03:09.973458 kernel: pci 0000:00:07.0: reg 0x14: [mem 0xfebf5000-0xfebf5fff] Feb 9 09:03:09.973606 kernel: pci 0000:00:07.0: reg 0x20: [mem 0xfe814000-0xfe817fff 64bit pref] Feb 9 09:03:09.973795 kernel: pci 0000:00:08.0: [1af4:1002] type 00 class 0x00ff00 Feb 9 09:03:09.973971 kernel: pci 0000:00:08.0: reg 0x10: [io 0xc140-0xc17f] Feb 9 09:03:09.974129 kernel: pci 0000:00:08.0: reg 0x20: [mem 0xfe818000-0xfe81bfff 64bit pref] Feb 9 09:03:09.974152 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Feb 9 09:03:09.974171 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Feb 9 09:03:09.974189 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Feb 9 09:03:09.974216 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Feb 9 09:03:09.974234 kernel: ACPI: PCI: Interrupt link LNKS configured for IRQ 9 Feb 9 09:03:09.974252 kernel: iommu: Default domain type: Translated Feb 9 09:03:09.974270 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Feb 9 09:03:09.974431 kernel: pci 0000:00:02.0: vgaarb: setting as boot VGA device Feb 9 09:03:09.974580 kernel: pci 0000:00:02.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Feb 9 09:03:09.974750 kernel: pci 0000:00:02.0: vgaarb: bridge control possible Feb 9 09:03:09.974773 kernel: vgaarb: loaded Feb 9 09:03:09.974797 kernel: pps_core: LinuxPPS API ver. 1 registered Feb 9 09:03:09.974816 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Feb 9 09:03:09.974834 kernel: PTP clock support registered Feb 9 09:03:09.974852 kernel: PCI: Using ACPI for IRQ routing Feb 9 09:03:09.974870 kernel: PCI: pci_cache_line_size set to 64 bytes Feb 9 09:03:09.974888 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] Feb 9 09:03:09.974906 kernel: e820: reserve RAM buffer [mem 0x7ffd8000-0x7fffffff] Feb 9 09:03:09.974925 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Feb 9 09:03:09.974943 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Feb 9 09:03:09.974964 kernel: clocksource: Switched to clocksource kvm-clock Feb 9 09:03:09.974981 kernel: VFS: Disk quotas dquot_6.6.0 Feb 9 09:03:09.974999 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Feb 9 09:03:09.975018 kernel: pnp: PnP ACPI init Feb 9 09:03:09.975036 kernel: pnp: PnP ACPI: found 4 devices Feb 9 09:03:09.975053 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Feb 9 09:03:09.975071 kernel: NET: Registered PF_INET protocol family Feb 9 09:03:09.975089 kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear) Feb 9 09:03:09.975107 kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear) Feb 9 09:03:09.975128 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Feb 9 09:03:09.975146 kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear) Feb 9 09:03:09.975164 kernel: TCP bind hash table entries: 16384 (order: 6, 262144 bytes, linear) Feb 9 09:03:09.975182 kernel: TCP: Hash tables configured (established 16384 bind 16384) Feb 9 09:03:09.975200 kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear) Feb 9 09:03:09.975218 kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear) Feb 9 09:03:09.975236 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Feb 9 09:03:09.975253 kernel: NET: Registered PF_XDP protocol family Feb 9 09:03:09.983547 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Feb 9 09:03:09.983714 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Feb 9 09:03:09.983833 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Feb 9 09:03:09.983952 kernel: pci_bus 0000:00: resource 7 [mem 0x80000000-0xfebfffff window] Feb 9 09:03:09.984077 kernel: pci_bus 0000:00: resource 8 [mem 0x100000000-0x17fffffff window] Feb 9 09:03:09.984241 kernel: pci 0000:00:01.0: PIIX3: Enabling Passive Release Feb 9 09:03:09.984433 kernel: pci 0000:00:00.0: Limiting direct PCI/PCI transfers Feb 9 09:03:09.984599 kernel: pci 0000:00:01.0: Activating ISA DMA hang workarounds Feb 9 09:03:09.984634 kernel: ACPI: \_SB_.LNKD: Enabled at IRQ 11 Feb 9 09:03:09.984784 kernel: pci 0000:00:01.2: quirk_usb_early_handoff+0x0/0x730 took 41209 usecs Feb 9 09:03:09.984806 kernel: PCI: CLS 0 bytes, default 64 Feb 9 09:03:09.984825 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer Feb 9 09:03:09.984844 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x21134f58f0d, max_idle_ns: 440795217993 ns Feb 9 09:03:09.984862 kernel: Initialise system trusted keyrings Feb 9 09:03:09.984880 kernel: workingset: timestamp_bits=39 max_order=19 bucket_order=0 Feb 9 09:03:09.984899 kernel: Key type asymmetric registered Feb 9 09:03:09.984916 kernel: Asymmetric key parser 'x509' registered Feb 9 09:03:09.984939 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 249) Feb 9 09:03:09.984956 kernel: io scheduler mq-deadline registered Feb 9 09:03:09.984974 kernel: io scheduler kyber registered Feb 9 09:03:09.984992 kernel: io scheduler bfq registered Feb 9 09:03:09.985010 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Feb 9 09:03:09.985028 kernel: ACPI: \_SB_.LNKB: Enabled at IRQ 10 Feb 9 09:03:09.985046 kernel: ACPI: \_SB_.LNKC: Enabled at IRQ 11 Feb 9 09:03:09.985066 kernel: ACPI: \_SB_.LNKA: Enabled at IRQ 10 Feb 9 09:03:09.985093 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Feb 9 09:03:09.985120 kernel: 00:00: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Feb 9 09:03:09.985141 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Feb 9 09:03:09.985154 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Feb 9 09:03:09.985166 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Feb 9 09:03:09.985379 kernel: rtc_cmos 00:03: RTC can wake from S4 Feb 9 09:03:09.985414 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Feb 9 09:03:09.985552 kernel: rtc_cmos 00:03: registered as rtc0 Feb 9 09:03:09.985699 kernel: rtc_cmos 00:03: setting system clock to 2024-02-09T09:03:09 UTC (1707469389) Feb 9 09:03:09.985844 kernel: rtc_cmos 00:03: alarms up to one day, y3k, 242 bytes nvram Feb 9 09:03:09.985862 kernel: intel_pstate: CPU model not supported Feb 9 09:03:09.985876 kernel: NET: Registered PF_INET6 protocol family Feb 9 09:03:09.985888 kernel: Segment Routing with IPv6 Feb 9 09:03:09.985900 kernel: In-situ OAM (IOAM) with IPv6 Feb 9 09:03:09.985915 kernel: NET: Registered PF_PACKET protocol family Feb 9 09:03:09.985933 kernel: Key type dns_resolver registered Feb 9 09:03:09.985951 kernel: IPI shorthand broadcast: enabled Feb 9 09:03:09.985969 kernel: sched_clock: Marking stable (869214713, 194476469)->(1304618381, -240927199) Feb 9 09:03:09.985992 kernel: registered taskstats version 1 Feb 9 09:03:09.986010 kernel: Loading compiled-in X.509 certificates Feb 9 09:03:09.986029 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 5.15.148-flatcar: e9d857ae0e8100c174221878afd1046acbb054a6' Feb 9 09:03:09.986046 kernel: Key type .fscrypt registered Feb 9 09:03:09.986064 kernel: Key type fscrypt-provisioning registered Feb 9 09:03:09.986083 kernel: ima: No TPM chip found, activating TPM-bypass! Feb 9 09:03:09.986101 kernel: ima: Allocated hash algorithm: sha1 Feb 9 09:03:09.986119 kernel: ima: No architecture policies found Feb 9 09:03:09.986137 kernel: Freeing unused kernel image (initmem) memory: 45496K Feb 9 09:03:09.986158 kernel: Write protecting the kernel read-only data: 28672k Feb 9 09:03:09.986175 kernel: Freeing unused kernel image (text/rodata gap) memory: 2040K Feb 9 09:03:09.986193 kernel: Freeing unused kernel image (rodata/data gap) memory: 636K Feb 9 09:03:09.986211 kernel: Run /init as init process Feb 9 09:03:09.986229 kernel: with arguments: Feb 9 09:03:09.986248 kernel: /init Feb 9 09:03:09.986290 kernel: with environment: Feb 9 09:03:09.986312 kernel: HOME=/ Feb 9 09:03:09.986330 kernel: TERM=linux Feb 9 09:03:09.986369 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Feb 9 09:03:09.986392 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) Feb 9 09:03:09.986418 systemd[1]: Detected virtualization kvm. Feb 9 09:03:09.986440 systemd[1]: Detected architecture x86-64. Feb 9 09:03:09.986456 systemd[1]: Running in initrd. Feb 9 09:03:09.986470 systemd[1]: No hostname configured, using default hostname. Feb 9 09:03:09.986485 systemd[1]: Hostname set to . Feb 9 09:03:09.986511 systemd[1]: Initializing machine ID from VM UUID. Feb 9 09:03:09.986531 systemd[1]: Queued start job for default target initrd.target. Feb 9 09:03:09.986551 systemd[1]: Started systemd-ask-password-console.path. Feb 9 09:03:09.986570 systemd[1]: Reached target cryptsetup.target. Feb 9 09:03:09.986590 systemd[1]: Reached target paths.target. Feb 9 09:03:09.986609 systemd[1]: Reached target slices.target. Feb 9 09:03:09.986630 systemd[1]: Reached target swap.target. Feb 9 09:03:09.986646 systemd[1]: Reached target timers.target. Feb 9 09:03:09.986673 systemd[1]: Listening on iscsid.socket. Feb 9 09:03:09.986693 systemd[1]: Listening on iscsiuio.socket. Feb 9 09:03:09.986713 systemd[1]: Listening on systemd-journald-audit.socket. Feb 9 09:03:09.986733 systemd[1]: Listening on systemd-journald-dev-log.socket. Feb 9 09:03:09.986758 systemd[1]: Listening on systemd-journald.socket. Feb 9 09:03:09.986786 systemd[1]: Listening on systemd-networkd.socket. Feb 9 09:03:09.986806 systemd[1]: Listening on systemd-udevd-control.socket. Feb 9 09:03:09.986825 systemd[1]: Listening on systemd-udevd-kernel.socket. Feb 9 09:03:09.986857 systemd[1]: Reached target sockets.target. Feb 9 09:03:09.986876 systemd[1]: Starting kmod-static-nodes.service... Feb 9 09:03:09.986891 systemd[1]: Finished network-cleanup.service. Feb 9 09:03:09.986911 systemd[1]: Starting systemd-fsck-usr.service... Feb 9 09:03:09.986931 systemd[1]: Starting systemd-journald.service... Feb 9 09:03:09.986945 systemd[1]: Starting systemd-modules-load.service... Feb 9 09:03:09.986968 systemd[1]: Starting systemd-resolved.service... Feb 9 09:03:09.986983 systemd[1]: Starting systemd-vconsole-setup.service... Feb 9 09:03:09.986998 systemd[1]: Finished kmod-static-nodes.service. Feb 9 09:03:09.987019 systemd-journald[184]: Journal started Feb 9 09:03:09.987114 systemd-journald[184]: Runtime Journal (/run/log/journal/ed9135f0930245b7812d42f261728148) is 4.9M, max 39.5M, 34.5M free. Feb 9 09:03:09.959478 systemd-modules-load[185]: Inserted module 'overlay' Feb 9 09:03:10.066756 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Feb 9 09:03:10.066806 kernel: Bridge firewalling registered Feb 9 09:03:10.066826 kernel: SCSI subsystem initialized Feb 9 09:03:10.066849 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Feb 9 09:03:10.066875 kernel: device-mapper: uevent: version 1.0.3 Feb 9 09:03:10.066899 kernel: device-mapper: ioctl: 4.45.0-ioctl (2021-03-22) initialised: dm-devel@redhat.com Feb 9 09:03:10.066925 systemd[1]: Started systemd-journald.service. Feb 9 09:03:10.066957 kernel: audit: type=1130 audit(1707469390.058:2): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:10.058000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:10.004203 systemd-modules-load[185]: Inserted module 'br_netfilter' Feb 9 09:03:10.073123 kernel: audit: type=1130 audit(1707469390.067:3): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:10.067000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:10.030069 systemd-resolved[186]: Positive Trust Anchors: Feb 9 09:03:10.080554 kernel: audit: type=1130 audit(1707469390.073:4): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:10.080595 kernel: audit: type=1130 audit(1707469390.079:5): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:10.073000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:10.079000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:10.030090 systemd-resolved[186]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Feb 9 09:03:10.091591 kernel: audit: type=1130 audit(1707469390.085:6): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:10.085000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:10.030164 systemd-resolved[186]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test Feb 9 09:03:10.099979 kernel: audit: type=1130 audit(1707469390.091:7): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:10.091000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:10.037834 systemd-resolved[186]: Defaulting to hostname 'linux'. Feb 9 09:03:10.043120 systemd-modules-load[185]: Inserted module 'dm_multipath' Feb 9 09:03:10.067833 systemd[1]: Started systemd-resolved.service. Feb 9 09:03:10.074062 systemd[1]: Finished systemd-fsck-usr.service. Feb 9 09:03:10.080335 systemd[1]: Finished systemd-modules-load.service. Feb 9 09:03:10.086040 systemd[1]: Finished systemd-vconsole-setup.service. Feb 9 09:03:10.092561 systemd[1]: Reached target nss-lookup.target. Feb 9 09:03:10.101985 systemd[1]: Starting dracut-cmdline-ask.service... Feb 9 09:03:10.104166 systemd[1]: Starting systemd-sysctl.service... Feb 9 09:03:10.111385 systemd[1]: Starting systemd-tmpfiles-setup-dev.service... Feb 9 09:03:10.125976 systemd[1]: Finished systemd-sysctl.service. Feb 9 09:03:10.148336 kernel: audit: type=1130 audit(1707469390.126:8): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:10.126000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:10.148000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:10.126932 systemd[1]: Finished systemd-tmpfiles-setup-dev.service. Feb 9 09:03:10.155711 kernel: audit: type=1130 audit(1707469390.148:9): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:10.148922 systemd[1]: Finished dracut-cmdline-ask.service. Feb 9 09:03:10.154000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:10.156776 systemd[1]: Starting dracut-cmdline.service... Feb 9 09:03:10.163735 kernel: audit: type=1130 audit(1707469390.154:10): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:10.177094 dracut-cmdline[206]: dracut-dracut-053 Feb 9 09:03:10.181443 dracut-cmdline[206]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=ae7db544026ede4699ee2036449b75950d3fb7929b25a6731d0ad396f1aa37c9 Feb 9 09:03:10.290409 kernel: Loading iSCSI transport class v2.0-870. Feb 9 09:03:10.310407 kernel: iscsi: registered transport (tcp) Feb 9 09:03:10.352137 kernel: iscsi: registered transport (qla4xxx) Feb 9 09:03:10.352235 kernel: QLogic iSCSI HBA Driver Feb 9 09:03:10.410008 systemd[1]: Finished dracut-cmdline.service. Feb 9 09:03:10.410000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:10.412428 systemd[1]: Starting dracut-pre-udev.service... Feb 9 09:03:10.478445 kernel: raid6: avx2x4 gen() 15121 MB/s Feb 9 09:03:10.496423 kernel: raid6: avx2x4 xor() 7552 MB/s Feb 9 09:03:10.514437 kernel: raid6: avx2x2 gen() 13278 MB/s Feb 9 09:03:10.532438 kernel: raid6: avx2x2 xor() 10901 MB/s Feb 9 09:03:10.550420 kernel: raid6: avx2x1 gen() 10123 MB/s Feb 9 09:03:10.568446 kernel: raid6: avx2x1 xor() 9497 MB/s Feb 9 09:03:10.586431 kernel: raid6: sse2x4 gen() 8608 MB/s Feb 9 09:03:10.604424 kernel: raid6: sse2x4 xor() 5022 MB/s Feb 9 09:03:10.622415 kernel: raid6: sse2x2 gen() 11973 MB/s Feb 9 09:03:10.640421 kernel: raid6: sse2x2 xor() 7428 MB/s Feb 9 09:03:10.658421 kernel: raid6: sse2x1 gen() 11406 MB/s Feb 9 09:03:10.676921 kernel: raid6: sse2x1 xor() 5583 MB/s Feb 9 09:03:10.677008 kernel: raid6: using algorithm avx2x4 gen() 15121 MB/s Feb 9 09:03:10.677029 kernel: raid6: .... xor() 7552 MB/s, rmw enabled Feb 9 09:03:10.678169 kernel: raid6: using avx2x2 recovery algorithm Feb 9 09:03:10.695461 kernel: xor: automatically using best checksumming function avx Feb 9 09:03:10.812393 kernel: Btrfs loaded, crc32c=crc32c-intel, zoned=no, fsverity=no Feb 9 09:03:10.827150 systemd[1]: Finished dracut-pre-udev.service. Feb 9 09:03:10.827000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:10.828000 audit: BPF prog-id=7 op=LOAD Feb 9 09:03:10.828000 audit: BPF prog-id=8 op=LOAD Feb 9 09:03:10.829072 systemd[1]: Starting systemd-udevd.service... Feb 9 09:03:10.846358 systemd-udevd[383]: Using default interface naming scheme 'v252'. Feb 9 09:03:10.854472 systemd[1]: Started systemd-udevd.service. Feb 9 09:03:10.858000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:10.860498 systemd[1]: Starting dracut-pre-trigger.service... Feb 9 09:03:10.881461 dracut-pre-trigger[399]: rd.md=0: removing MD RAID activation Feb 9 09:03:10.928000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:10.928197 systemd[1]: Finished dracut-pre-trigger.service. Feb 9 09:03:10.930110 systemd[1]: Starting systemd-udev-trigger.service... Feb 9 09:03:10.987000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:10.987443 systemd[1]: Finished systemd-udev-trigger.service. Feb 9 09:03:11.062379 kernel: virtio_blk virtio4: [vda] 125829120 512-byte logical blocks (64.4 GB/60.0 GiB) Feb 9 09:03:11.075429 kernel: cryptd: max_cpu_qlen set to 1000 Feb 9 09:03:11.084478 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Feb 9 09:03:11.084544 kernel: GPT:9289727 != 125829119 Feb 9 09:03:11.084557 kernel: GPT:Alternate GPT header not at the end of the disk. Feb 9 09:03:11.084569 kernel: GPT:9289727 != 125829119 Feb 9 09:03:11.084580 kernel: GPT: Use GNU Parted to correct GPT errors. Feb 9 09:03:11.086812 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Feb 9 09:03:11.100372 kernel: virtio_blk virtio5: [vdb] 948 512-byte logical blocks (485 kB/474 KiB) Feb 9 09:03:11.102390 kernel: scsi host0: Virtio SCSI HBA Feb 9 09:03:11.136380 kernel: AVX2 version of gcm_enc/dec engaged. Feb 9 09:03:11.147384 kernel: AES CTR mode by8 optimization enabled Feb 9 09:03:11.157783 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device. Feb 9 09:03:11.160221 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/vda6 scanned by (udev-worker) (436) Feb 9 09:03:11.161063 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device. Feb 9 09:03:11.180528 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device. Feb 9 09:03:11.279873 kernel: libata version 3.00 loaded. Feb 9 09:03:11.279912 kernel: ACPI: bus type USB registered Feb 9 09:03:11.279931 kernel: usbcore: registered new interface driver usbfs Feb 9 09:03:11.279949 kernel: usbcore: registered new interface driver hub Feb 9 09:03:11.279980 kernel: usbcore: registered new device driver usb Feb 9 09:03:11.279997 kernel: ata_piix 0000:00:01.1: version 2.13 Feb 9 09:03:11.280255 kernel: scsi host1: ata_piix Feb 9 09:03:11.280506 kernel: ehci_hcd: USB 2.0 'Enhanced' Host Controller (EHCI) Driver Feb 9 09:03:11.280527 kernel: ehci-pci: EHCI PCI platform driver Feb 9 09:03:11.280545 kernel: scsi host2: ata_piix Feb 9 09:03:11.280724 kernel: ata1: PATA max MWDMA2 cmd 0x1f0 ctl 0x3f6 bmdma 0xc1e0 irq 14 Feb 9 09:03:11.280745 kernel: ata2: PATA max MWDMA2 cmd 0x170 ctl 0x376 bmdma 0xc1e8 irq 15 Feb 9 09:03:11.280772 kernel: uhci_hcd: USB Universal Host Controller Interface driver Feb 9 09:03:11.280792 kernel: uhci_hcd 0000:00:01.2: UHCI Host Controller Feb 9 09:03:11.285815 kernel: uhci_hcd 0000:00:01.2: new USB bus registered, assigned bus number 1 Feb 9 09:03:11.286115 kernel: uhci_hcd 0000:00:01.2: detected 2 ports Feb 9 09:03:11.286303 kernel: uhci_hcd 0000:00:01.2: irq 11, io base 0x0000c180 Feb 9 09:03:11.291539 kernel: hub 1-0:1.0: USB hub found Feb 9 09:03:11.291887 kernel: hub 1-0:1.0: 2 ports detected Feb 9 09:03:11.292310 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device. Feb 9 09:03:11.301445 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device. Feb 9 09:03:11.306238 systemd[1]: Starting disk-uuid.service... Feb 9 09:03:11.316325 disk-uuid[489]: Primary Header is updated. Feb 9 09:03:11.316325 disk-uuid[489]: Secondary Entries is updated. Feb 9 09:03:11.316325 disk-uuid[489]: Secondary Header is updated. Feb 9 09:03:11.326386 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Feb 9 09:03:11.335482 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Feb 9 09:03:12.333962 disk-uuid[498]: The operation has completed successfully. Feb 9 09:03:12.334887 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Feb 9 09:03:12.393846 systemd[1]: disk-uuid.service: Deactivated successfully. Feb 9 09:03:12.394512 systemd[1]: Finished disk-uuid.service. Feb 9 09:03:12.395000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:12.395000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:12.413485 systemd[1]: Starting verity-setup.service... Feb 9 09:03:12.431411 kernel: device-mapper: verity: sha256 using implementation "sha256-avx2" Feb 9 09:03:12.485287 systemd[1]: Found device dev-mapper-usr.device. Feb 9 09:03:12.487145 systemd[1]: Mounting sysusr-usr.mount... Feb 9 09:03:12.488896 systemd[1]: Finished verity-setup.service. Feb 9 09:03:12.489000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=verity-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:12.581399 kernel: EXT4-fs (dm-0): mounted filesystem without journal. Opts: norecovery. Quota mode: none. Feb 9 09:03:12.582254 systemd[1]: Mounted sysusr-usr.mount. Feb 9 09:03:12.583118 systemd[1]: afterburn-network-kargs.service was skipped because no trigger condition checks were met. Feb 9 09:03:12.584215 systemd[1]: Starting ignition-setup.service... Feb 9 09:03:12.586224 systemd[1]: Starting parse-ip-for-networkd.service... Feb 9 09:03:12.609076 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Feb 9 09:03:12.609153 kernel: BTRFS info (device vda6): using free space tree Feb 9 09:03:12.609177 kernel: BTRFS info (device vda6): has skinny extents Feb 9 09:03:12.639617 systemd[1]: mnt-oem.mount: Deactivated successfully. Feb 9 09:03:12.648623 systemd[1]: Finished ignition-setup.service. Feb 9 09:03:12.649000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:12.650772 systemd[1]: Starting ignition-fetch-offline.service... Feb 9 09:03:12.754326 systemd[1]: Finished parse-ip-for-networkd.service. Feb 9 09:03:12.754000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:12.756000 audit: BPF prog-id=9 op=LOAD Feb 9 09:03:12.757556 systemd[1]: Starting systemd-networkd.service... Feb 9 09:03:12.824824 systemd-networkd[688]: lo: Link UP Feb 9 09:03:12.824838 systemd-networkd[688]: lo: Gained carrier Feb 9 09:03:12.825876 systemd-networkd[688]: Enumeration completed Feb 9 09:03:12.826472 systemd-networkd[688]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. Feb 9 09:03:12.830000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:12.827992 systemd-networkd[688]: eth0: Configuring with /usr/lib/systemd/network/yy-digitalocean.network. Feb 9 09:03:12.828643 systemd[1]: Started systemd-networkd.service. Feb 9 09:03:12.829334 systemd-networkd[688]: eth1: Link UP Feb 9 09:03:12.829341 systemd-networkd[688]: eth1: Gained carrier Feb 9 09:03:12.831291 systemd[1]: Reached target network.target. Feb 9 09:03:12.849030 ignition[626]: Ignition 2.14.0 Feb 9 09:03:12.860000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:12.832796 systemd-networkd[688]: eth0: Link UP Feb 9 09:03:12.849047 ignition[626]: Stage: fetch-offline Feb 9 09:03:12.832803 systemd-networkd[688]: eth0: Gained carrier Feb 9 09:03:12.849179 ignition[626]: reading system config file "/usr/lib/ignition/base.d/base.ign" Feb 9 09:03:12.839894 systemd[1]: Starting iscsiuio.service... Feb 9 09:03:12.849224 ignition[626]: parsing config with SHA512: 865c03baa79b8c74023d13a0b3666474fa06a165421a1e05731b76e0f557d42c5c89d4870a0b9c4182ad7d4d8209de20dca9c9da63d637e0410fbd60314cac6c Feb 9 09:03:12.846549 systemd-networkd[688]: eth1: DHCPv4 address 10.124.0.3/20 acquired from 169.254.169.253 Feb 9 09:03:12.856271 ignition[626]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Feb 9 09:03:12.888000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:12.854607 systemd-networkd[688]: eth0: DHCPv4 address 24.144.92.253/20, gateway 24.144.80.1 acquired from 169.254.169.253 Feb 9 09:03:12.856518 ignition[626]: parsed url from cmdline: "" Feb 9 09:03:12.859583 systemd[1]: Finished ignition-fetch-offline.service. Feb 9 09:03:12.856526 ignition[626]: no config URL provided Feb 9 09:03:12.862788 systemd[1]: Starting ignition-fetch.service... Feb 9 09:03:12.856539 ignition[626]: reading system config file "/usr/lib/ignition/user.ign" Feb 9 09:03:12.886177 systemd[1]: Started iscsiuio.service. Feb 9 09:03:12.899479 iscsid[698]: iscsid: can't open InitiatorName configuration file /etc/iscsi/initiatorname.iscsi Feb 9 09:03:12.899479 iscsid[698]: iscsid: Warning: InitiatorName file /etc/iscsi/initiatorname.iscsi does not exist or does not contain a properly formatted InitiatorName. If using software iscsi (iscsi_tcp or ib_iser) or partial offload (bnx2i or cxgbi iscsi), you may not be able to log into or discover targets. Please create a file /etc/iscsi/initiatorname.iscsi that contains a sting with the format: InitiatorName=iqn.yyyy-mm.[:identifier]. Feb 9 09:03:12.899479 iscsid[698]: Example: InitiatorName=iqn.2001-04.com.redhat:fc6. Feb 9 09:03:12.899479 iscsid[698]: If using hardware iscsi like qla4xxx this message can be ignored. Feb 9 09:03:12.899479 iscsid[698]: iscsid: can't open InitiatorAlias configuration file /etc/iscsi/initiatorname.iscsi Feb 9 09:03:12.899479 iscsid[698]: iscsid: can't open iscsid.safe_logout configuration file /etc/iscsi/iscsid.conf Feb 9 09:03:12.901000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:12.856560 ignition[626]: no config at "/usr/lib/ignition/user.ign" Feb 9 09:03:12.891654 systemd[1]: Starting iscsid.service... Feb 9 09:03:12.856571 ignition[626]: failed to fetch config: resource requires networking Feb 9 09:03:12.901275 systemd[1]: Started iscsid.service. Feb 9 09:03:12.856774 ignition[626]: Ignition finished successfully Feb 9 09:03:12.903907 systemd[1]: Starting dracut-initqueue.service... Feb 9 09:03:12.890698 ignition[692]: Ignition 2.14.0 Feb 9 09:03:12.899214 ignition[692]: Stage: fetch Feb 9 09:03:12.928000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:12.904763 ignition[692]: reading system config file "/usr/lib/ignition/base.d/base.ign" Feb 9 09:03:12.927999 systemd[1]: Finished dracut-initqueue.service. Feb 9 09:03:12.904829 ignition[692]: parsing config with SHA512: 865c03baa79b8c74023d13a0b3666474fa06a165421a1e05731b76e0f557d42c5c89d4870a0b9c4182ad7d4d8209de20dca9c9da63d637e0410fbd60314cac6c Feb 9 09:03:12.928897 systemd[1]: Reached target remote-fs-pre.target. Feb 9 09:03:12.908233 ignition[692]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Feb 9 09:03:12.929610 systemd[1]: Reached target remote-cryptsetup.target. Feb 9 09:03:12.911799 ignition[692]: parsed url from cmdline: "" Feb 9 09:03:12.930230 systemd[1]: Reached target remote-fs.target. Feb 9 09:03:12.911809 ignition[692]: no config URL provided Feb 9 09:03:12.932628 systemd[1]: Starting dracut-pre-mount.service... Feb 9 09:03:12.911824 ignition[692]: reading system config file "/usr/lib/ignition/user.ign" Feb 9 09:03:12.911847 ignition[692]: no config at "/usr/lib/ignition/user.ign" Feb 9 09:03:12.911905 ignition[692]: GET http://169.254.169.254/metadata/v1/user-data: attempt #1 Feb 9 09:03:12.947785 systemd[1]: Finished dracut-pre-mount.service. Feb 9 09:03:12.948000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:12.954683 ignition[692]: GET result: OK Feb 9 09:03:12.954845 ignition[692]: parsing config with SHA512: cba01bafaee59f8349c9e9e170be4a279efd58109be073b04b3a2963203bed711c6b9661ada6e21274716eac89508be762891d98e91a247dac52e7eae73f571b Feb 9 09:03:12.986129 unknown[692]: fetched base config from "system" Feb 9 09:03:12.987037 unknown[692]: fetched base config from "system" Feb 9 09:03:12.987802 unknown[692]: fetched user config from "digitalocean" Feb 9 09:03:12.989473 ignition[692]: fetch: fetch complete Feb 9 09:03:12.989490 ignition[692]: fetch: fetch passed Feb 9 09:03:12.989579 ignition[692]: Ignition finished successfully Feb 9 09:03:12.992444 systemd[1]: Finished ignition-fetch.service. Feb 9 09:03:12.992000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:12.994683 systemd[1]: Starting ignition-kargs.service... Feb 9 09:03:13.012486 ignition[713]: Ignition 2.14.0 Feb 9 09:03:13.012500 ignition[713]: Stage: kargs Feb 9 09:03:13.012671 ignition[713]: reading system config file "/usr/lib/ignition/base.d/base.ign" Feb 9 09:03:13.012697 ignition[713]: parsing config with SHA512: 865c03baa79b8c74023d13a0b3666474fa06a165421a1e05731b76e0f557d42c5c89d4870a0b9c4182ad7d4d8209de20dca9c9da63d637e0410fbd60314cac6c Feb 9 09:03:13.014812 ignition[713]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Feb 9 09:03:13.016503 ignition[713]: kargs: kargs passed Feb 9 09:03:13.016604 ignition[713]: Ignition finished successfully Feb 9 09:03:13.018078 systemd[1]: Finished ignition-kargs.service. Feb 9 09:03:13.018000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:13.020746 systemd[1]: Starting ignition-disks.service... Feb 9 09:03:13.040152 ignition[719]: Ignition 2.14.0 Feb 9 09:03:13.040170 ignition[719]: Stage: disks Feb 9 09:03:13.040379 ignition[719]: reading system config file "/usr/lib/ignition/base.d/base.ign" Feb 9 09:03:13.040412 ignition[719]: parsing config with SHA512: 865c03baa79b8c74023d13a0b3666474fa06a165421a1e05731b76e0f557d42c5c89d4870a0b9c4182ad7d4d8209de20dca9c9da63d637e0410fbd60314cac6c Feb 9 09:03:13.042996 ignition[719]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Feb 9 09:03:13.045236 ignition[719]: disks: disks passed Feb 9 09:03:13.046472 systemd[1]: Finished ignition-disks.service. Feb 9 09:03:13.047000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:13.045308 ignition[719]: Ignition finished successfully Feb 9 09:03:13.047990 systemd[1]: Reached target initrd-root-device.target. Feb 9 09:03:13.049297 systemd[1]: Reached target local-fs-pre.target. Feb 9 09:03:13.050488 systemd[1]: Reached target local-fs.target. Feb 9 09:03:13.051875 systemd[1]: Reached target sysinit.target. Feb 9 09:03:13.052984 systemd[1]: Reached target basic.target. Feb 9 09:03:13.055776 systemd[1]: Starting systemd-fsck-root.service... Feb 9 09:03:13.078156 systemd-fsck[727]: ROOT: clean, 602/553520 files, 56014/553472 blocks Feb 9 09:03:13.084000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:13.083487 systemd[1]: Finished systemd-fsck-root.service. Feb 9 09:03:13.086018 systemd[1]: Mounting sysroot.mount... Feb 9 09:03:13.102418 kernel: EXT4-fs (vda9): mounted filesystem with ordered data mode. Opts: (null). Quota mode: none. Feb 9 09:03:13.103085 systemd[1]: Mounted sysroot.mount. Feb 9 09:03:13.103922 systemd[1]: Reached target initrd-root-fs.target. Feb 9 09:03:13.106794 systemd[1]: Mounting sysroot-usr.mount... Feb 9 09:03:13.108957 systemd[1]: Starting flatcar-digitalocean-network.service... Feb 9 09:03:13.112760 systemd[1]: Starting flatcar-metadata-hostname.service... Feb 9 09:03:13.113556 systemd[1]: ignition-remount-sysroot.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Feb 9 09:03:13.113621 systemd[1]: Reached target ignition-diskful.target. Feb 9 09:03:13.123058 systemd[1]: Mounted sysroot-usr.mount. Feb 9 09:03:13.126587 systemd[1]: Starting initrd-setup-root.service... Feb 9 09:03:13.138102 initrd-setup-root[739]: cut: /sysroot/etc/passwd: No such file or directory Feb 9 09:03:13.146162 initrd-setup-root[747]: cut: /sysroot/etc/group: No such file or directory Feb 9 09:03:13.154197 initrd-setup-root[755]: cut: /sysroot/etc/shadow: No such file or directory Feb 9 09:03:13.164705 initrd-setup-root[763]: cut: /sysroot/etc/gshadow: No such file or directory Feb 9 09:03:13.254781 systemd[1]: Finished initrd-setup-root.service. Feb 9 09:03:13.255000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:13.257336 systemd[1]: Starting ignition-mount.service... Feb 9 09:03:13.259959 systemd[1]: Starting sysroot-boot.service... Feb 9 09:03:13.279672 bash[785]: umount: /sysroot/usr/share/oem: not mounted. Feb 9 09:03:13.281046 coreos-metadata[733]: Feb 09 09:03:13.280 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Feb 9 09:03:13.294926 ignition[786]: INFO : Ignition 2.14.0 Feb 9 09:03:13.296000 ignition[786]: INFO : Stage: mount Feb 9 09:03:13.296965 ignition[786]: INFO : reading system config file "/usr/lib/ignition/base.d/base.ign" Feb 9 09:03:13.297975 ignition[786]: DEBUG : parsing config with SHA512: 865c03baa79b8c74023d13a0b3666474fa06a165421a1e05731b76e0f557d42c5c89d4870a0b9c4182ad7d4d8209de20dca9c9da63d637e0410fbd60314cac6c Feb 9 09:03:13.302994 coreos-metadata[733]: Feb 09 09:03:13.302 INFO Fetch successful Feb 9 09:03:13.304198 ignition[786]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Feb 9 09:03:13.307522 ignition[786]: INFO : mount: mount passed Feb 9 09:03:13.308488 ignition[786]: INFO : Ignition finished successfully Feb 9 09:03:13.310000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=flatcar-digitalocean-network comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:13.310000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=flatcar-digitalocean-network comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:13.311000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:13.310162 systemd[1]: flatcar-digitalocean-network.service: Deactivated successfully. Feb 9 09:03:13.310267 systemd[1]: Finished flatcar-digitalocean-network.service. Feb 9 09:03:13.311029 systemd[1]: Finished ignition-mount.service. Feb 9 09:03:13.320629 systemd[1]: Finished sysroot-boot.service. Feb 9 09:03:13.320000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:13.330634 coreos-metadata[734]: Feb 09 09:03:13.330 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Feb 9 09:03:13.343116 coreos-metadata[734]: Feb 09 09:03:13.342 INFO Fetch successful Feb 9 09:03:13.349716 coreos-metadata[734]: Feb 09 09:03:13.349 INFO wrote hostname ci-3510.3.2-6-d69396b42b to /sysroot/etc/hostname Feb 9 09:03:13.352193 systemd[1]: Finished flatcar-metadata-hostname.service. Feb 9 09:03:13.352000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=flatcar-metadata-hostname comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:13.509841 systemd[1]: Mounting sysroot-usr-share-oem.mount... Feb 9 09:03:13.519485 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/vda6 scanned by mount (794) Feb 9 09:03:13.538002 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Feb 9 09:03:13.538094 kernel: BTRFS info (device vda6): using free space tree Feb 9 09:03:13.538118 kernel: BTRFS info (device vda6): has skinny extents Feb 9 09:03:13.545537 systemd[1]: Mounted sysroot-usr-share-oem.mount. Feb 9 09:03:13.547753 systemd[1]: Starting ignition-files.service... Feb 9 09:03:13.572143 ignition[814]: INFO : Ignition 2.14.0 Feb 9 09:03:13.572143 ignition[814]: INFO : Stage: files Feb 9 09:03:13.573844 ignition[814]: INFO : reading system config file "/usr/lib/ignition/base.d/base.ign" Feb 9 09:03:13.573844 ignition[814]: DEBUG : parsing config with SHA512: 865c03baa79b8c74023d13a0b3666474fa06a165421a1e05731b76e0f557d42c5c89d4870a0b9c4182ad7d4d8209de20dca9c9da63d637e0410fbd60314cac6c Feb 9 09:03:13.576231 ignition[814]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Feb 9 09:03:13.581543 ignition[814]: DEBUG : files: compiled without relabeling support, skipping Feb 9 09:03:13.586211 ignition[814]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Feb 9 09:03:13.586211 ignition[814]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Feb 9 09:03:13.590661 ignition[814]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Feb 9 09:03:13.591679 ignition[814]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Feb 9 09:03:13.592541 ignition[814]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Feb 9 09:03:13.592458 unknown[814]: wrote ssh authorized keys file for user: core Feb 9 09:03:13.594347 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/cni-plugins-linux-amd64-v1.3.0.tgz" Feb 9 09:03:13.594347 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://github.com/containernetworking/plugins/releases/download/v1.3.0/cni-plugins-linux-amd64-v1.3.0.tgz: attempt #1 Feb 9 09:03:14.065092 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Feb 9 09:03:14.267762 systemd-networkd[688]: eth0: Gained IPv6LL Feb 9 09:03:14.306681 ignition[814]: DEBUG : files: createFilesystemsFiles: createFiles: op(3): file matches expected sum of: 5d0324ca8a3c90c680b6e1fddb245a2255582fa15949ba1f3c6bb7323df9d3af754dae98d6e40ac9ccafb2999c932df2c4288d418949a4915d928eb23c090540 Feb 9 09:03:14.308800 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/cni-plugins-linux-amd64-v1.3.0.tgz" Feb 9 09:03:14.308800 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/crictl-v1.27.0-linux-amd64.tar.gz" Feb 9 09:03:14.308800 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/kubernetes-sigs/cri-tools/releases/download/v1.27.0/crictl-v1.27.0-linux-amd64.tar.gz: attempt #1 Feb 9 09:03:14.651580 systemd-networkd[688]: eth1: Gained IPv6LL Feb 9 09:03:14.745501 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Feb 9 09:03:14.861249 ignition[814]: DEBUG : files: createFilesystemsFiles: createFiles: op(4): file matches expected sum of: aa622325bf05520939f9e020d7a28ab48ac23e2fae6f47d5a4e52174c88c1ebc31b464853e4fd65bd8f5331f330a6ca96fd370d247d3eeaed042da4ee2d1219a Feb 9 09:03:14.863090 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/crictl-v1.27.0-linux-amd64.tar.gz" Feb 9 09:03:14.863090 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/opt/bin/kubeadm" Feb 9 09:03:14.863090 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(5): GET https://dl.k8s.io/release/v1.28.1/bin/linux/amd64/kubeadm: attempt #1 Feb 9 09:03:14.925586 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(5): GET result: OK Feb 9 09:03:15.229308 ignition[814]: DEBUG : files: createFilesystemsFiles: createFiles: op(5): file matches expected sum of: f4daad200c8378dfdc6cb69af28eaca4215f2b4a2dbdf75f29f9210171cb5683bc873fc000319022e6b3ad61175475d77190734713ba9136644394e8a8faafa1 Feb 9 09:03:15.231241 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/opt/bin/kubeadm" Feb 9 09:03:15.231241 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/opt/bin/kubelet" Feb 9 09:03:15.231241 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(6): GET https://dl.k8s.io/release/v1.28.1/bin/linux/amd64/kubelet: attempt #1 Feb 9 09:03:15.276162 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(6): GET result: OK Feb 9 09:03:15.817754 ignition[814]: DEBUG : files: createFilesystemsFiles: createFiles: op(6): file matches expected sum of: ce6ba764274162d38ac1c44e1fb1f0f835346f3afc5b508bb755b1b7d7170910f5812b0a1941b32e29d950e905bbd08ae761c87befad921db4d44969c8562e75 Feb 9 09:03:15.820526 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/opt/bin/kubelet" Feb 9 09:03:15.820526 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/install.sh" Feb 9 09:03:15.820526 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/install.sh" Feb 9 09:03:15.820526 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/docker/daemon.json" Feb 9 09:03:15.820526 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/docker/daemon.json" Feb 9 09:03:15.820526 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Feb 9 09:03:15.820526 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Feb 9 09:03:15.820526 ignition[814]: INFO : files: op(a): [started] processing unit "coreos-metadata-sshkeys@.service" Feb 9 09:03:15.820526 ignition[814]: INFO : files: op(a): [finished] processing unit "coreos-metadata-sshkeys@.service" Feb 9 09:03:15.820526 ignition[814]: INFO : files: op(b): [started] processing unit "prepare-cni-plugins.service" Feb 9 09:03:15.820526 ignition[814]: INFO : files: op(b): op(c): [started] writing unit "prepare-cni-plugins.service" at "/sysroot/etc/systemd/system/prepare-cni-plugins.service" Feb 9 09:03:15.820526 ignition[814]: INFO : files: op(b): op(c): [finished] writing unit "prepare-cni-plugins.service" at "/sysroot/etc/systemd/system/prepare-cni-plugins.service" Feb 9 09:03:15.820526 ignition[814]: INFO : files: op(b): [finished] processing unit "prepare-cni-plugins.service" Feb 9 09:03:15.820526 ignition[814]: INFO : files: op(d): [started] processing unit "prepare-critools.service" Feb 9 09:03:15.820526 ignition[814]: INFO : files: op(d): op(e): [started] writing unit "prepare-critools.service" at "/sysroot/etc/systemd/system/prepare-critools.service" Feb 9 09:03:15.820526 ignition[814]: INFO : files: op(d): op(e): [finished] writing unit "prepare-critools.service" at "/sysroot/etc/systemd/system/prepare-critools.service" Feb 9 09:03:15.820526 ignition[814]: INFO : files: op(d): [finished] processing unit "prepare-critools.service" Feb 9 09:03:15.820526 ignition[814]: INFO : files: op(f): [started] setting preset to enabled for "coreos-metadata-sshkeys@.service " Feb 9 09:03:15.874051 kernel: kauditd_printk_skb: 29 callbacks suppressed Feb 9 09:03:15.874090 kernel: audit: type=1130 audit(1707469395.831:40): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:15.874113 kernel: audit: type=1130 audit(1707469395.846:41): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:15.874141 kernel: audit: type=1131 audit(1707469395.847:42): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:15.874159 kernel: audit: type=1130 audit(1707469395.859:43): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:15.831000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:15.846000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:15.847000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:15.859000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:15.874414 ignition[814]: INFO : files: op(f): [finished] setting preset to enabled for "coreos-metadata-sshkeys@.service " Feb 9 09:03:15.874414 ignition[814]: INFO : files: op(10): [started] setting preset to enabled for "prepare-cni-plugins.service" Feb 9 09:03:15.874414 ignition[814]: INFO : files: op(10): [finished] setting preset to enabled for "prepare-cni-plugins.service" Feb 9 09:03:15.874414 ignition[814]: INFO : files: op(11): [started] setting preset to enabled for "prepare-critools.service" Feb 9 09:03:15.874414 ignition[814]: INFO : files: op(11): [finished] setting preset to enabled for "prepare-critools.service" Feb 9 09:03:15.874414 ignition[814]: INFO : files: createResultFile: createFiles: op(12): [started] writing file "/sysroot/etc/.ignition-result.json" Feb 9 09:03:15.874414 ignition[814]: INFO : files: createResultFile: createFiles: op(12): [finished] writing file "/sysroot/etc/.ignition-result.json" Feb 9 09:03:15.874414 ignition[814]: INFO : files: files passed Feb 9 09:03:15.874414 ignition[814]: INFO : Ignition finished successfully Feb 9 09:03:15.830381 systemd[1]: Finished ignition-files.service. Feb 9 09:03:15.833484 systemd[1]: Starting initrd-setup-root-after-ignition.service... Feb 9 09:03:15.886859 initrd-setup-root-after-ignition[839]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Feb 9 09:03:15.840508 systemd[1]: torcx-profile-populate.service was skipped because of an unmet condition check (ConditionPathExists=/sysroot/etc/torcx/next-profile). Feb 9 09:03:15.841595 systemd[1]: Starting ignition-quench.service... Feb 9 09:03:15.846700 systemd[1]: ignition-quench.service: Deactivated successfully. Feb 9 09:03:15.846798 systemd[1]: Finished ignition-quench.service. Feb 9 09:03:15.859344 systemd[1]: Finished initrd-setup-root-after-ignition.service. Feb 9 09:03:15.860400 systemd[1]: Reached target ignition-complete.target. Feb 9 09:03:15.868332 systemd[1]: Starting initrd-parse-etc.service... Feb 9 09:03:15.918417 kernel: audit: type=1130 audit(1707469395.894:44): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:15.918470 kernel: audit: type=1131 audit(1707469395.894:45): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:15.894000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:15.894000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:15.893789 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Feb 9 09:03:15.893894 systemd[1]: Finished initrd-parse-etc.service. Feb 9 09:03:15.895038 systemd[1]: Reached target initrd-fs.target. Feb 9 09:03:15.919064 systemd[1]: Reached target initrd.target. Feb 9 09:03:15.920323 systemd[1]: dracut-mount.service was skipped because no trigger condition checks were met. Feb 9 09:03:15.921921 systemd[1]: Starting dracut-pre-pivot.service... Feb 9 09:03:15.938825 systemd[1]: Finished dracut-pre-pivot.service. Feb 9 09:03:15.939000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:15.941382 systemd[1]: Starting initrd-cleanup.service... Feb 9 09:03:15.946899 kernel: audit: type=1130 audit(1707469395.939:46): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:15.957923 systemd[1]: Stopped target nss-lookup.target. Feb 9 09:03:15.959574 systemd[1]: Stopped target remote-cryptsetup.target. Feb 9 09:03:15.961162 systemd[1]: Stopped target timers.target. Feb 9 09:03:15.962904 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Feb 9 09:03:15.964062 systemd[1]: Stopped dracut-pre-pivot.service. Feb 9 09:03:15.965000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:15.971121 systemd[1]: Stopped target initrd.target. Feb 9 09:03:15.972742 kernel: audit: type=1131 audit(1707469395.965:47): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:15.972159 systemd[1]: Stopped target basic.target. Feb 9 09:03:15.973318 systemd[1]: Stopped target ignition-complete.target. Feb 9 09:03:15.974441 systemd[1]: Stopped target ignition-diskful.target. Feb 9 09:03:15.975725 systemd[1]: Stopped target initrd-root-device.target. Feb 9 09:03:15.976987 systemd[1]: Stopped target remote-fs.target. Feb 9 09:03:15.978179 systemd[1]: Stopped target remote-fs-pre.target. Feb 9 09:03:15.979242 systemd[1]: Stopped target sysinit.target. Feb 9 09:03:15.980536 systemd[1]: Stopped target local-fs.target. Feb 9 09:03:15.981758 systemd[1]: Stopped target local-fs-pre.target. Feb 9 09:03:15.982916 systemd[1]: Stopped target swap.target. Feb 9 09:03:15.984186 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Feb 9 09:03:15.991140 kernel: audit: type=1131 audit(1707469395.985:48): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:15.985000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:15.984457 systemd[1]: Stopped dracut-pre-mount.service. Feb 9 09:03:15.985744 systemd[1]: Stopped target cryptsetup.target. Feb 9 09:03:15.998472 kernel: audit: type=1131 audit(1707469395.992:49): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:15.992000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:15.991932 systemd[1]: dracut-initqueue.service: Deactivated successfully. Feb 9 09:03:15.992250 systemd[1]: Stopped dracut-initqueue.service. Feb 9 09:03:15.993234 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Feb 9 09:03:15.993504 systemd[1]: Stopped initrd-setup-root-after-ignition.service. Feb 9 09:03:16.002850 systemd[1]: ignition-files.service: Deactivated successfully. Feb 9 09:03:16.002000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:16.003098 systemd[1]: Stopped ignition-files.service. Feb 9 09:03:16.004000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:16.004774 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Feb 9 09:03:16.004985 systemd[1]: Stopped flatcar-metadata-hostname.service. Feb 9 09:03:16.005000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=flatcar-metadata-hostname comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:16.007723 systemd[1]: Stopping ignition-mount.service... Feb 9 09:03:16.008740 systemd[1]: Stopping iscsid.service... Feb 9 09:03:16.019114 iscsid[698]: iscsid shutting down. Feb 9 09:03:16.019000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:16.017339 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Feb 9 09:03:16.017619 systemd[1]: Stopped kmod-static-nodes.service. Feb 9 09:03:16.029000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:16.021223 systemd[1]: Stopping sysroot-boot.service... Feb 9 09:03:16.027810 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Feb 9 09:03:16.028080 systemd[1]: Stopped systemd-udev-trigger.service. Feb 9 09:03:16.031000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:16.030062 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Feb 9 09:03:16.030272 systemd[1]: Stopped dracut-pre-trigger.service. Feb 9 09:03:16.034102 systemd[1]: iscsid.service: Deactivated successfully. Feb 9 09:03:16.036206 systemd[1]: Stopped iscsid.service. Feb 9 09:03:16.037000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:16.049417 ignition[852]: INFO : Ignition 2.14.0 Feb 9 09:03:16.049417 ignition[852]: INFO : Stage: umount Feb 9 09:03:16.049417 ignition[852]: INFO : reading system config file "/usr/lib/ignition/base.d/base.ign" Feb 9 09:03:16.049417 ignition[852]: DEBUG : parsing config with SHA512: 865c03baa79b8c74023d13a0b3666474fa06a165421a1e05731b76e0f557d42c5c89d4870a0b9c4182ad7d4d8209de20dca9c9da63d637e0410fbd60314cac6c Feb 9 09:03:16.049417 ignition[852]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Feb 9 09:03:16.051000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:16.057000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:16.057000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:16.047763 systemd[1]: Stopping iscsiuio.service... Feb 9 09:03:16.048814 systemd[1]: iscsiuio.service: Deactivated successfully. Feb 9 09:03:16.049436 systemd[1]: Stopped iscsiuio.service. Feb 9 09:03:16.052051 systemd[1]: initrd-cleanup.service: Deactivated successfully. Feb 9 09:03:16.067211 ignition[852]: INFO : umount: umount passed Feb 9 09:03:16.067211 ignition[852]: INFO : Ignition finished successfully Feb 9 09:03:16.067000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:16.068000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:16.052172 systemd[1]: Finished initrd-cleanup.service. Feb 9 09:03:16.069000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:16.058952 systemd[1]: sysroot-boot.mount: Deactivated successfully. Feb 9 09:03:16.070000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:16.066857 systemd[1]: sysroot-boot.service: Deactivated successfully. Feb 9 09:03:16.072000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:16.066972 systemd[1]: Stopped sysroot-boot.service. Feb 9 09:03:16.067994 systemd[1]: ignition-mount.service: Deactivated successfully. Feb 9 09:03:16.074000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:16.068113 systemd[1]: Stopped ignition-mount.service. Feb 9 09:03:16.069018 systemd[1]: ignition-disks.service: Deactivated successfully. Feb 9 09:03:16.069077 systemd[1]: Stopped ignition-disks.service. Feb 9 09:03:16.070098 systemd[1]: ignition-kargs.service: Deactivated successfully. Feb 9 09:03:16.070164 systemd[1]: Stopped ignition-kargs.service. Feb 9 09:03:16.071262 systemd[1]: ignition-fetch.service: Deactivated successfully. Feb 9 09:03:16.071394 systemd[1]: Stopped ignition-fetch.service. Feb 9 09:03:16.072673 systemd[1]: Stopped target network.target. Feb 9 09:03:16.073757 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Feb 9 09:03:16.084000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:16.073807 systemd[1]: Stopped ignition-fetch-offline.service. Feb 9 09:03:16.086000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:16.075008 systemd[1]: Stopped target paths.target. Feb 9 09:03:16.076286 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Feb 9 09:03:16.078475 systemd[1]: Stopped systemd-ask-password-console.path. Feb 9 09:03:16.079196 systemd[1]: Stopped target slices.target. Feb 9 09:03:16.080495 systemd[1]: Stopped target sockets.target. Feb 9 09:03:16.081707 systemd[1]: iscsid.socket: Deactivated successfully. Feb 9 09:03:16.081782 systemd[1]: Closed iscsid.socket. Feb 9 09:03:16.082891 systemd[1]: iscsiuio.socket: Deactivated successfully. Feb 9 09:03:16.082940 systemd[1]: Closed iscsiuio.socket. Feb 9 09:03:16.095000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:16.084055 systemd[1]: ignition-setup.service: Deactivated successfully. Feb 9 09:03:16.084118 systemd[1]: Stopped ignition-setup.service. Feb 9 09:03:16.097000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:16.098000 audit: BPF prog-id=6 op=UNLOAD Feb 9 09:03:16.085297 systemd[1]: initrd-setup-root.service: Deactivated successfully. Feb 9 09:03:16.085343 systemd[1]: Stopped initrd-setup-root.service. Feb 9 09:03:16.087411 systemd[1]: Stopping systemd-networkd.service... Feb 9 09:03:16.088401 systemd[1]: Stopping systemd-resolved.service... Feb 9 09:03:16.105000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:16.090444 systemd-networkd[688]: eth1: DHCPv6 lease lost Feb 9 09:03:16.106000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:16.093543 systemd-networkd[688]: eth0: DHCPv6 lease lost Feb 9 09:03:16.107000 audit: BPF prog-id=9 op=UNLOAD Feb 9 09:03:16.107000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:16.094915 systemd[1]: systemd-resolved.service: Deactivated successfully. Feb 9 09:03:16.095065 systemd[1]: Stopped systemd-resolved.service. Feb 9 09:03:16.096972 systemd[1]: systemd-networkd.service: Deactivated successfully. Feb 9 09:03:16.097082 systemd[1]: Stopped systemd-networkd.service. Feb 9 09:03:16.098685 systemd[1]: systemd-networkd.socket: Deactivated successfully. Feb 9 09:03:16.098723 systemd[1]: Closed systemd-networkd.socket. Feb 9 09:03:16.100719 systemd[1]: Stopping network-cleanup.service... Feb 9 09:03:16.136000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:16.101840 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Feb 9 09:03:16.137000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=network-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:16.101924 systemd[1]: Stopped parse-ip-for-networkd.service. Feb 9 09:03:16.105490 systemd[1]: systemd-sysctl.service: Deactivated successfully. Feb 9 09:03:16.105560 systemd[1]: Stopped systemd-sysctl.service. Feb 9 09:03:16.141000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:16.107063 systemd[1]: systemd-modules-load.service: Deactivated successfully. Feb 9 09:03:16.143000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:16.107122 systemd[1]: Stopped systemd-modules-load.service. Feb 9 09:03:16.144000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:16.108140 systemd[1]: Stopping systemd-udevd.service... Feb 9 09:03:16.115517 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Feb 9 09:03:16.134697 systemd[1]: systemd-udevd.service: Deactivated successfully. Feb 9 09:03:16.135231 systemd[1]: Stopped systemd-udevd.service. Feb 9 09:03:16.137248 systemd[1]: network-cleanup.service: Deactivated successfully. Feb 9 09:03:16.156000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:16.137414 systemd[1]: Stopped network-cleanup.service. Feb 9 09:03:16.158000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:16.158000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:16.138331 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Feb 9 09:03:16.138396 systemd[1]: Closed systemd-udevd-control.socket. Feb 9 09:03:16.139725 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Feb 9 09:03:16.139774 systemd[1]: Closed systemd-udevd-kernel.socket. Feb 9 09:03:16.140980 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Feb 9 09:03:16.141047 systemd[1]: Stopped dracut-pre-udev.service. Feb 9 09:03:16.142469 systemd[1]: dracut-cmdline.service: Deactivated successfully. Feb 9 09:03:16.142535 systemd[1]: Stopped dracut-cmdline.service. Feb 9 09:03:16.143770 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Feb 9 09:03:16.143834 systemd[1]: Stopped dracut-cmdline-ask.service. Feb 9 09:03:16.146330 systemd[1]: Starting initrd-udevadm-cleanup-db.service... Feb 9 09:03:16.147090 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Feb 9 09:03:16.147170 systemd[1]: Stopped systemd-vconsole-setup.service. Feb 9 09:03:16.157796 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Feb 9 09:03:16.157923 systemd[1]: Finished initrd-udevadm-cleanup-db.service. Feb 9 09:03:16.158736 systemd[1]: Reached target initrd-switch-root.target. Feb 9 09:03:16.161132 systemd[1]: Starting initrd-switch-root.service... Feb 9 09:03:16.176856 systemd[1]: Switching root. Feb 9 09:03:16.200230 systemd-journald[184]: Journal stopped Feb 9 09:03:20.009907 systemd-journald[184]: Received SIGTERM from PID 1 (systemd). Feb 9 09:03:20.009991 kernel: SELinux: Class mctp_socket not defined in policy. Feb 9 09:03:20.010016 kernel: SELinux: Class anon_inode not defined in policy. Feb 9 09:03:20.010038 kernel: SELinux: the above unknown classes and permissions will be allowed Feb 9 09:03:20.010085 kernel: SELinux: policy capability network_peer_controls=1 Feb 9 09:03:20.010107 kernel: SELinux: policy capability open_perms=1 Feb 9 09:03:20.010134 kernel: SELinux: policy capability extended_socket_class=1 Feb 9 09:03:20.010156 kernel: SELinux: policy capability always_check_network=0 Feb 9 09:03:20.010178 kernel: SELinux: policy capability cgroup_seclabel=1 Feb 9 09:03:20.010201 kernel: SELinux: policy capability nnp_nosuid_transition=1 Feb 9 09:03:20.010222 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Feb 9 09:03:20.010243 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Feb 9 09:03:20.010272 systemd[1]: Successfully loaded SELinux policy in 55.248ms. Feb 9 09:03:20.010312 systemd[1]: Relabelled /dev, /dev/shm, /run, /sys/fs/cgroup in 13.910ms. Feb 9 09:03:20.010339 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) Feb 9 09:03:20.010410 systemd[1]: Detected virtualization kvm. Feb 9 09:03:20.010433 systemd[1]: Detected architecture x86-64. Feb 9 09:03:20.010457 systemd[1]: Detected first boot. Feb 9 09:03:20.010481 systemd[1]: Hostname set to . Feb 9 09:03:20.010505 systemd[1]: Initializing machine ID from VM UUID. Feb 9 09:03:20.010542 kernel: SELinux: Context system_u:object_r:container_file_t:s0:c1022,c1023 is not valid (left unmapped). Feb 9 09:03:20.010565 systemd[1]: Populated /etc with preset unit settings. Feb 9 09:03:20.010591 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Feb 9 09:03:20.010616 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Feb 9 09:03:20.010643 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 9 09:03:20.010666 systemd[1]: initrd-switch-root.service: Deactivated successfully. Feb 9 09:03:20.010691 systemd[1]: Stopped initrd-switch-root.service. Feb 9 09:03:20.010722 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Feb 9 09:03:20.010746 systemd[1]: Created slice system-addon\x2dconfig.slice. Feb 9 09:03:20.010770 systemd[1]: Created slice system-addon\x2drun.slice. Feb 9 09:03:20.010792 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice. Feb 9 09:03:20.010817 systemd[1]: Created slice system-getty.slice. Feb 9 09:03:20.010841 systemd[1]: Created slice system-modprobe.slice. Feb 9 09:03:20.010865 systemd[1]: Created slice system-serial\x2dgetty.slice. Feb 9 09:03:20.010889 systemd[1]: Created slice system-system\x2dcloudinit.slice. Feb 9 09:03:20.010913 systemd[1]: Created slice system-systemd\x2dfsck.slice. Feb 9 09:03:20.010947 systemd[1]: Created slice user.slice. Feb 9 09:03:20.010971 systemd[1]: Started systemd-ask-password-console.path. Feb 9 09:03:20.010995 systemd[1]: Started systemd-ask-password-wall.path. Feb 9 09:03:20.011018 systemd[1]: Set up automount boot.automount. Feb 9 09:03:20.011042 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount. Feb 9 09:03:20.011066 systemd[1]: Stopped target initrd-switch-root.target. Feb 9 09:03:20.011098 systemd[1]: Stopped target initrd-fs.target. Feb 9 09:03:20.011123 systemd[1]: Stopped target initrd-root-fs.target. Feb 9 09:03:20.011147 systemd[1]: Reached target integritysetup.target. Feb 9 09:03:20.011172 systemd[1]: Reached target remote-cryptsetup.target. Feb 9 09:03:20.011196 systemd[1]: Reached target remote-fs.target. Feb 9 09:03:20.011220 systemd[1]: Reached target slices.target. Feb 9 09:03:20.011259 systemd[1]: Reached target swap.target. Feb 9 09:03:20.011286 systemd[1]: Reached target torcx.target. Feb 9 09:03:20.011311 systemd[1]: Reached target veritysetup.target. Feb 9 09:03:20.011335 systemd[1]: Listening on systemd-coredump.socket. Feb 9 09:03:20.011408 systemd[1]: Listening on systemd-initctl.socket. Feb 9 09:03:20.011433 systemd[1]: Listening on systemd-networkd.socket. Feb 9 09:03:20.011457 systemd[1]: Listening on systemd-udevd-control.socket. Feb 9 09:03:20.011481 systemd[1]: Listening on systemd-udevd-kernel.socket. Feb 9 09:03:20.011504 systemd[1]: Listening on systemd-userdbd.socket. Feb 9 09:03:20.011527 systemd[1]: Mounting dev-hugepages.mount... Feb 9 09:03:20.011552 systemd[1]: Mounting dev-mqueue.mount... Feb 9 09:03:20.011576 systemd[1]: Mounting media.mount... Feb 9 09:03:20.011600 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 9 09:03:20.011639 systemd[1]: Mounting sys-kernel-debug.mount... Feb 9 09:03:20.011665 systemd[1]: Mounting sys-kernel-tracing.mount... Feb 9 09:03:20.011688 systemd[1]: Mounting tmp.mount... Feb 9 09:03:20.011717 systemd[1]: Starting flatcar-tmpfiles.service... Feb 9 09:03:20.011740 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Feb 9 09:03:20.011767 systemd[1]: Starting kmod-static-nodes.service... Feb 9 09:03:20.011786 systemd[1]: Starting modprobe@configfs.service... Feb 9 09:03:20.011804 systemd[1]: Starting modprobe@dm_mod.service... Feb 9 09:03:20.011827 systemd[1]: Starting modprobe@drm.service... Feb 9 09:03:20.011863 systemd[1]: Starting modprobe@efi_pstore.service... Feb 9 09:03:20.011877 systemd[1]: Starting modprobe@fuse.service... Feb 9 09:03:20.011890 systemd[1]: Starting modprobe@loop.service... Feb 9 09:03:20.011904 systemd[1]: setup-nsswitch.service was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Feb 9 09:03:20.011918 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Feb 9 09:03:20.011931 systemd[1]: Stopped systemd-fsck-root.service. Feb 9 09:03:20.011948 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Feb 9 09:03:20.011961 systemd[1]: Stopped systemd-fsck-usr.service. Feb 9 09:03:20.011975 systemd[1]: Stopped systemd-journald.service. Feb 9 09:03:20.011996 systemd[1]: Starting systemd-journald.service... Feb 9 09:03:20.012009 systemd[1]: Starting systemd-modules-load.service... Feb 9 09:03:20.012023 kernel: loop: module loaded Feb 9 09:03:20.012037 systemd[1]: Starting systemd-network-generator.service... Feb 9 09:03:20.012051 systemd[1]: Starting systemd-remount-fs.service... Feb 9 09:03:20.012065 systemd[1]: Starting systemd-udev-trigger.service... Feb 9 09:03:20.012079 systemd[1]: verity-setup.service: Deactivated successfully. Feb 9 09:03:20.012092 systemd[1]: Stopped verity-setup.service. Feb 9 09:03:20.012106 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 9 09:03:20.012126 systemd[1]: Mounted dev-hugepages.mount. Feb 9 09:03:20.012140 kernel: fuse: init (API version 7.34) Feb 9 09:03:20.012153 systemd[1]: Mounted dev-mqueue.mount. Feb 9 09:03:20.012166 systemd[1]: Mounted media.mount. Feb 9 09:03:20.012180 systemd[1]: Mounted sys-kernel-debug.mount. Feb 9 09:03:20.012193 systemd[1]: Mounted sys-kernel-tracing.mount. Feb 9 09:03:20.012207 systemd[1]: Mounted tmp.mount. Feb 9 09:03:20.012223 systemd[1]: Finished kmod-static-nodes.service. Feb 9 09:03:20.012236 systemd[1]: modprobe@configfs.service: Deactivated successfully. Feb 9 09:03:20.012255 systemd[1]: Finished modprobe@configfs.service. Feb 9 09:03:20.012269 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Feb 9 09:03:20.012282 systemd[1]: Finished modprobe@dm_mod.service. Feb 9 09:03:20.012295 systemd[1]: modprobe@drm.service: Deactivated successfully. Feb 9 09:03:20.012308 systemd[1]: Finished modprobe@drm.service. Feb 9 09:03:20.012329 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Feb 9 09:03:20.012343 systemd[1]: Finished modprobe@efi_pstore.service. Feb 9 09:03:20.012369 systemd[1]: modprobe@fuse.service: Deactivated successfully. Feb 9 09:03:20.012397 systemd-journald[963]: Journal started Feb 9 09:03:20.012457 systemd-journald[963]: Runtime Journal (/run/log/journal/ed9135f0930245b7812d42f261728148) is 4.9M, max 39.5M, 34.5M free. Feb 9 09:03:16.370000 audit: MAC_POLICY_LOAD auid=4294967295 ses=4294967295 lsm=selinux res=1 Feb 9 09:03:16.442000 audit[1]: AVC avc: denied { bpf } for pid=1 comm="systemd" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=1 Feb 9 09:03:16.442000 audit[1]: AVC avc: denied { perfmon } for pid=1 comm="systemd" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=1 Feb 9 09:03:16.442000 audit: BPF prog-id=10 op=LOAD Feb 9 09:03:16.442000 audit: BPF prog-id=10 op=UNLOAD Feb 9 09:03:16.442000 audit: BPF prog-id=11 op=LOAD Feb 9 09:03:16.442000 audit: BPF prog-id=11 op=UNLOAD Feb 9 09:03:16.548000 audit[885]: AVC avc: denied { associate } for pid=885 comm="torcx-generator" name="docker" dev="tmpfs" ino=2 scontext=system_u:object_r:unlabeled_t:s0 tcontext=system_u:object_r:tmpfs_t:s0 tclass=filesystem permissive=1 srawcon="system_u:object_r:container_file_t:s0:c1022,c1023" Feb 9 09:03:16.548000 audit[885]: SYSCALL arch=c000003e syscall=188 success=yes exit=0 a0=c00014d8a2 a1=c0000cedf8 a2=c0000d70c0 a3=32 items=0 ppid=868 pid=885 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="torcx-generator" exe="/usr/lib/systemd/system-generators/torcx-generator" subj=system_u:system_r:kernel_t:s0 key=(null) Feb 9 09:03:16.548000 audit: PROCTITLE proctitle=2F7573722F6C69622F73797374656D642F73797374656D2D67656E657261746F72732F746F7263782D67656E657261746F72002F72756E2F73797374656D642F67656E657261746F72002F72756E2F73797374656D642F67656E657261746F722E6561726C79002F72756E2F73797374656D642F67656E657261746F722E6C61 Feb 9 09:03:16.551000 audit[885]: AVC avc: denied { associate } for pid=885 comm="torcx-generator" name="bin" scontext=system_u:object_r:unlabeled_t:s0 tcontext=system_u:object_r:tmpfs_t:s0 tclass=filesystem permissive=1 Feb 9 09:03:16.551000 audit[885]: SYSCALL arch=c000003e syscall=258 success=yes exit=0 a0=ffffffffffffff9c a1=c00014d979 a2=1ed a3=0 items=2 ppid=868 pid=885 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="torcx-generator" exe="/usr/lib/systemd/system-generators/torcx-generator" subj=system_u:system_r:kernel_t:s0 key=(null) Feb 9 09:03:16.551000 audit: CWD cwd="/" Feb 9 09:03:16.551000 audit: PATH item=0 name=(null) inode=2 dev=00:1a mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:unlabeled_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:16.551000 audit: PATH item=1 name=(null) inode=3 dev=00:1a mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:unlabeled_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:16.551000 audit: PROCTITLE proctitle=2F7573722F6C69622F73797374656D642F73797374656D2D67656E657261746F72732F746F7263782D67656E657261746F72002F72756E2F73797374656D642F67656E657261746F72002F72756E2F73797374656D642F67656E657261746F722E6561726C79002F72756E2F73797374656D642F67656E657261746F722E6C61 Feb 9 09:03:19.729000 audit: BPF prog-id=12 op=LOAD Feb 9 09:03:19.730000 audit: BPF prog-id=3 op=UNLOAD Feb 9 09:03:19.730000 audit: BPF prog-id=13 op=LOAD Feb 9 09:03:19.730000 audit: BPF prog-id=14 op=LOAD Feb 9 09:03:19.730000 audit: BPF prog-id=4 op=UNLOAD Feb 9 09:03:19.730000 audit: BPF prog-id=5 op=UNLOAD Feb 9 09:03:19.731000 audit: BPF prog-id=15 op=LOAD Feb 9 09:03:19.731000 audit: BPF prog-id=12 op=UNLOAD Feb 9 09:03:19.731000 audit: BPF prog-id=16 op=LOAD Feb 9 09:03:19.731000 audit: BPF prog-id=17 op=LOAD Feb 9 09:03:19.731000 audit: BPF prog-id=13 op=UNLOAD Feb 9 09:03:19.731000 audit: BPF prog-id=14 op=UNLOAD Feb 9 09:03:19.732000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:19.736000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:19.736000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:19.739000 audit: BPF prog-id=15 op=UNLOAD Feb 9 09:03:19.907000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:19.911000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:19.914000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:19.914000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:19.915000 audit: BPF prog-id=18 op=LOAD Feb 9 09:03:20.014436 systemd[1]: Finished modprobe@fuse.service. Feb 9 09:03:19.915000 audit: BPF prog-id=19 op=LOAD Feb 9 09:03:19.915000 audit: BPF prog-id=20 op=LOAD Feb 9 09:03:19.915000 audit: BPF prog-id=16 op=UNLOAD Feb 9 09:03:19.915000 audit: BPF prog-id=17 op=UNLOAD Feb 9 09:03:19.950000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=verity-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:19.981000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:19.987000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:19.987000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:19.995000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:19.995000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:20.001000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:20.002000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:20.007000 audit: CONFIG_CHANGE op=set audit_enabled=1 old=1 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 res=1 Feb 9 09:03:20.007000 audit[963]: SYSCALL arch=c000003e syscall=46 success=yes exit=60 a0=3 a1=7ffda0db6f30 a2=4000 a3=7ffda0db6fcc items=0 ppid=1 pid=963 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="systemd-journal" exe="/usr/lib/systemd/systemd-journald" subj=system_u:system_r:kernel_t:s0 key=(null) Feb 9 09:03:20.007000 audit: PROCTITLE proctitle="/usr/lib/systemd/systemd-journald" Feb 9 09:03:20.008000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:20.008000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:20.014000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:20.014000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:16.545960 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2024-02-09T09:03:16Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.2 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.2 /var/lib/torcx/store]" Feb 9 09:03:19.728447 systemd[1]: Queued start job for default target multi-user.target. Feb 9 09:03:20.016999 systemd[1]: Started systemd-journald.service. Feb 9 09:03:20.016000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:16.546560 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2024-02-09T09:03:16Z" level=debug msg="profile found" name=docker-1.12-no path=/usr/share/torcx/profiles/docker-1.12-no.json Feb 9 09:03:19.728462 systemd[1]: Unnecessary job was removed for dev-vda6.device. Feb 9 09:03:16.546586 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2024-02-09T09:03:16Z" level=debug msg="profile found" name=vendor path=/usr/share/torcx/profiles/vendor.json Feb 9 09:03:19.732630 systemd[1]: systemd-journald.service: Deactivated successfully. Feb 9 09:03:16.546628 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2024-02-09T09:03:16Z" level=info msg="no vendor profile selected by /etc/flatcar/docker-1.12" Feb 9 09:03:20.017016 systemd[1]: modprobe@loop.service: Deactivated successfully. Feb 9 09:03:16.546644 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2024-02-09T09:03:16Z" level=debug msg="skipped missing lower profile" missing profile=oem Feb 9 09:03:20.017165 systemd[1]: Finished modprobe@loop.service. Feb 9 09:03:16.546688 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2024-02-09T09:03:16Z" level=warning msg="no next profile: unable to read profile file: open /etc/torcx/next-profile: no such file or directory" Feb 9 09:03:16.546707 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2024-02-09T09:03:16Z" level=debug msg="apply configuration parsed" lower profiles (vendor/oem)="[vendor]" upper profile (user)= Feb 9 09:03:16.546938 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2024-02-09T09:03:16Z" level=debug msg="mounted tmpfs" target=/run/torcx/unpack Feb 9 09:03:20.018000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:20.018000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:16.546992 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2024-02-09T09:03:16Z" level=debug msg="profile found" name=docker-1.12-no path=/usr/share/torcx/profiles/docker-1.12-no.json Feb 9 09:03:20.019012 systemd[1]: Finished systemd-modules-load.service. Feb 9 09:03:20.019000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:16.547012 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2024-02-09T09:03:16Z" level=debug msg="profile found" name=vendor path=/usr/share/torcx/profiles/vendor.json Feb 9 09:03:20.019847 systemd[1]: Finished systemd-network-generator.service. Feb 9 09:03:16.547828 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2024-02-09T09:03:16Z" level=debug msg="new archive/reference added to cache" format=tgz name=docker path="/usr/share/torcx/store/docker:20.10.torcx.tgz" reference=20.10 Feb 9 09:03:16.547876 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2024-02-09T09:03:16Z" level=debug msg="new archive/reference added to cache" format=tgz name=docker path="/usr/share/torcx/store/docker:com.coreos.cl.torcx.tgz" reference=com.coreos.cl Feb 9 09:03:16.547903 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2024-02-09T09:03:16Z" level=info msg="store skipped" err="open /usr/share/oem/torcx/store/3510.3.2: no such file or directory" path=/usr/share/oem/torcx/store/3510.3.2 Feb 9 09:03:16.547926 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2024-02-09T09:03:16Z" level=info msg="store skipped" err="open /usr/share/oem/torcx/store: no such file or directory" path=/usr/share/oem/torcx/store Feb 9 09:03:16.547951 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2024-02-09T09:03:16Z" level=info msg="store skipped" err="open /var/lib/torcx/store/3510.3.2: no such file or directory" path=/var/lib/torcx/store/3510.3.2 Feb 9 09:03:16.547971 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2024-02-09T09:03:16Z" level=info msg="store skipped" err="open /var/lib/torcx/store: no such file or directory" path=/var/lib/torcx/store Feb 9 09:03:19.224971 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2024-02-09T09:03:19Z" level=debug msg="image unpacked" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Feb 9 09:03:20.022000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:19.225376 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2024-02-09T09:03:19Z" level=debug msg="binaries propagated" assets="[/bin/containerd /bin/containerd-shim /bin/ctr /bin/docker /bin/docker-containerd /bin/docker-containerd-shim /bin/docker-init /bin/docker-proxy /bin/docker-runc /bin/dockerd /bin/runc /bin/tini]" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Feb 9 09:03:19.225530 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2024-02-09T09:03:19Z" level=debug msg="networkd units propagated" assets="[/lib/systemd/network/50-docker.network /lib/systemd/network/90-docker-veth.network]" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Feb 9 09:03:19.225820 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2024-02-09T09:03:19Z" level=debug msg="systemd units propagated" assets="[/lib/systemd/system/containerd.service /lib/systemd/system/docker.service /lib/systemd/system/docker.socket /lib/systemd/system/sockets.target.wants /lib/systemd/system/multi-user.target.wants]" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Feb 9 09:03:19.225899 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2024-02-09T09:03:19Z" level=debug msg="profile applied" sealed profile=/run/torcx/profile.json upper profile= Feb 9 09:03:19.226003 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2024-02-09T09:03:19Z" level=debug msg="system state sealed" content="[TORCX_LOWER_PROFILES=\"vendor\" TORCX_UPPER_PROFILE=\"\" TORCX_PROFILE_PATH=\"/run/torcx/profile.json\" TORCX_BINDIR=\"/run/torcx/bin\" TORCX_UNPACKDIR=\"/run/torcx/unpack\"]" path=/run/metadata/torcx Feb 9 09:03:20.023388 systemd[1]: Finished systemd-remount-fs.service. Feb 9 09:03:20.023000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-remount-fs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:20.025088 systemd[1]: Reached target network-pre.target. Feb 9 09:03:20.028149 systemd[1]: Mounting sys-fs-fuse-connections.mount... Feb 9 09:03:20.030752 systemd[1]: Mounting sys-kernel-config.mount... Feb 9 09:03:20.031409 systemd[1]: remount-root.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Feb 9 09:03:20.037909 systemd[1]: Starting systemd-hwdb-update.service... Feb 9 09:03:20.040145 systemd[1]: Starting systemd-journal-flush.service... Feb 9 09:03:20.040978 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Feb 9 09:03:20.043498 systemd[1]: Starting systemd-random-seed.service... Feb 9 09:03:20.044707 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Feb 9 09:03:20.048687 systemd[1]: Starting systemd-sysctl.service... Feb 9 09:03:20.053000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=flatcar-tmpfiles comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:20.053383 systemd[1]: Finished flatcar-tmpfiles.service. Feb 9 09:03:20.054421 systemd[1]: Mounted sys-fs-fuse-connections.mount. Feb 9 09:03:20.055090 systemd[1]: Mounted sys-kernel-config.mount. Feb 9 09:03:20.059037 systemd[1]: Starting systemd-sysusers.service... Feb 9 09:03:20.059614 systemd-journald[963]: Time spent on flushing to /var/log/journal/ed9135f0930245b7812d42f261728148 is 46.885ms for 1160 entries. Feb 9 09:03:20.059614 systemd-journald[963]: System Journal (/var/log/journal/ed9135f0930245b7812d42f261728148) is 8.0M, max 195.6M, 187.6M free. Feb 9 09:03:20.122073 systemd-journald[963]: Received client request to flush runtime journal. Feb 9 09:03:20.078000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-random-seed comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:20.101000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:20.117000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysusers comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:20.078443 systemd[1]: Finished systemd-random-seed.service. Feb 9 09:03:20.079132 systemd[1]: Reached target first-boot-complete.target. Feb 9 09:03:20.100986 systemd[1]: Finished systemd-sysctl.service. Feb 9 09:03:20.117592 systemd[1]: Finished systemd-sysusers.service. Feb 9 09:03:20.127000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-flush comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:20.126592 systemd[1]: Finished systemd-journal-flush.service. Feb 9 09:03:20.129407 systemd[1]: Finished systemd-udev-trigger.service. Feb 9 09:03:20.129000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:20.132241 systemd[1]: Starting systemd-udev-settle.service... Feb 9 09:03:20.142589 udevadm[996]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in. Feb 9 09:03:20.850326 systemd[1]: Finished systemd-hwdb-update.service. Feb 9 09:03:20.857167 kernel: kauditd_printk_skb: 101 callbacks suppressed Feb 9 09:03:20.857323 kernel: audit: type=1130 audit(1707469400.850:142): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-hwdb-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:20.850000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-hwdb-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:20.860447 kernel: audit: type=1334 audit(1707469400.856:143): prog-id=21 op=LOAD Feb 9 09:03:20.856000 audit: BPF prog-id=21 op=LOAD Feb 9 09:03:20.859416 systemd[1]: Starting systemd-udevd.service... Feb 9 09:03:20.856000 audit: BPF prog-id=22 op=LOAD Feb 9 09:03:20.856000 audit: BPF prog-id=7 op=UNLOAD Feb 9 09:03:20.864258 kernel: audit: type=1334 audit(1707469400.856:144): prog-id=22 op=LOAD Feb 9 09:03:20.864367 kernel: audit: type=1334 audit(1707469400.856:145): prog-id=7 op=UNLOAD Feb 9 09:03:20.864400 kernel: audit: type=1334 audit(1707469400.856:146): prog-id=8 op=UNLOAD Feb 9 09:03:20.856000 audit: BPF prog-id=8 op=UNLOAD Feb 9 09:03:20.887252 systemd-udevd[997]: Using default interface naming scheme 'v252'. Feb 9 09:03:20.934027 kernel: audit: type=1130 audit(1707469400.923:147): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:20.934123 kernel: audit: type=1334 audit(1707469400.924:148): prog-id=23 op=LOAD Feb 9 09:03:20.923000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:20.924000 audit: BPF prog-id=23 op=LOAD Feb 9 09:03:20.922970 systemd[1]: Started systemd-udevd.service. Feb 9 09:03:20.930734 systemd[1]: Starting systemd-networkd.service... Feb 9 09:03:20.946682 kernel: audit: type=1334 audit(1707469400.938:149): prog-id=24 op=LOAD Feb 9 09:03:20.946764 kernel: audit: type=1334 audit(1707469400.938:150): prog-id=25 op=LOAD Feb 9 09:03:20.946787 kernel: audit: type=1334 audit(1707469400.938:151): prog-id=26 op=LOAD Feb 9 09:03:20.938000 audit: BPF prog-id=24 op=LOAD Feb 9 09:03:20.938000 audit: BPF prog-id=25 op=LOAD Feb 9 09:03:20.938000 audit: BPF prog-id=26 op=LOAD Feb 9 09:03:20.943177 systemd[1]: Starting systemd-userdbd.service... Feb 9 09:03:20.991000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-userdbd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:20.991305 systemd[1]: Started systemd-userdbd.service. Feb 9 09:03:21.018529 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 9 09:03:21.018999 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Feb 9 09:03:21.020966 systemd[1]: Starting modprobe@dm_mod.service... Feb 9 09:03:21.024256 systemd[1]: Starting modprobe@efi_pstore.service... Feb 9 09:03:21.031000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:21.031000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:21.032000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:21.032000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:21.029535 systemd[1]: Starting modprobe@loop.service... Feb 9 09:03:21.030115 systemd[1]: remount-root.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Feb 9 09:03:21.030198 systemd[1]: setup-nsswitch.service was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Feb 9 09:03:21.030375 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 9 09:03:21.030982 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Feb 9 09:03:21.031200 systemd[1]: Finished modprobe@dm_mod.service. Feb 9 09:03:21.032268 systemd[1]: modprobe@loop.service: Deactivated successfully. Feb 9 09:03:21.032448 systemd[1]: Finished modprobe@loop.service. Feb 9 09:03:21.041460 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Feb 9 09:03:21.041620 systemd[1]: Finished modprobe@efi_pstore.service. Feb 9 09:03:21.044000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:21.044000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:21.044970 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Feb 9 09:03:21.045026 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Feb 9 09:03:21.056543 systemd[1]: Condition check resulted in dev-ttyS0.device being skipped. Feb 9 09:03:21.064783 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device. Feb 9 09:03:21.115917 systemd-networkd[1002]: lo: Link UP Feb 9 09:03:21.115927 systemd-networkd[1002]: lo: Gained carrier Feb 9 09:03:21.116594 systemd-networkd[1002]: Enumeration completed Feb 9 09:03:21.116000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:21.116715 systemd[1]: Started systemd-networkd.service. Feb 9 09:03:21.117610 systemd-networkd[1002]: eth1: Configuring with /run/systemd/network/10-52:dc:4c:08:54:ef.network. Feb 9 09:03:21.118828 systemd-networkd[1002]: eth0: Configuring with /run/systemd/network/10-26:bd:54:55:23:ff.network. Feb 9 09:03:21.119729 systemd-networkd[1002]: eth1: Link UP Feb 9 09:03:21.119737 systemd-networkd[1002]: eth1: Gained carrier Feb 9 09:03:21.123465 systemd-networkd[1002]: eth0: Link UP Feb 9 09:03:21.123479 systemd-networkd[1002]: eth0: Gained carrier Feb 9 09:03:21.133374 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input2 Feb 9 09:03:21.145400 kernel: ACPI: button: Power Button [PWRF] Feb 9 09:03:21.149000 audit[1003]: AVC avc: denied { confidentiality } for pid=1003 comm="(udev-worker)" lockdown_reason="use of tracefs" scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=lockdown permissive=1 Feb 9 09:03:21.149000 audit[1003]: SYSCALL arch=c000003e syscall=175 success=yes exit=0 a0=563f39d52c20 a1=32194 a2=7ff925550bc5 a3=5 items=108 ppid=997 pid=1003 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="(udev-worker)" exe="/usr/bin/udevadm" subj=system_u:system_r:kernel_t:s0 key=(null) Feb 9 09:03:21.149000 audit: CWD cwd="/" Feb 9 09:03:21.149000 audit: PATH item=0 name=(null) inode=45 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=1 name=(null) inode=14077 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=2 name=(null) inode=14077 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=3 name=(null) inode=14078 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=4 name=(null) inode=14077 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=5 name=(null) inode=14079 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=6 name=(null) inode=14077 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=7 name=(null) inode=14080 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=8 name=(null) inode=14080 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=9 name=(null) inode=14081 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=10 name=(null) inode=14080 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=11 name=(null) inode=14082 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=12 name=(null) inode=14080 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=13 name=(null) inode=14083 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=14 name=(null) inode=14080 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=15 name=(null) inode=14084 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=16 name=(null) inode=14080 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=17 name=(null) inode=14085 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=18 name=(null) inode=14077 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=19 name=(null) inode=14086 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=20 name=(null) inode=14086 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=21 name=(null) inode=14087 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=22 name=(null) inode=14086 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=23 name=(null) inode=14088 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=24 name=(null) inode=14086 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=25 name=(null) inode=14089 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=26 name=(null) inode=14086 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=27 name=(null) inode=14090 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=28 name=(null) inode=14086 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=29 name=(null) inode=14091 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=30 name=(null) inode=14077 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=31 name=(null) inode=14092 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=32 name=(null) inode=14092 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=33 name=(null) inode=14093 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=34 name=(null) inode=14092 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=35 name=(null) inode=14094 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=36 name=(null) inode=14092 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=37 name=(null) inode=14095 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=38 name=(null) inode=14092 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=39 name=(null) inode=14096 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=40 name=(null) inode=14092 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=41 name=(null) inode=14097 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=42 name=(null) inode=14077 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=43 name=(null) inode=14098 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=44 name=(null) inode=14098 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=45 name=(null) inode=14099 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=46 name=(null) inode=14098 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=47 name=(null) inode=14100 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=48 name=(null) inode=14098 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=49 name=(null) inode=14101 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=50 name=(null) inode=14098 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=51 name=(null) inode=14102 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=52 name=(null) inode=14098 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=53 name=(null) inode=14431 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=54 name=(null) inode=45 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=55 name=(null) inode=14432 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=56 name=(null) inode=14432 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=57 name=(null) inode=14433 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=58 name=(null) inode=14432 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=59 name=(null) inode=14434 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=60 name=(null) inode=14432 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=61 name=(null) inode=14435 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=62 name=(null) inode=14435 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=63 name=(null) inode=14436 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=64 name=(null) inode=14435 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=65 name=(null) inode=14437 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=66 name=(null) inode=14435 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=67 name=(null) inode=14438 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=68 name=(null) inode=14435 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=69 name=(null) inode=14439 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=70 name=(null) inode=14435 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=71 name=(null) inode=14440 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=72 name=(null) inode=14432 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=73 name=(null) inode=14441 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=74 name=(null) inode=14441 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=75 name=(null) inode=14442 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=76 name=(null) inode=14441 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=77 name=(null) inode=14443 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=78 name=(null) inode=14441 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=79 name=(null) inode=14444 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=80 name=(null) inode=14441 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=81 name=(null) inode=14445 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=82 name=(null) inode=14441 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=83 name=(null) inode=14446 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=84 name=(null) inode=14432 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=85 name=(null) inode=14447 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=86 name=(null) inode=14447 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=87 name=(null) inode=14448 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=88 name=(null) inode=14447 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=89 name=(null) inode=14449 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=90 name=(null) inode=14447 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=91 name=(null) inode=14450 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=92 name=(null) inode=14447 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=93 name=(null) inode=14451 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=94 name=(null) inode=14447 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=95 name=(null) inode=14452 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=96 name=(null) inode=14432 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=97 name=(null) inode=14453 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=98 name=(null) inode=14453 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=99 name=(null) inode=14454 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=100 name=(null) inode=14453 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=101 name=(null) inode=14455 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=102 name=(null) inode=14453 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=103 name=(null) inode=14456 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=104 name=(null) inode=14453 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=105 name=(null) inode=14457 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=106 name=(null) inode=14453 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PATH item=107 name=(null) inode=14458 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 09:03:21.149000 audit: PROCTITLE proctitle="(udev-worker)" Feb 9 09:03:21.206376 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input3 Feb 9 09:03:21.212410 kernel: piix4_smbus 0000:00:01.3: SMBus Host Controller at 0x700, revision 0 Feb 9 09:03:21.212720 kernel: mousedev: PS/2 mouse device common for all mice Feb 9 09:03:21.345378 kernel: EDAC MC: Ver: 3.0.0 Feb 9 09:03:21.362043 systemd[1]: Finished systemd-udev-settle.service. Feb 9 09:03:21.362000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-settle comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:21.365204 systemd[1]: Starting lvm2-activation-early.service... Feb 9 09:03:21.387286 lvm[1035]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Feb 9 09:03:21.419858 systemd[1]: Finished lvm2-activation-early.service. Feb 9 09:03:21.420000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:21.420626 systemd[1]: Reached target cryptsetup.target. Feb 9 09:03:21.422543 systemd[1]: Starting lvm2-activation.service... Feb 9 09:03:21.429053 lvm[1036]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Feb 9 09:03:21.459964 systemd[1]: Finished lvm2-activation.service. Feb 9 09:03:21.460759 systemd[1]: Reached target local-fs-pre.target. Feb 9 09:03:21.460000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:21.463568 systemd[1]: Mounting media-configdrive.mount... Feb 9 09:03:21.464253 systemd[1]: var-lib-machines.mount was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Feb 9 09:03:21.464320 systemd[1]: Reached target machines.target. Feb 9 09:03:21.466279 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service... Feb 9 09:03:21.481789 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service. Feb 9 09:03:21.482000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-OEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:21.486369 kernel: ISO 9660 Extensions: RRIP_1991A Feb 9 09:03:21.488378 systemd[1]: Mounted media-configdrive.mount. Feb 9 09:03:21.489012 systemd[1]: Reached target local-fs.target. Feb 9 09:03:21.491398 systemd[1]: Starting ldconfig.service... Feb 9 09:03:21.492677 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Feb 9 09:03:21.492728 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Feb 9 09:03:21.494031 systemd[1]: Starting systemd-boot-update.service... Feb 9 09:03:21.497534 systemd[1]: Starting systemd-machine-id-commit.service... Feb 9 09:03:21.501674 systemd[1]: systemd-sysext.service was skipped because no trigger condition checks were met. Feb 9 09:03:21.501742 systemd[1]: ensure-sysext.service was skipped because no trigger condition checks were met. Feb 9 09:03:21.503116 systemd[1]: Starting systemd-tmpfiles-setup.service... Feb 9 09:03:21.512162 systemd[1]: boot.automount: Got automount request for /boot, triggered by 1042 (bootctl) Feb 9 09:03:21.513721 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service... Feb 9 09:03:21.569371 systemd-tmpfiles[1044]: /usr/lib/tmpfiles.d/legacy.conf:13: Duplicate line for path "/run/lock", ignoring. Feb 9 09:03:21.580135 systemd-tmpfiles[1044]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Feb 9 09:03:21.580752 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Feb 9 09:03:21.581430 systemd[1]: Finished systemd-machine-id-commit.service. Feb 9 09:03:21.581000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-machine-id-commit comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:21.588791 systemd-tmpfiles[1044]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Feb 9 09:03:21.639135 systemd-fsck[1048]: fsck.fat 4.2 (2021-01-31) Feb 9 09:03:21.639135 systemd-fsck[1048]: /dev/vda1: 789 files, 115332/258078 clusters Feb 9 09:03:21.642478 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service. Feb 9 09:03:21.642000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:21.645014 systemd[1]: Mounting boot.mount... Feb 9 09:03:21.669466 systemd[1]: Mounted boot.mount. Feb 9 09:03:21.693674 systemd[1]: Finished systemd-boot-update.service. Feb 9 09:03:21.694000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-boot-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:21.802194 systemd[1]: Finished systemd-tmpfiles-setup.service. Feb 9 09:03:21.802000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:21.804871 systemd[1]: Starting audit-rules.service... Feb 9 09:03:21.807538 systemd[1]: Starting clean-ca-certificates.service... Feb 9 09:03:21.811773 systemd[1]: Starting systemd-journal-catalog-update.service... Feb 9 09:03:21.814000 audit: BPF prog-id=27 op=LOAD Feb 9 09:03:21.818578 systemd[1]: Starting systemd-resolved.service... Feb 9 09:03:21.820000 audit: BPF prog-id=28 op=LOAD Feb 9 09:03:21.821874 systemd[1]: Starting systemd-timesyncd.service... Feb 9 09:03:21.824596 systemd[1]: Starting systemd-update-utmp.service... Feb 9 09:03:21.827658 systemd[1]: Finished clean-ca-certificates.service. Feb 9 09:03:21.827000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=clean-ca-certificates comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:21.828778 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Feb 9 09:03:21.839000 audit[1059]: SYSTEM_BOOT pid=1059 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg=' comm="systemd-update-utmp" exe="/usr/lib/systemd/systemd-update-utmp" hostname=? addr=? terminal=? res=success' Feb 9 09:03:21.846110 systemd[1]: Finished systemd-update-utmp.service. Feb 9 09:03:21.846000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-update-utmp comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:21.907986 systemd[1]: Finished systemd-journal-catalog-update.service. Feb 9 09:03:21.908000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-catalog-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 09:03:21.910000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=add_rule key=(null) list=5 res=1 Feb 9 09:03:21.910000 audit[1071]: SYSCALL arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7ffdb43e3a40 a2=420 a3=0 items=0 ppid=1051 pid=1071 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/sbin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Feb 9 09:03:21.910000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Feb 9 09:03:21.911694 augenrules[1071]: No rules Feb 9 09:03:21.915307 systemd[1]: Finished audit-rules.service. Feb 9 09:03:21.947063 systemd[1]: Started systemd-timesyncd.service. Feb 9 09:03:21.948169 systemd[1]: Reached target time-set.target. Feb 9 09:03:21.952476 ldconfig[1041]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Feb 9 09:03:21.959979 systemd[1]: Finished ldconfig.service. Feb 9 09:03:21.962546 systemd[1]: Starting systemd-update-done.service... Feb 9 09:03:21.966641 systemd-resolved[1055]: Positive Trust Anchors: Feb 9 09:03:21.966658 systemd-resolved[1055]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Feb 9 09:03:21.966738 systemd-resolved[1055]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test Feb 9 09:03:21.976204 systemd[1]: Finished systemd-update-done.service. Feb 9 09:03:21.977099 systemd-resolved[1055]: Using system hostname 'ci-3510.3.2-6-d69396b42b'. Feb 9 09:03:21.980151 systemd[1]: Started systemd-resolved.service. Feb 9 09:03:21.980930 systemd[1]: Reached target network.target. Feb 9 09:03:21.981492 systemd[1]: Reached target nss-lookup.target. Feb 9 09:03:21.982012 systemd[1]: Reached target sysinit.target. Feb 9 09:03:21.982691 systemd[1]: Started motdgen.path. Feb 9 09:03:21.983283 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path. Feb 9 09:03:21.984258 systemd[1]: Started logrotate.timer. Feb 9 09:03:21.985034 systemd[1]: Started mdadm.timer. Feb 9 09:03:21.985684 systemd[1]: Started systemd-tmpfiles-clean.timer. Feb 9 09:03:22.002649 systemd[1]: update-engine-stub.timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Feb 9 09:03:22.002719 systemd[1]: Reached target paths.target. Feb 9 09:03:22.003443 systemd[1]: Reached target timers.target. Feb 9 09:03:22.004392 systemd[1]: Listening on dbus.socket. Feb 9 09:03:22.006570 systemd[1]: Starting docker.socket... Feb 9 09:03:22.011693 systemd[1]: Listening on sshd.socket. Feb 9 09:03:22.012340 systemd[1]: systemd-pcrphase-sysinit.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Feb 9 09:03:22.012903 systemd[1]: Listening on docker.socket. Feb 9 09:03:22.013499 systemd[1]: Reached target sockets.target. Feb 9 09:03:22.013974 systemd[1]: Reached target basic.target. Feb 9 09:03:22.014606 systemd[1]: addon-config@usr-share-oem.service was skipped because no trigger condition checks were met. Feb 9 09:03:22.014650 systemd[1]: addon-run@usr-share-oem.service was skipped because no trigger condition checks were met. Feb 9 09:03:22.016277 systemd[1]: Starting containerd.service... Feb 9 09:03:22.019701 systemd[1]: Starting coreos-metadata-sshkeys@core.service... Feb 9 09:03:22.849722 systemd-timesyncd[1058]: Contacted time server 108.175.15.67:123 (0.flatcar.pool.ntp.org). Feb 9 09:03:22.849787 systemd-timesyncd[1058]: Initial clock synchronization to Fri 2024-02-09 09:03:22.849593 UTC. Feb 9 09:03:22.850095 systemd[1]: Starting dbus.service... Feb 9 09:03:22.852518 systemd-resolved[1055]: Clock change detected. Flushing caches. Feb 9 09:03:22.854630 systemd[1]: Starting enable-oem-cloudinit.service... Feb 9 09:03:22.857308 systemd[1]: Starting extend-filesystems.service... Feb 9 09:03:22.858109 systemd[1]: flatcar-setup-environment.service was skipped because of an unmet condition check (ConditionPathExists=/usr/share/oem/bin/flatcar-setup-environment). Feb 9 09:03:22.860998 systemd[1]: Starting motdgen.service... Feb 9 09:03:22.864389 systemd[1]: Starting prepare-cni-plugins.service... Feb 9 09:03:22.869720 systemd[1]: Starting prepare-critools.service... Feb 9 09:03:22.874265 systemd[1]: Starting ssh-key-proc-cmdline.service... Feb 9 09:03:22.876788 systemd[1]: Starting sshd-keygen.service... Feb 9 09:03:22.881970 systemd[1]: Starting systemd-logind.service... Feb 9 09:03:22.884341 systemd[1]: systemd-pcrphase.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Feb 9 09:03:22.884441 systemd[1]: tcsd.service was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Feb 9 09:03:22.885019 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Feb 9 09:03:22.926950 dbus-daemon[1082]: [system] SELinux support is enabled Feb 9 09:03:22.935024 jq[1085]: false Feb 9 09:03:22.935191 tar[1106]: ./ Feb 9 09:03:22.935191 tar[1106]: ./loopback Feb 9 09:03:22.886050 systemd[1]: Starting update-engine.service... Feb 9 09:03:22.890686 systemd[1]: Starting update-ssh-keys-after-ignition.service... Feb 9 09:03:22.944636 jq[1100]: true Feb 9 09:03:22.918593 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Feb 9 09:03:22.918841 systemd[1]: Finished ssh-key-proc-cmdline.service. Feb 9 09:03:22.948747 tar[1107]: crictl Feb 9 09:03:22.927122 systemd[1]: Started dbus.service. Feb 9 09:03:22.930743 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Feb 9 09:03:22.930776 systemd[1]: Reached target system-config.target. Feb 9 09:03:22.931424 systemd[1]: user-cloudinit-proc-cmdline.service was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Feb 9 09:03:22.932985 systemd[1]: Starting user-configdrive.service... Feb 9 09:03:22.945177 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Feb 9 09:03:22.945350 systemd[1]: Condition check resulted in enable-oem-cloudinit.service being skipped. Feb 9 09:03:22.961592 jq[1112]: true Feb 9 09:03:22.982250 extend-filesystems[1086]: Found vda Feb 9 09:03:22.982250 extend-filesystems[1086]: Found vda1 Feb 9 09:03:22.982250 extend-filesystems[1086]: Found vda2 Feb 9 09:03:22.982250 extend-filesystems[1086]: Found vda3 Feb 9 09:03:22.982250 extend-filesystems[1086]: Found usr Feb 9 09:03:22.982250 extend-filesystems[1086]: Found vda4 Feb 9 09:03:22.982250 extend-filesystems[1086]: Found vda6 Feb 9 09:03:22.982250 extend-filesystems[1086]: Found vda7 Feb 9 09:03:22.982250 extend-filesystems[1086]: Found vda9 Feb 9 09:03:22.982250 extend-filesystems[1086]: Checking size of /dev/vda9 Feb 9 09:03:22.991389 systemd[1]: motdgen.service: Deactivated successfully. Feb 9 09:03:23.018762 coreos-cloudinit[1114]: 2024/02/09 09:03:23 Checking availability of "cloud-drive" Feb 9 09:03:23.018762 coreos-cloudinit[1114]: 2024/02/09 09:03:23 Fetching user-data from datasource of type "cloud-drive" Feb 9 09:03:23.018762 coreos-cloudinit[1114]: 2024/02/09 09:03:23 Attempting to read from "/media/configdrive/openstack/latest/user_data" Feb 9 09:03:23.018762 coreos-cloudinit[1114]: 2024/02/09 09:03:23 Fetching meta-data from datasource of type "cloud-drive" Feb 9 09:03:23.018762 coreos-cloudinit[1114]: 2024/02/09 09:03:23 Attempting to read from "/media/configdrive/openstack/latest/meta_data.json" Feb 9 09:03:22.991642 systemd[1]: Finished motdgen.service. Feb 9 09:03:23.068759 update_engine[1099]: I0209 09:03:23.067862 1099 main.cc:92] Flatcar Update Engine starting Feb 9 09:03:23.077033 systemd[1]: Started update-engine.service. Feb 9 09:03:23.083007 coreos-cloudinit[1114]: Detected an Ignition config. Exiting... Feb 9 09:03:23.078054 systemd[1]: Finished user-configdrive.service. Feb 9 09:03:23.083677 update_engine[1099]: I0209 09:03:23.083390 1099 update_check_scheduler.cc:74] Next update check in 5m48s Feb 9 09:03:23.078732 systemd[1]: Reached target user-config.target. Feb 9 09:03:23.081287 systemd[1]: Started locksmithd.service. Feb 9 09:03:23.086078 extend-filesystems[1086]: Resized partition /dev/vda9 Feb 9 09:03:23.089814 extend-filesystems[1144]: resize2fs 1.46.5 (30-Dec-2021) Feb 9 09:03:23.099290 bash[1140]: Updated "/home/core/.ssh/authorized_keys" Feb 9 09:03:23.099148 systemd[1]: Finished update-ssh-keys-after-ignition.service. Feb 9 09:03:23.101959 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 15121403 blocks Feb 9 09:03:23.118691 tar[1106]: ./bandwidth Feb 9 09:03:23.160549 env[1108]: time="2024-02-09T09:03:23.160465919Z" level=info msg="starting containerd" revision=92b3a9d6f1b3bcc6dc74875cfdea653fe39f09c2 version=1.6.16 Feb 9 09:03:23.168721 systemd-logind[1096]: Watching system buttons on /dev/input/event1 (Power Button) Feb 9 09:03:23.169192 systemd-logind[1096]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Feb 9 09:03:23.176271 systemd-logind[1096]: New seat seat0. Feb 9 09:03:23.178298 kernel: EXT4-fs (vda9): resized filesystem to 15121403 Feb 9 09:03:23.191098 systemd[1]: Started systemd-logind.service. Feb 9 09:03:23.196586 extend-filesystems[1144]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Feb 9 09:03:23.196586 extend-filesystems[1144]: old_desc_blocks = 1, new_desc_blocks = 8 Feb 9 09:03:23.196586 extend-filesystems[1144]: The filesystem on /dev/vda9 is now 15121403 (4k) blocks long. Feb 9 09:03:23.201983 extend-filesystems[1086]: Resized filesystem in /dev/vda9 Feb 9 09:03:23.201983 extend-filesystems[1086]: Found vdb Feb 9 09:03:23.197586 systemd[1]: extend-filesystems.service: Deactivated successfully. Feb 9 09:03:23.197775 systemd[1]: Finished extend-filesystems.service. Feb 9 09:03:23.252699 env[1108]: time="2024-02-09T09:03:23.252623877Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Feb 9 09:03:23.252863 env[1108]: time="2024-02-09T09:03:23.252840325Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Feb 9 09:03:23.254886 env[1108]: time="2024-02-09T09:03:23.254847967Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/5.15.148-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Feb 9 09:03:23.255144 env[1108]: time="2024-02-09T09:03:23.255114791Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Feb 9 09:03:23.256826 env[1108]: time="2024-02-09T09:03:23.256785035Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Feb 9 09:03:23.256826 env[1108]: time="2024-02-09T09:03:23.256816683Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Feb 9 09:03:23.256949 env[1108]: time="2024-02-09T09:03:23.256831837Z" level=warning msg="failed to load plugin io.containerd.snapshotter.v1.devmapper" error="devmapper not configured" Feb 9 09:03:23.256949 env[1108]: time="2024-02-09T09:03:23.256842066Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Feb 9 09:03:23.256949 env[1108]: time="2024-02-09T09:03:23.256921122Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Feb 9 09:03:23.257165 env[1108]: time="2024-02-09T09:03:23.257145108Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Feb 9 09:03:23.257307 env[1108]: time="2024-02-09T09:03:23.257286192Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Feb 9 09:03:23.257307 env[1108]: time="2024-02-09T09:03:23.257305484Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Feb 9 09:03:23.257376 env[1108]: time="2024-02-09T09:03:23.257358436Z" level=warning msg="could not use snapshotter devmapper in metadata plugin" error="devmapper not configured" Feb 9 09:03:23.257376 env[1108]: time="2024-02-09T09:03:23.257369557Z" level=info msg="metadata content store policy set" policy=shared Feb 9 09:03:23.271465 env[1108]: time="2024-02-09T09:03:23.271324161Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Feb 9 09:03:23.271465 env[1108]: time="2024-02-09T09:03:23.271398570Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Feb 9 09:03:23.271465 env[1108]: time="2024-02-09T09:03:23.271421947Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Feb 9 09:03:23.271723 env[1108]: time="2024-02-09T09:03:23.271484518Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Feb 9 09:03:23.271723 env[1108]: time="2024-02-09T09:03:23.271535609Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Feb 9 09:03:23.271723 env[1108]: time="2024-02-09T09:03:23.271558987Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Feb 9 09:03:23.271723 env[1108]: time="2024-02-09T09:03:23.271571769Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Feb 9 09:03:23.271723 env[1108]: time="2024-02-09T09:03:23.271585138Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Feb 9 09:03:23.271723 env[1108]: time="2024-02-09T09:03:23.271597796Z" level=info msg="loading plugin \"io.containerd.service.v1.leases-service\"..." type=io.containerd.service.v1 Feb 9 09:03:23.271723 env[1108]: time="2024-02-09T09:03:23.271611166Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Feb 9 09:03:23.271723 env[1108]: time="2024-02-09T09:03:23.271623242Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Feb 9 09:03:23.271723 env[1108]: time="2024-02-09T09:03:23.271636288Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Feb 9 09:03:23.272109 env[1108]: time="2024-02-09T09:03:23.271775190Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Feb 9 09:03:23.272109 env[1108]: time="2024-02-09T09:03:23.271853157Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Feb 9 09:03:23.272109 env[1108]: time="2024-02-09T09:03:23.272097046Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Feb 9 09:03:23.272242 env[1108]: time="2024-02-09T09:03:23.272122108Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Feb 9 09:03:23.272242 env[1108]: time="2024-02-09T09:03:23.272137382Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Feb 9 09:03:23.272242 env[1108]: time="2024-02-09T09:03:23.272186897Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Feb 9 09:03:23.272242 env[1108]: time="2024-02-09T09:03:23.272200411Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Feb 9 09:03:23.272242 env[1108]: time="2024-02-09T09:03:23.272212254Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Feb 9 09:03:23.272242 env[1108]: time="2024-02-09T09:03:23.272223512Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Feb 9 09:03:23.272242 env[1108]: time="2024-02-09T09:03:23.272235307Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Feb 9 09:03:23.272557 env[1108]: time="2024-02-09T09:03:23.272247216Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Feb 9 09:03:23.272557 env[1108]: time="2024-02-09T09:03:23.272259310Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Feb 9 09:03:23.272557 env[1108]: time="2024-02-09T09:03:23.272271517Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Feb 9 09:03:23.272557 env[1108]: time="2024-02-09T09:03:23.272285846Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Feb 9 09:03:23.272557 env[1108]: time="2024-02-09T09:03:23.272403008Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Feb 9 09:03:23.272557 env[1108]: time="2024-02-09T09:03:23.272418772Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Feb 9 09:03:23.272557 env[1108]: time="2024-02-09T09:03:23.272430337Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Feb 9 09:03:23.272557 env[1108]: time="2024-02-09T09:03:23.272441602Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Feb 9 09:03:23.272557 env[1108]: time="2024-02-09T09:03:23.272456654Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="no OpenTelemetry endpoint: skip plugin" type=io.containerd.tracing.processor.v1 Feb 9 09:03:23.272557 env[1108]: time="2024-02-09T09:03:23.272470290Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Feb 9 09:03:23.272557 env[1108]: time="2024-02-09T09:03:23.272516797Z" level=error msg="failed to initialize a tracing processor \"otlp\"" error="no OpenTelemetry endpoint: skip plugin" Feb 9 09:03:23.272557 env[1108]: time="2024-02-09T09:03:23.272556282Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Feb 9 09:03:23.272927 env[1108]: time="2024-02-09T09:03:23.272761571Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.6 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Feb 9 09:03:23.272927 env[1108]: time="2024-02-09T09:03:23.272814976Z" level=info msg="Connect containerd service" Feb 9 09:03:23.272927 env[1108]: time="2024-02-09T09:03:23.272854093Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Feb 9 09:03:23.276316 env[1108]: time="2024-02-09T09:03:23.273405820Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Feb 9 09:03:23.276316 env[1108]: time="2024-02-09T09:03:23.274082847Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Feb 9 09:03:23.276316 env[1108]: time="2024-02-09T09:03:23.274230407Z" level=info msg=serving... address=/run/containerd/containerd.sock Feb 9 09:03:23.276316 env[1108]: time="2024-02-09T09:03:23.274279636Z" level=info msg="containerd successfully booted in 0.122237s" Feb 9 09:03:23.274822 systemd[1]: Started containerd.service. Feb 9 09:03:23.277282 env[1108]: time="2024-02-09T09:03:23.276757564Z" level=info msg="Start subscribing containerd event" Feb 9 09:03:23.277282 env[1108]: time="2024-02-09T09:03:23.276819561Z" level=info msg="Start recovering state" Feb 9 09:03:23.277282 env[1108]: time="2024-02-09T09:03:23.276892642Z" level=info msg="Start event monitor" Feb 9 09:03:23.277282 env[1108]: time="2024-02-09T09:03:23.276908980Z" level=info msg="Start snapshots syncer" Feb 9 09:03:23.277282 env[1108]: time="2024-02-09T09:03:23.276919483Z" level=info msg="Start cni network conf syncer for default" Feb 9 09:03:23.277282 env[1108]: time="2024-02-09T09:03:23.276930970Z" level=info msg="Start streaming server" Feb 9 09:03:23.290582 coreos-metadata[1081]: Feb 09 09:03:23.285 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Feb 9 09:03:23.291595 tar[1106]: ./ptp Feb 9 09:03:23.307236 coreos-metadata[1081]: Feb 09 09:03:23.307 INFO Fetch successful Feb 9 09:03:23.314200 unknown[1081]: wrote ssh authorized keys file for user: core Feb 9 09:03:23.329218 update-ssh-keys[1151]: Updated "/home/core/.ssh/authorized_keys" Feb 9 09:03:23.329858 systemd[1]: Finished coreos-metadata-sshkeys@core.service. Feb 9 09:03:23.366140 tar[1106]: ./vlan Feb 9 09:03:23.425224 tar[1106]: ./host-device Feb 9 09:03:23.484282 tar[1106]: ./tuning Feb 9 09:03:23.537981 tar[1106]: ./vrf Feb 9 09:03:23.543844 systemd-networkd[1002]: eth1: Gained IPv6LL Feb 9 09:03:23.594528 tar[1106]: ./sbr Feb 9 09:03:23.650794 tar[1106]: ./tap Feb 9 09:03:23.672780 systemd-networkd[1002]: eth0: Gained IPv6LL Feb 9 09:03:23.710305 tar[1106]: ./dhcp Feb 9 09:03:23.879264 tar[1106]: ./static Feb 9 09:03:23.940813 tar[1106]: ./firewall Feb 9 09:03:23.996301 tar[1106]: ./macvlan Feb 9 09:03:24.049382 tar[1106]: ./dummy Feb 9 09:03:24.137982 tar[1106]: ./bridge Feb 9 09:03:24.142057 systemd[1]: Finished prepare-critools.service. Feb 9 09:03:24.186929 tar[1106]: ./ipvlan Feb 9 09:03:24.229155 tar[1106]: ./portmap Feb 9 09:03:24.269528 tar[1106]: ./host-local Feb 9 09:03:24.316807 systemd[1]: Finished prepare-cni-plugins.service. Feb 9 09:03:24.384136 locksmithd[1143]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Feb 9 09:03:24.905535 sshd_keygen[1110]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Feb 9 09:03:24.943855 systemd[1]: Finished sshd-keygen.service. Feb 9 09:03:24.946349 systemd[1]: Starting issuegen.service... Feb 9 09:03:24.953991 systemd[1]: issuegen.service: Deactivated successfully. Feb 9 09:03:24.954214 systemd[1]: Finished issuegen.service. Feb 9 09:03:24.957286 systemd[1]: Starting systemd-user-sessions.service... Feb 9 09:03:24.968527 systemd[1]: Finished systemd-user-sessions.service. Feb 9 09:03:24.971932 systemd[1]: Started getty@tty1.service. Feb 9 09:03:24.975674 systemd[1]: Started serial-getty@ttyS0.service. Feb 9 09:03:24.976895 systemd[1]: Reached target getty.target. Feb 9 09:03:24.978011 systemd[1]: Reached target multi-user.target. Feb 9 09:03:24.981026 systemd[1]: Starting systemd-update-utmp-runlevel.service... Feb 9 09:03:24.992047 systemd[1]: systemd-update-utmp-runlevel.service: Deactivated successfully. Feb 9 09:03:24.992303 systemd[1]: Finished systemd-update-utmp-runlevel.service. Feb 9 09:03:24.993204 systemd[1]: Startup finished in 1.041s (kernel) + 6.617s (initrd) + 7.861s (userspace) = 15.520s. Feb 9 09:03:32.414439 systemd[1]: Created slice system-sshd.slice. Feb 9 09:03:32.416601 systemd[1]: Started sshd@0-24.144.92.253:22-139.178.89.65:33968.service. Feb 9 09:03:32.490510 sshd[1175]: Accepted publickey for core from 139.178.89.65 port 33968 ssh2: RSA SHA256:zxCjWE6I1sqRNr8f+A5DoPj4YLVmU7ObDiNpO/GSq00 Feb 9 09:03:32.492522 sshd[1175]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Feb 9 09:03:32.505387 systemd[1]: Created slice user-500.slice. Feb 9 09:03:32.506818 systemd[1]: Starting user-runtime-dir@500.service... Feb 9 09:03:32.514941 systemd-logind[1096]: New session 1 of user core. Feb 9 09:03:32.520767 systemd[1]: Finished user-runtime-dir@500.service. Feb 9 09:03:32.523101 systemd[1]: Starting user@500.service... Feb 9 09:03:32.529198 (systemd)[1178]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Feb 9 09:03:32.666027 systemd[1178]: Queued start job for default target default.target. Feb 9 09:03:32.668197 systemd[1178]: Reached target paths.target. Feb 9 09:03:32.668526 systemd[1178]: Reached target sockets.target. Feb 9 09:03:32.668723 systemd[1178]: Reached target timers.target. Feb 9 09:03:32.668883 systemd[1178]: Reached target basic.target. Feb 9 09:03:32.669184 systemd[1]: Started user@500.service. Feb 9 09:03:32.670625 systemd[1178]: Reached target default.target. Feb 9 09:03:32.670813 systemd[1]: Started session-1.scope. Feb 9 09:03:32.671657 systemd[1178]: Startup finished in 131ms. Feb 9 09:03:32.736369 systemd[1]: Started sshd@1-24.144.92.253:22-139.178.89.65:33972.service. Feb 9 09:03:32.787200 sshd[1187]: Accepted publickey for core from 139.178.89.65 port 33972 ssh2: RSA SHA256:zxCjWE6I1sqRNr8f+A5DoPj4YLVmU7ObDiNpO/GSq00 Feb 9 09:03:32.789388 sshd[1187]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Feb 9 09:03:32.796853 systemd-logind[1096]: New session 2 of user core. Feb 9 09:03:32.798474 systemd[1]: Started session-2.scope. Feb 9 09:03:32.869048 sshd[1187]: pam_unix(sshd:session): session closed for user core Feb 9 09:03:32.875359 systemd[1]: sshd@1-24.144.92.253:22-139.178.89.65:33972.service: Deactivated successfully. Feb 9 09:03:32.876127 systemd[1]: session-2.scope: Deactivated successfully. Feb 9 09:03:32.878031 systemd-logind[1096]: Session 2 logged out. Waiting for processes to exit. Feb 9 09:03:32.879814 systemd[1]: Started sshd@2-24.144.92.253:22-139.178.89.65:33984.service. Feb 9 09:03:32.881153 systemd-logind[1096]: Removed session 2. Feb 9 09:03:32.933614 sshd[1193]: Accepted publickey for core from 139.178.89.65 port 33984 ssh2: RSA SHA256:zxCjWE6I1sqRNr8f+A5DoPj4YLVmU7ObDiNpO/GSq00 Feb 9 09:03:32.935407 sshd[1193]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Feb 9 09:03:32.943594 systemd-logind[1096]: New session 3 of user core. Feb 9 09:03:32.943791 systemd[1]: Started session-3.scope. Feb 9 09:03:33.005230 sshd[1193]: pam_unix(sshd:session): session closed for user core Feb 9 09:03:33.010954 systemd[1]: sshd@2-24.144.92.253:22-139.178.89.65:33984.service: Deactivated successfully. Feb 9 09:03:33.011774 systemd[1]: session-3.scope: Deactivated successfully. Feb 9 09:03:33.012539 systemd-logind[1096]: Session 3 logged out. Waiting for processes to exit. Feb 9 09:03:33.014319 systemd[1]: Started sshd@3-24.144.92.253:22-139.178.89.65:33986.service. Feb 9 09:03:33.017906 systemd-logind[1096]: Removed session 3. Feb 9 09:03:33.067400 sshd[1199]: Accepted publickey for core from 139.178.89.65 port 33986 ssh2: RSA SHA256:zxCjWE6I1sqRNr8f+A5DoPj4YLVmU7ObDiNpO/GSq00 Feb 9 09:03:33.069986 sshd[1199]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Feb 9 09:03:33.077095 systemd-logind[1096]: New session 4 of user core. Feb 9 09:03:33.077953 systemd[1]: Started session-4.scope. Feb 9 09:03:33.146378 sshd[1199]: pam_unix(sshd:session): session closed for user core Feb 9 09:03:33.153094 systemd[1]: sshd@3-24.144.92.253:22-139.178.89.65:33986.service: Deactivated successfully. Feb 9 09:03:33.154153 systemd[1]: session-4.scope: Deactivated successfully. Feb 9 09:03:33.155221 systemd-logind[1096]: Session 4 logged out. Waiting for processes to exit. Feb 9 09:03:33.157108 systemd[1]: Started sshd@4-24.144.92.253:22-139.178.89.65:34000.service. Feb 9 09:03:33.158208 systemd-logind[1096]: Removed session 4. Feb 9 09:03:33.209174 sshd[1205]: Accepted publickey for core from 139.178.89.65 port 34000 ssh2: RSA SHA256:zxCjWE6I1sqRNr8f+A5DoPj4YLVmU7ObDiNpO/GSq00 Feb 9 09:03:33.212343 sshd[1205]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Feb 9 09:03:33.220786 systemd-logind[1096]: New session 5 of user core. Feb 9 09:03:33.221062 systemd[1]: Started session-5.scope. Feb 9 09:03:33.299064 sudo[1208]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Feb 9 09:03:33.299886 sudo[1208]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500) Feb 9 09:03:33.904390 systemd[1]: Reloading. Feb 9 09:03:34.008795 /usr/lib/systemd/system-generators/torcx-generator[1237]: time="2024-02-09T09:03:34Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.2 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.2 /var/lib/torcx/store]" Feb 9 09:03:34.008833 /usr/lib/systemd/system-generators/torcx-generator[1237]: time="2024-02-09T09:03:34Z" level=info msg="torcx already run" Feb 9 09:03:34.142949 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Feb 9 09:03:34.143235 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Feb 9 09:03:34.164120 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 9 09:03:34.290322 systemd[1]: Starting systemd-networkd-wait-online.service... Feb 9 09:03:34.299572 systemd[1]: Finished systemd-networkd-wait-online.service. Feb 9 09:03:34.301483 systemd[1]: Reached target network-online.target. Feb 9 09:03:34.304867 systemd[1]: Started kubelet.service. Feb 9 09:03:34.326834 systemd[1]: Starting coreos-metadata.service... Feb 9 09:03:34.377537 coreos-metadata[1292]: Feb 09 09:03:34.377 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Feb 9 09:03:34.391411 coreos-metadata[1292]: Feb 09 09:03:34.391 INFO Fetch successful Feb 9 09:03:34.404966 kubelet[1284]: E0209 09:03:34.404889 1284 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 9 09:03:34.407689 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 9 09:03:34.407894 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 9 09:03:34.413162 systemd[1]: Finished coreos-metadata.service. Feb 9 09:03:34.832699 systemd[1]: Stopped kubelet.service. Feb 9 09:03:34.856479 systemd[1]: Reloading. Feb 9 09:03:34.940325 /usr/lib/systemd/system-generators/torcx-generator[1347]: time="2024-02-09T09:03:34Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.2 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.2 /var/lib/torcx/store]" Feb 9 09:03:34.940356 /usr/lib/systemd/system-generators/torcx-generator[1347]: time="2024-02-09T09:03:34Z" level=info msg="torcx already run" Feb 9 09:03:35.056116 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Feb 9 09:03:35.056458 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Feb 9 09:03:35.077321 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 9 09:03:35.181955 systemd[1]: Started kubelet.service. Feb 9 09:03:35.258806 kubelet[1395]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 9 09:03:35.258806 kubelet[1395]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Feb 9 09:03:35.258806 kubelet[1395]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 9 09:03:35.259256 kubelet[1395]: I0209 09:03:35.258886 1395 server.go:203] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 9 09:03:35.770610 kubelet[1395]: I0209 09:03:35.770562 1395 server.go:467] "Kubelet version" kubeletVersion="v1.28.1" Feb 9 09:03:35.770861 kubelet[1395]: I0209 09:03:35.770840 1395 server.go:469] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 9 09:03:35.771476 kubelet[1395]: I0209 09:03:35.771451 1395 server.go:895] "Client rotation is on, will bootstrap in background" Feb 9 09:03:35.775639 kubelet[1395]: I0209 09:03:35.775599 1395 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Feb 9 09:03:35.784315 kubelet[1395]: I0209 09:03:35.784273 1395 server.go:725] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Feb 9 09:03:35.784907 kubelet[1395]: I0209 09:03:35.784884 1395 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 9 09:03:35.785263 kubelet[1395]: I0209 09:03:35.785238 1395 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} Feb 9 09:03:35.785466 kubelet[1395]: I0209 09:03:35.785447 1395 topology_manager.go:138] "Creating topology manager with none policy" Feb 9 09:03:35.785643 kubelet[1395]: I0209 09:03:35.785590 1395 container_manager_linux.go:301] "Creating device plugin manager" Feb 9 09:03:35.785908 kubelet[1395]: I0209 09:03:35.785890 1395 state_mem.go:36] "Initialized new in-memory state store" Feb 9 09:03:35.786130 kubelet[1395]: I0209 09:03:35.786117 1395 kubelet.go:393] "Attempting to sync node with API server" Feb 9 09:03:35.790713 kubelet[1395]: I0209 09:03:35.790676 1395 kubelet.go:298] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 9 09:03:35.790927 kubelet[1395]: I0209 09:03:35.790911 1395 kubelet.go:309] "Adding apiserver pod source" Feb 9 09:03:35.791070 kubelet[1395]: I0209 09:03:35.791054 1395 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 9 09:03:35.791551 kubelet[1395]: E0209 09:03:35.791530 1395 file.go:98] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:03:35.791765 kubelet[1395]: E0209 09:03:35.791739 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:03:35.791958 kubelet[1395]: I0209 09:03:35.791943 1395 kuberuntime_manager.go:257] "Container runtime initialized" containerRuntime="containerd" version="1.6.16" apiVersion="v1" Feb 9 09:03:35.792783 kubelet[1395]: W0209 09:03:35.792765 1395 probe.go:268] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Feb 9 09:03:35.795485 kubelet[1395]: I0209 09:03:35.795435 1395 server.go:1232] "Started kubelet" Feb 9 09:03:35.797029 kubelet[1395]: E0209 09:03:35.796988 1395 cri_stats_provider.go:448] "Failed to get the info of the filesystem with mountpoint" err="unable to find data in memory cache" mountpoint="/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs" Feb 9 09:03:35.797029 kubelet[1395]: E0209 09:03:35.797028 1395 kubelet.go:1431] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Feb 9 09:03:35.801405 kernel: SELinux: Context system_u:object_r:container_file_t:s0 is not valid (left unmapped). Feb 9 09:03:35.801564 kubelet[1395]: E0209 09:03:35.799844 1395 event.go:280] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"24.144.92.253.17b2266319540026", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"24.144.92.253", UID:"24.144.92.253", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"Starting", Message:"Starting kubelet.", Source:v1.EventSource{Component:"kubelet", Host:"24.144.92.253"}, FirstTimestamp:time.Date(2024, time.February, 9, 9, 3, 35, 795392550, time.Local), LastTimestamp:time.Date(2024, time.February, 9, 9, 3, 35, 795392550, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"kubelet", ReportingInstance:"24.144.92.253"}': 'events is forbidden: User "system:anonymous" cannot create resource "events" in API group "" in the namespace "default"' (will not retry!) Feb 9 09:03:35.801721 kubelet[1395]: W0209 09:03:35.801685 1395 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.Node: nodes "24.144.92.253" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Feb 9 09:03:35.801793 kubelet[1395]: E0209 09:03:35.801724 1395 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.Node: failed to list *v1.Node: nodes "24.144.92.253" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Feb 9 09:03:35.801793 kubelet[1395]: W0209 09:03:35.801765 1395 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Feb 9 09:03:35.801793 kubelet[1395]: E0209 09:03:35.801778 1395 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Feb 9 09:03:35.802171 kubelet[1395]: I0209 09:03:35.802149 1395 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 9 09:03:35.804455 kubelet[1395]: E0209 09:03:35.803852 1395 event.go:280] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"24.144.92.253.17b22663196cbcd8", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"24.144.92.253", UID:"24.144.92.253", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"InvalidDiskCapacity", Message:"invalid capacity 0 on image filesystem", Source:v1.EventSource{Component:"kubelet", Host:"24.144.92.253"}, FirstTimestamp:time.Date(2024, time.February, 9, 9, 3, 35, 797013720, time.Local), LastTimestamp:time.Date(2024, time.February, 9, 9, 3, 35, 797013720, time.Local), Count:1, Type:"Warning", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"kubelet", ReportingInstance:"24.144.92.253"}': 'events is forbidden: User "system:anonymous" cannot create resource "events" in API group "" in the namespace "default"' (will not retry!) Feb 9 09:03:35.807765 kubelet[1395]: I0209 09:03:35.802183 1395 ratelimit.go:65] "Setting rate limiting for podresources endpoint" qps=100 burstTokens=10 Feb 9 09:03:35.807999 kubelet[1395]: I0209 09:03:35.807979 1395 server.go:233] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 9 09:03:35.808059 kubelet[1395]: I0209 09:03:35.802214 1395 server.go:162] "Starting to listen" address="0.0.0.0" port=10250 Feb 9 09:03:35.808931 kubelet[1395]: I0209 09:03:35.808886 1395 server.go:462] "Adding debug handlers to kubelet server" Feb 9 09:03:35.811162 kubelet[1395]: E0209 09:03:35.811126 1395 kubelet_node_status.go:458] "Error getting the current node from lister" err="node \"24.144.92.253\" not found" Feb 9 09:03:35.811162 kubelet[1395]: I0209 09:03:35.811155 1395 volume_manager.go:291] "Starting Kubelet Volume Manager" Feb 9 09:03:35.811407 kubelet[1395]: I0209 09:03:35.811265 1395 desired_state_of_world_populator.go:151] "Desired state populator starts to run" Feb 9 09:03:35.811407 kubelet[1395]: I0209 09:03:35.811318 1395 reconciler_new.go:29] "Reconciler: start to sync state" Feb 9 09:03:35.817862 kubelet[1395]: E0209 09:03:35.817821 1395 controller.go:146] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"24.144.92.253\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="200ms" Feb 9 09:03:35.818026 kubelet[1395]: W0209 09:03:35.817906 1395 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Feb 9 09:03:35.818026 kubelet[1395]: E0209 09:03:35.817948 1395 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Feb 9 09:03:35.853646 kubelet[1395]: I0209 09:03:35.853616 1395 cpu_manager.go:214] "Starting CPU manager" policy="none" Feb 9 09:03:35.853836 kubelet[1395]: I0209 09:03:35.853819 1395 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Feb 9 09:03:35.853970 kubelet[1395]: I0209 09:03:35.853957 1395 state_mem.go:36] "Initialized new in-memory state store" Feb 9 09:03:35.854304 kubelet[1395]: E0209 09:03:35.854141 1395 event.go:280] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"24.144.92.253.17b226631cbe3767", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"24.144.92.253", UID:"24.144.92.253", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientMemory", Message:"Node 24.144.92.253 status is now: NodeHasSufficientMemory", Source:v1.EventSource{Component:"kubelet", Host:"24.144.92.253"}, FirstTimestamp:time.Date(2024, time.February, 9, 9, 3, 35, 852685159, time.Local), LastTimestamp:time.Date(2024, time.February, 9, 9, 3, 35, 852685159, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"kubelet", ReportingInstance:"24.144.92.253"}': 'events is forbidden: User "system:anonymous" cannot create resource "events" in API group "" in the namespace "default"' (will not retry!) Feb 9 09:03:35.856241 kubelet[1395]: E0209 09:03:35.856100 1395 event.go:280] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"24.144.92.253.17b226631cbe4e7d", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"24.144.92.253", UID:"24.144.92.253", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasNoDiskPressure", Message:"Node 24.144.92.253 status is now: NodeHasNoDiskPressure", Source:v1.EventSource{Component:"kubelet", Host:"24.144.92.253"}, FirstTimestamp:time.Date(2024, time.February, 9, 9, 3, 35, 852691069, time.Local), LastTimestamp:time.Date(2024, time.February, 9, 9, 3, 35, 852691069, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"kubelet", ReportingInstance:"24.144.92.253"}': 'events is forbidden: User "system:anonymous" cannot create resource "events" in API group "" in the namespace "default"' (will not retry!) Feb 9 09:03:35.857280 kubelet[1395]: I0209 09:03:35.857258 1395 policy_none.go:49] "None policy: Start" Feb 9 09:03:35.857710 kubelet[1395]: E0209 09:03:35.857540 1395 event.go:280] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"24.144.92.253.17b226631cbe5bd8", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"24.144.92.253", UID:"24.144.92.253", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientPID", Message:"Node 24.144.92.253 status is now: NodeHasSufficientPID", Source:v1.EventSource{Component:"kubelet", Host:"24.144.92.253"}, FirstTimestamp:time.Date(2024, time.February, 9, 9, 3, 35, 852694488, time.Local), LastTimestamp:time.Date(2024, time.February, 9, 9, 3, 35, 852694488, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"kubelet", ReportingInstance:"24.144.92.253"}': 'events is forbidden: User "system:anonymous" cannot create resource "events" in API group "" in the namespace "default"' (will not retry!) Feb 9 09:03:35.859908 kubelet[1395]: I0209 09:03:35.859886 1395 memory_manager.go:169] "Starting memorymanager" policy="None" Feb 9 09:03:35.860049 kubelet[1395]: I0209 09:03:35.860035 1395 state_mem.go:35] "Initializing new in-memory state store" Feb 9 09:03:35.870737 systemd[1]: Created slice kubepods.slice. Feb 9 09:03:35.879507 systemd[1]: Created slice kubepods-burstable.slice. Feb 9 09:03:35.884774 systemd[1]: Created slice kubepods-besteffort.slice. Feb 9 09:03:35.891781 kubelet[1395]: I0209 09:03:35.891749 1395 manager.go:471] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 9 09:03:35.892337 kubelet[1395]: I0209 09:03:35.892315 1395 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 9 09:03:35.896466 kubelet[1395]: E0209 09:03:35.896422 1395 eviction_manager.go:258] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"24.144.92.253\" not found" Feb 9 09:03:35.898949 kubelet[1395]: E0209 09:03:35.898742 1395 event.go:280] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"24.144.92.253.17b226631f521c16", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"24.144.92.253", UID:"24.144.92.253", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeAllocatableEnforced", Message:"Updated Node Allocatable limit across pods", Source:v1.EventSource{Component:"kubelet", Host:"24.144.92.253"}, FirstTimestamp:time.Date(2024, time.February, 9, 9, 3, 35, 895931926, time.Local), LastTimestamp:time.Date(2024, time.February, 9, 9, 3, 35, 895931926, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"kubelet", ReportingInstance:"24.144.92.253"}': 'events is forbidden: User "system:anonymous" cannot create resource "events" in API group "" in the namespace "default"' (will not retry!) Feb 9 09:03:35.912427 kubelet[1395]: I0209 09:03:35.912392 1395 kubelet_node_status.go:70] "Attempting to register node" node="24.144.92.253" Feb 9 09:03:35.914336 kubelet[1395]: E0209 09:03:35.914298 1395 kubelet_node_status.go:92] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="24.144.92.253" Feb 9 09:03:35.915050 kubelet[1395]: E0209 09:03:35.914925 1395 event.go:280] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"24.144.92.253.17b226631cbe3767", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"24.144.92.253", UID:"24.144.92.253", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientMemory", Message:"Node 24.144.92.253 status is now: NodeHasSufficientMemory", Source:v1.EventSource{Component:"kubelet", Host:"24.144.92.253"}, FirstTimestamp:time.Date(2024, time.February, 9, 9, 3, 35, 852685159, time.Local), LastTimestamp:time.Date(2024, time.February, 9, 9, 3, 35, 912328378, time.Local), Count:2, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"kubelet", ReportingInstance:"24.144.92.253"}': 'events "24.144.92.253.17b226631cbe3767" is forbidden: User "system:anonymous" cannot patch resource "events" in API group "" in the namespace "default"' (will not retry!) Feb 9 09:03:35.916759 kubelet[1395]: E0209 09:03:35.916674 1395 event.go:280] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"24.144.92.253.17b226631cbe4e7d", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"24.144.92.253", UID:"24.144.92.253", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasNoDiskPressure", Message:"Node 24.144.92.253 status is now: NodeHasNoDiskPressure", Source:v1.EventSource{Component:"kubelet", Host:"24.144.92.253"}, FirstTimestamp:time.Date(2024, time.February, 9, 9, 3, 35, 852691069, time.Local), LastTimestamp:time.Date(2024, time.February, 9, 9, 3, 35, 912348221, time.Local), Count:2, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"kubelet", ReportingInstance:"24.144.92.253"}': 'events "24.144.92.253.17b226631cbe4e7d" is forbidden: User "system:anonymous" cannot patch resource "events" in API group "" in the namespace "default"' (will not retry!) Feb 9 09:03:35.918540 kubelet[1395]: E0209 09:03:35.918427 1395 event.go:280] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"24.144.92.253.17b226631cbe5bd8", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"24.144.92.253", UID:"24.144.92.253", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientPID", Message:"Node 24.144.92.253 status is now: NodeHasSufficientPID", Source:v1.EventSource{Component:"kubelet", Host:"24.144.92.253"}, FirstTimestamp:time.Date(2024, time.February, 9, 9, 3, 35, 852694488, time.Local), LastTimestamp:time.Date(2024, time.February, 9, 9, 3, 35, 912352238, time.Local), Count:2, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"kubelet", ReportingInstance:"24.144.92.253"}': 'events "24.144.92.253.17b226631cbe5bd8" is forbidden: User "system:anonymous" cannot patch resource "events" in API group "" in the namespace "default"' (will not retry!) Feb 9 09:03:35.931058 kubelet[1395]: I0209 09:03:35.931011 1395 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 9 09:03:35.932632 kubelet[1395]: I0209 09:03:35.932598 1395 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 9 09:03:35.932833 kubelet[1395]: I0209 09:03:35.932820 1395 status_manager.go:217] "Starting to sync pod status with apiserver" Feb 9 09:03:35.932959 kubelet[1395]: I0209 09:03:35.932945 1395 kubelet.go:2303] "Starting kubelet main sync loop" Feb 9 09:03:35.933833 kubelet[1395]: E0209 09:03:35.933803 1395 kubelet.go:2327] "Skipping pod synchronization" err="PLEG is not healthy: pleg has yet to be successful" Feb 9 09:03:35.935648 kubelet[1395]: W0209 09:03:35.935624 1395 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Feb 9 09:03:35.935793 kubelet[1395]: E0209 09:03:35.935778 1395 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Feb 9 09:03:36.021552 kubelet[1395]: E0209 09:03:36.020667 1395 controller.go:146] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"24.144.92.253\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="400ms" Feb 9 09:03:36.116165 kubelet[1395]: I0209 09:03:36.116121 1395 kubelet_node_status.go:70] "Attempting to register node" node="24.144.92.253" Feb 9 09:03:36.118303 kubelet[1395]: E0209 09:03:36.118263 1395 kubelet_node_status.go:92] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="24.144.92.253" Feb 9 09:03:36.118569 kubelet[1395]: E0209 09:03:36.118271 1395 event.go:280] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"24.144.92.253.17b226631cbe3767", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"24.144.92.253", UID:"24.144.92.253", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientMemory", Message:"Node 24.144.92.253 status is now: NodeHasSufficientMemory", Source:v1.EventSource{Component:"kubelet", Host:"24.144.92.253"}, FirstTimestamp:time.Date(2024, time.February, 9, 9, 3, 35, 852685159, time.Local), LastTimestamp:time.Date(2024, time.February, 9, 9, 3, 36, 116062698, time.Local), Count:3, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"kubelet", ReportingInstance:"24.144.92.253"}': 'events "24.144.92.253.17b226631cbe3767" is forbidden: User "system:anonymous" cannot patch resource "events" in API group "" in the namespace "default"' (will not retry!) Feb 9 09:03:36.120718 kubelet[1395]: E0209 09:03:36.120646 1395 event.go:280] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"24.144.92.253.17b226631cbe4e7d", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"24.144.92.253", UID:"24.144.92.253", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasNoDiskPressure", Message:"Node 24.144.92.253 status is now: NodeHasNoDiskPressure", Source:v1.EventSource{Component:"kubelet", Host:"24.144.92.253"}, FirstTimestamp:time.Date(2024, time.February, 9, 9, 3, 35, 852691069, time.Local), LastTimestamp:time.Date(2024, time.February, 9, 9, 3, 36, 116075238, time.Local), Count:3, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"kubelet", ReportingInstance:"24.144.92.253"}': 'events "24.144.92.253.17b226631cbe4e7d" is forbidden: User "system:anonymous" cannot patch resource "events" in API group "" in the namespace "default"' (will not retry!) Feb 9 09:03:36.122159 kubelet[1395]: E0209 09:03:36.122096 1395 event.go:280] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"24.144.92.253.17b226631cbe5bd8", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"24.144.92.253", UID:"24.144.92.253", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientPID", Message:"Node 24.144.92.253 status is now: NodeHasSufficientPID", Source:v1.EventSource{Component:"kubelet", Host:"24.144.92.253"}, FirstTimestamp:time.Date(2024, time.February, 9, 9, 3, 35, 852694488, time.Local), LastTimestamp:time.Date(2024, time.February, 9, 9, 3, 36, 116078742, time.Local), Count:3, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"kubelet", ReportingInstance:"24.144.92.253"}': 'events "24.144.92.253.17b226631cbe5bd8" is forbidden: User "system:anonymous" cannot patch resource "events" in API group "" in the namespace "default"' (will not retry!) Feb 9 09:03:36.422819 kubelet[1395]: E0209 09:03:36.422673 1395 controller.go:146] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"24.144.92.253\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="800ms" Feb 9 09:03:36.520567 kubelet[1395]: I0209 09:03:36.520519 1395 kubelet_node_status.go:70] "Attempting to register node" node="24.144.92.253" Feb 9 09:03:36.522182 kubelet[1395]: E0209 09:03:36.522079 1395 event.go:280] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"24.144.92.253.17b226631cbe3767", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"24.144.92.253", UID:"24.144.92.253", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientMemory", Message:"Node 24.144.92.253 status is now: NodeHasSufficientMemory", Source:v1.EventSource{Component:"kubelet", Host:"24.144.92.253"}, FirstTimestamp:time.Date(2024, time.February, 9, 9, 3, 35, 852685159, time.Local), LastTimestamp:time.Date(2024, time.February, 9, 9, 3, 36, 520059225, time.Local), Count:4, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"kubelet", ReportingInstance:"24.144.92.253"}': 'events "24.144.92.253.17b226631cbe3767" is forbidden: User "system:anonymous" cannot patch resource "events" in API group "" in the namespace "default"' (will not retry!) Feb 9 09:03:36.522536 kubelet[1395]: E0209 09:03:36.522516 1395 kubelet_node_status.go:92] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="24.144.92.253" Feb 9 09:03:36.523647 kubelet[1395]: E0209 09:03:36.523569 1395 event.go:280] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"24.144.92.253.17b226631cbe4e7d", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"24.144.92.253", UID:"24.144.92.253", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasNoDiskPressure", Message:"Node 24.144.92.253 status is now: NodeHasNoDiskPressure", Source:v1.EventSource{Component:"kubelet", Host:"24.144.92.253"}, FirstTimestamp:time.Date(2024, time.February, 9, 9, 3, 35, 852691069, time.Local), LastTimestamp:time.Date(2024, time.February, 9, 9, 3, 36, 520425806, time.Local), Count:4, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"kubelet", ReportingInstance:"24.144.92.253"}': 'events "24.144.92.253.17b226631cbe4e7d" is forbidden: User "system:anonymous" cannot patch resource "events" in API group "" in the namespace "default"' (will not retry!) Feb 9 09:03:36.525201 kubelet[1395]: E0209 09:03:36.525104 1395 event.go:280] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"24.144.92.253.17b226631cbe5bd8", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"24.144.92.253", UID:"24.144.92.253", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientPID", Message:"Node 24.144.92.253 status is now: NodeHasSufficientPID", Source:v1.EventSource{Component:"kubelet", Host:"24.144.92.253"}, FirstTimestamp:time.Date(2024, time.February, 9, 9, 3, 35, 852694488, time.Local), LastTimestamp:time.Date(2024, time.February, 9, 9, 3, 36, 520441067, time.Local), Count:4, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"kubelet", ReportingInstance:"24.144.92.253"}': 'events "24.144.92.253.17b226631cbe5bd8" is forbidden: User "system:anonymous" cannot patch resource "events" in API group "" in the namespace "default"' (will not retry!) Feb 9 09:03:36.637635 kubelet[1395]: W0209 09:03:36.637592 1395 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.Node: nodes "24.144.92.253" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Feb 9 09:03:36.637897 kubelet[1395]: E0209 09:03:36.637876 1395 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.Node: failed to list *v1.Node: nodes "24.144.92.253" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Feb 9 09:03:36.765361 kubelet[1395]: W0209 09:03:36.765320 1395 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Feb 9 09:03:36.765628 kubelet[1395]: E0209 09:03:36.765608 1395 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Feb 9 09:03:36.774548 kubelet[1395]: I0209 09:03:36.774485 1395 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Feb 9 09:03:36.792862 kubelet[1395]: E0209 09:03:36.792811 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:03:37.173976 kubelet[1395]: E0209 09:03:37.173832 1395 csi_plugin.go:295] Failed to initialize CSINode: error updating CSINode annotation: timed out waiting for the condition; caused by: nodes "24.144.92.253" not found Feb 9 09:03:37.232083 kubelet[1395]: E0209 09:03:37.232038 1395 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"24.144.92.253\" not found" node="24.144.92.253" Feb 9 09:03:37.324645 kubelet[1395]: I0209 09:03:37.324609 1395 kubelet_node_status.go:70] "Attempting to register node" node="24.144.92.253" Feb 9 09:03:37.342802 kubelet[1395]: I0209 09:03:37.342751 1395 kubelet_node_status.go:73] "Successfully registered node" node="24.144.92.253" Feb 9 09:03:37.351037 kubelet[1395]: I0209 09:03:37.350984 1395 kuberuntime_manager.go:1463] "Updating runtime config through cri with podcidr" CIDR="192.168.1.0/24" Feb 9 09:03:37.352175 env[1108]: time="2024-02-09T09:03:37.351984253Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Feb 9 09:03:37.352652 kubelet[1395]: I0209 09:03:37.352289 1395 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.1.0/24" Feb 9 09:03:37.370219 kubelet[1395]: E0209 09:03:37.370167 1395 kubelet_node_status.go:458] "Error getting the current node from lister" err="node \"24.144.92.253\" not found" Feb 9 09:03:37.471585 kubelet[1395]: E0209 09:03:37.471404 1395 kubelet_node_status.go:458] "Error getting the current node from lister" err="node \"24.144.92.253\" not found" Feb 9 09:03:37.572271 kubelet[1395]: E0209 09:03:37.572224 1395 kubelet_node_status.go:458] "Error getting the current node from lister" err="node \"24.144.92.253\" not found" Feb 9 09:03:37.673029 kubelet[1395]: E0209 09:03:37.672975 1395 kubelet_node_status.go:458] "Error getting the current node from lister" err="node \"24.144.92.253\" not found" Feb 9 09:03:37.712647 sudo[1208]: pam_unix(sudo:session): session closed for user root Feb 9 09:03:37.717804 sshd[1205]: pam_unix(sshd:session): session closed for user core Feb 9 09:03:37.722017 systemd[1]: sshd@4-24.144.92.253:22-139.178.89.65:34000.service: Deactivated successfully. Feb 9 09:03:37.723320 systemd[1]: session-5.scope: Deactivated successfully. Feb 9 09:03:37.724522 systemd-logind[1096]: Session 5 logged out. Waiting for processes to exit. Feb 9 09:03:37.726105 systemd-logind[1096]: Removed session 5. Feb 9 09:03:37.774325 kubelet[1395]: E0209 09:03:37.774217 1395 kubelet_node_status.go:458] "Error getting the current node from lister" err="node \"24.144.92.253\" not found" Feb 9 09:03:37.793709 kubelet[1395]: E0209 09:03:37.793588 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:03:37.874531 kubelet[1395]: E0209 09:03:37.874387 1395 kubelet_node_status.go:458] "Error getting the current node from lister" err="node \"24.144.92.253\" not found" Feb 9 09:03:37.975687 kubelet[1395]: E0209 09:03:37.974901 1395 kubelet_node_status.go:458] "Error getting the current node from lister" err="node \"24.144.92.253\" not found" Feb 9 09:03:38.075767 kubelet[1395]: E0209 09:03:38.075700 1395 kubelet_node_status.go:458] "Error getting the current node from lister" err="node \"24.144.92.253\" not found" Feb 9 09:03:38.176599 kubelet[1395]: E0209 09:03:38.176551 1395 kubelet_node_status.go:458] "Error getting the current node from lister" err="node \"24.144.92.253\" not found" Feb 9 09:03:38.277551 kubelet[1395]: E0209 09:03:38.277504 1395 kubelet_node_status.go:458] "Error getting the current node from lister" err="node \"24.144.92.253\" not found" Feb 9 09:03:38.793928 kubelet[1395]: E0209 09:03:38.793891 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:03:38.794372 kubelet[1395]: I0209 09:03:38.794350 1395 apiserver.go:52] "Watching apiserver" Feb 9 09:03:38.797961 kubelet[1395]: I0209 09:03:38.797924 1395 topology_manager.go:215] "Topology Admit Handler" podUID="d05f098b-a09d-4a49-8adf-e1321bc177c6" podNamespace="kube-system" podName="cilium-mwqzv" Feb 9 09:03:38.798106 kubelet[1395]: I0209 09:03:38.798063 1395 topology_manager.go:215] "Topology Admit Handler" podUID="198088a3-bb8d-4902-8081-529a415c0dab" podNamespace="kube-system" podName="kube-proxy-n2hp7" Feb 9 09:03:38.804958 systemd[1]: Created slice kubepods-besteffort-pod198088a3_bb8d_4902_8081_529a415c0dab.slice. Feb 9 09:03:38.818441 kubelet[1395]: I0209 09:03:38.818414 1395 desired_state_of_world_populator.go:159] "Finished populating initial desired state of world" Feb 9 09:03:38.828092 kubelet[1395]: I0209 09:03:38.828051 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-lib-modules\") pod \"cilium-mwqzv\" (UID: \"d05f098b-a09d-4a49-8adf-e1321bc177c6\") " pod="kube-system/cilium-mwqzv" Feb 9 09:03:38.828367 kubelet[1395]: I0209 09:03:38.828346 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/d05f098b-a09d-4a49-8adf-e1321bc177c6-clustermesh-secrets\") pod \"cilium-mwqzv\" (UID: \"d05f098b-a09d-4a49-8adf-e1321bc177c6\") " pod="kube-system/cilium-mwqzv" Feb 9 09:03:38.828546 kubelet[1395]: I0209 09:03:38.828530 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-host-proc-sys-net\") pod \"cilium-mwqzv\" (UID: \"d05f098b-a09d-4a49-8adf-e1321bc177c6\") " pod="kube-system/cilium-mwqzv" Feb 9 09:03:38.828686 kubelet[1395]: I0209 09:03:38.828671 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-cilium-cgroup\") pod \"cilium-mwqzv\" (UID: \"d05f098b-a09d-4a49-8adf-e1321bc177c6\") " pod="kube-system/cilium-mwqzv" Feb 9 09:03:38.828819 kubelet[1395]: I0209 09:03:38.828805 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-xtables-lock\") pod \"cilium-mwqzv\" (UID: \"d05f098b-a09d-4a49-8adf-e1321bc177c6\") " pod="kube-system/cilium-mwqzv" Feb 9 09:03:38.828969 kubelet[1395]: I0209 09:03:38.828954 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/d05f098b-a09d-4a49-8adf-e1321bc177c6-cilium-config-path\") pod \"cilium-mwqzv\" (UID: \"d05f098b-a09d-4a49-8adf-e1321bc177c6\") " pod="kube-system/cilium-mwqzv" Feb 9 09:03:38.829119 kubelet[1395]: I0209 09:03:38.829101 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djcfj\" (UniqueName: \"kubernetes.io/projected/d05f098b-a09d-4a49-8adf-e1321bc177c6-kube-api-access-djcfj\") pod \"cilium-mwqzv\" (UID: \"d05f098b-a09d-4a49-8adf-e1321bc177c6\") " pod="kube-system/cilium-mwqzv" Feb 9 09:03:38.829260 kubelet[1395]: I0209 09:03:38.829244 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/198088a3-bb8d-4902-8081-529a415c0dab-xtables-lock\") pod \"kube-proxy-n2hp7\" (UID: \"198088a3-bb8d-4902-8081-529a415c0dab\") " pod="kube-system/kube-proxy-n2hp7" Feb 9 09:03:38.829397 kubelet[1395]: I0209 09:03:38.829382 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rgcm\" (UniqueName: \"kubernetes.io/projected/198088a3-bb8d-4902-8081-529a415c0dab-kube-api-access-8rgcm\") pod \"kube-proxy-n2hp7\" (UID: \"198088a3-bb8d-4902-8081-529a415c0dab\") " pod="kube-system/kube-proxy-n2hp7" Feb 9 09:03:38.829574 kubelet[1395]: I0209 09:03:38.829557 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-cilium-run\") pod \"cilium-mwqzv\" (UID: \"d05f098b-a09d-4a49-8adf-e1321bc177c6\") " pod="kube-system/cilium-mwqzv" Feb 9 09:03:38.829788 kubelet[1395]: I0209 09:03:38.829773 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-bpf-maps\") pod \"cilium-mwqzv\" (UID: \"d05f098b-a09d-4a49-8adf-e1321bc177c6\") " pod="kube-system/cilium-mwqzv" Feb 9 09:03:38.829939 kubelet[1395]: I0209 09:03:38.829924 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-cni-path\") pod \"cilium-mwqzv\" (UID: \"d05f098b-a09d-4a49-8adf-e1321bc177c6\") " pod="kube-system/cilium-mwqzv" Feb 9 09:03:38.830104 kubelet[1395]: I0209 09:03:38.830086 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/d05f098b-a09d-4a49-8adf-e1321bc177c6-hubble-tls\") pod \"cilium-mwqzv\" (UID: \"d05f098b-a09d-4a49-8adf-e1321bc177c6\") " pod="kube-system/cilium-mwqzv" Feb 9 09:03:38.830249 kubelet[1395]: I0209 09:03:38.830232 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/198088a3-bb8d-4902-8081-529a415c0dab-kube-proxy\") pod \"kube-proxy-n2hp7\" (UID: \"198088a3-bb8d-4902-8081-529a415c0dab\") " pod="kube-system/kube-proxy-n2hp7" Feb 9 09:03:38.830394 kubelet[1395]: I0209 09:03:38.830379 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/198088a3-bb8d-4902-8081-529a415c0dab-lib-modules\") pod \"kube-proxy-n2hp7\" (UID: \"198088a3-bb8d-4902-8081-529a415c0dab\") " pod="kube-system/kube-proxy-n2hp7" Feb 9 09:03:38.830562 kubelet[1395]: I0209 09:03:38.830543 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-hostproc\") pod \"cilium-mwqzv\" (UID: \"d05f098b-a09d-4a49-8adf-e1321bc177c6\") " pod="kube-system/cilium-mwqzv" Feb 9 09:03:38.830716 kubelet[1395]: I0209 09:03:38.830700 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-etc-cni-netd\") pod \"cilium-mwqzv\" (UID: \"d05f098b-a09d-4a49-8adf-e1321bc177c6\") " pod="kube-system/cilium-mwqzv" Feb 9 09:03:38.831391 kubelet[1395]: I0209 09:03:38.831364 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-host-proc-sys-kernel\") pod \"cilium-mwqzv\" (UID: \"d05f098b-a09d-4a49-8adf-e1321bc177c6\") " pod="kube-system/cilium-mwqzv" Feb 9 09:03:38.835864 systemd[1]: Created slice kubepods-burstable-podd05f098b_a09d_4a49_8adf_e1321bc177c6.slice. Feb 9 09:03:39.134518 kubelet[1395]: E0209 09:03:39.133428 1395 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Feb 9 09:03:39.135344 env[1108]: time="2024-02-09T09:03:39.135290839Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-n2hp7,Uid:198088a3-bb8d-4902-8081-529a415c0dab,Namespace:kube-system,Attempt:0,}" Feb 9 09:03:39.144386 kubelet[1395]: E0209 09:03:39.144316 1395 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Feb 9 09:03:39.146310 env[1108]: time="2024-02-09T09:03:39.145734308Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-mwqzv,Uid:d05f098b-a09d-4a49-8adf-e1321bc177c6,Namespace:kube-system,Attempt:0,}" Feb 9 09:03:39.729436 env[1108]: time="2024-02-09T09:03:39.729366936Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 09:03:39.731465 env[1108]: time="2024-02-09T09:03:39.731419240Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 09:03:39.735088 env[1108]: time="2024-02-09T09:03:39.735036020Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:6270bb605e12e581514ada5fd5b3216f727db55dc87d5889c790e4c760683fee,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 09:03:39.736276 env[1108]: time="2024-02-09T09:03:39.736234692Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:6270bb605e12e581514ada5fd5b3216f727db55dc87d5889c790e4c760683fee,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 09:03:39.747608 env[1108]: time="2024-02-09T09:03:39.747554870Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 09:03:39.750705 env[1108]: time="2024-02-09T09:03:39.750662399Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 09:03:39.751368 env[1108]: time="2024-02-09T09:03:39.751327592Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 09:03:39.751973 env[1108]: time="2024-02-09T09:03:39.751930070Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 09:03:39.783355 env[1108]: time="2024-02-09T09:03:39.779431237Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 9 09:03:39.783355 env[1108]: time="2024-02-09T09:03:39.779469494Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 9 09:03:39.783355 env[1108]: time="2024-02-09T09:03:39.779480564Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 9 09:03:39.783355 env[1108]: time="2024-02-09T09:03:39.779666630Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/12b0726dbea16a1b5a212140b2992b123b82bc45a0e3a27f90285e5163c7b086 pid=1455 runtime=io.containerd.runc.v2 Feb 9 09:03:39.783953 env[1108]: time="2024-02-09T09:03:39.776457018Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 9 09:03:39.783953 env[1108]: time="2024-02-09T09:03:39.776513419Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 9 09:03:39.783953 env[1108]: time="2024-02-09T09:03:39.776529124Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 9 09:03:39.783953 env[1108]: time="2024-02-09T09:03:39.777021126Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/3f55ee1ffb7b986f1ae81b075696d310d94fa86e0069101af0741e798b3bc5e4 pid=1457 runtime=io.containerd.runc.v2 Feb 9 09:03:39.794724 kubelet[1395]: E0209 09:03:39.794674 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:03:39.804138 systemd[1]: Started cri-containerd-3f55ee1ffb7b986f1ae81b075696d310d94fa86e0069101af0741e798b3bc5e4.scope. Feb 9 09:03:39.814038 systemd[1]: Started cri-containerd-12b0726dbea16a1b5a212140b2992b123b82bc45a0e3a27f90285e5163c7b086.scope. Feb 9 09:03:39.863831 env[1108]: time="2024-02-09T09:03:39.863784643Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-mwqzv,Uid:d05f098b-a09d-4a49-8adf-e1321bc177c6,Namespace:kube-system,Attempt:0,} returns sandbox id \"3f55ee1ffb7b986f1ae81b075696d310d94fa86e0069101af0741e798b3bc5e4\"" Feb 9 09:03:39.865058 kubelet[1395]: E0209 09:03:39.865029 1395 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Feb 9 09:03:39.866709 env[1108]: time="2024-02-09T09:03:39.866670238Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Feb 9 09:03:39.873948 env[1108]: time="2024-02-09T09:03:39.873907865Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-n2hp7,Uid:198088a3-bb8d-4902-8081-529a415c0dab,Namespace:kube-system,Attempt:0,} returns sandbox id \"12b0726dbea16a1b5a212140b2992b123b82bc45a0e3a27f90285e5163c7b086\"" Feb 9 09:03:39.875002 kubelet[1395]: E0209 09:03:39.874958 1395 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Feb 9 09:03:39.945709 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount136371685.mount: Deactivated successfully. Feb 9 09:03:40.795329 kubelet[1395]: E0209 09:03:40.795285 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:03:41.796732 kubelet[1395]: E0209 09:03:41.796677 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:03:42.797832 kubelet[1395]: E0209 09:03:42.797783 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:03:43.798624 kubelet[1395]: E0209 09:03:43.798571 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:03:44.799303 kubelet[1395]: E0209 09:03:44.799260 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:03:45.724275 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1180955109.mount: Deactivated successfully. Feb 9 09:03:45.800010 kubelet[1395]: E0209 09:03:45.799935 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:03:46.800745 kubelet[1395]: E0209 09:03:46.800692 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:03:47.800975 kubelet[1395]: E0209 09:03:47.800865 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:03:48.802003 kubelet[1395]: E0209 09:03:48.801938 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:03:49.323652 env[1108]: time="2024-02-09T09:03:49.323579208Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 09:03:49.327694 env[1108]: time="2024-02-09T09:03:49.327629970Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 09:03:49.329908 env[1108]: time="2024-02-09T09:03:49.329855752Z" level=info msg="ImageUpdate event &ImageUpdate{Name:quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 09:03:49.331335 env[1108]: time="2024-02-09T09:03:49.331284425Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"" Feb 9 09:03:49.333269 env[1108]: time="2024-02-09T09:03:49.333222383Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.28.6\"" Feb 9 09:03:49.335291 env[1108]: time="2024-02-09T09:03:49.335257677Z" level=info msg="CreateContainer within sandbox \"3f55ee1ffb7b986f1ae81b075696d310d94fa86e0069101af0741e798b3bc5e4\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Feb 9 09:03:49.349322 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3707162159.mount: Deactivated successfully. Feb 9 09:03:49.355965 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2861189391.mount: Deactivated successfully. Feb 9 09:03:49.365076 env[1108]: time="2024-02-09T09:03:49.365013801Z" level=info msg="CreateContainer within sandbox \"3f55ee1ffb7b986f1ae81b075696d310d94fa86e0069101af0741e798b3bc5e4\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"fa22ea78685b767657fec5a8225818e4e3491ed5dfcac1158214eb78bac1e67d\"" Feb 9 09:03:49.366111 env[1108]: time="2024-02-09T09:03:49.366066668Z" level=info msg="StartContainer for \"fa22ea78685b767657fec5a8225818e4e3491ed5dfcac1158214eb78bac1e67d\"" Feb 9 09:03:49.398473 systemd[1]: Started cri-containerd-fa22ea78685b767657fec5a8225818e4e3491ed5dfcac1158214eb78bac1e67d.scope. Feb 9 09:03:49.468847 env[1108]: time="2024-02-09T09:03:49.468778090Z" level=info msg="StartContainer for \"fa22ea78685b767657fec5a8225818e4e3491ed5dfcac1158214eb78bac1e67d\" returns successfully" Feb 9 09:03:49.478969 systemd[1]: cri-containerd-fa22ea78685b767657fec5a8225818e4e3491ed5dfcac1158214eb78bac1e67d.scope: Deactivated successfully. Feb 9 09:03:49.577993 env[1108]: time="2024-02-09T09:03:49.576794692Z" level=info msg="shim disconnected" id=fa22ea78685b767657fec5a8225818e4e3491ed5dfcac1158214eb78bac1e67d Feb 9 09:03:49.577993 env[1108]: time="2024-02-09T09:03:49.576852308Z" level=warning msg="cleaning up after shim disconnected" id=fa22ea78685b767657fec5a8225818e4e3491ed5dfcac1158214eb78bac1e67d namespace=k8s.io Feb 9 09:03:49.577993 env[1108]: time="2024-02-09T09:03:49.576864912Z" level=info msg="cleaning up dead shim" Feb 9 09:03:49.591957 env[1108]: time="2024-02-09T09:03:49.591888805Z" level=warning msg="cleanup warnings time=\"2024-02-09T09:03:49Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=1572 runtime=io.containerd.runc.v2\n" Feb 9 09:03:49.803011 kubelet[1395]: E0209 09:03:49.802933 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:03:49.966904 kubelet[1395]: E0209 09:03:49.966352 1395 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Feb 9 09:03:49.980699 env[1108]: time="2024-02-09T09:03:49.980628866Z" level=info msg="CreateContainer within sandbox \"3f55ee1ffb7b986f1ae81b075696d310d94fa86e0069101af0741e798b3bc5e4\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Feb 9 09:03:50.007096 env[1108]: time="2024-02-09T09:03:50.006982728Z" level=info msg="CreateContainer within sandbox \"3f55ee1ffb7b986f1ae81b075696d310d94fa86e0069101af0741e798b3bc5e4\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"e19abab8f3746b464980f337f5669da3391695f31e285c11842661765089f999\"" Feb 9 09:03:50.008307 env[1108]: time="2024-02-09T09:03:50.008255671Z" level=info msg="StartContainer for \"e19abab8f3746b464980f337f5669da3391695f31e285c11842661765089f999\"" Feb 9 09:03:50.043820 systemd[1]: Started cri-containerd-e19abab8f3746b464980f337f5669da3391695f31e285c11842661765089f999.scope. Feb 9 09:03:50.142228 systemd[1]: systemd-sysctl.service: Deactivated successfully. Feb 9 09:03:50.142462 systemd[1]: Stopped systemd-sysctl.service. Feb 9 09:03:50.142880 systemd[1]: Stopping systemd-sysctl.service... Feb 9 09:03:50.144885 systemd[1]: Starting systemd-sysctl.service... Feb 9 09:03:50.145262 systemd[1]: cri-containerd-e19abab8f3746b464980f337f5669da3391695f31e285c11842661765089f999.scope: Deactivated successfully. Feb 9 09:03:50.158864 systemd[1]: Finished systemd-sysctl.service. Feb 9 09:03:50.159671 env[1108]: time="2024-02-09T09:03:50.158376529Z" level=info msg="StartContainer for \"e19abab8f3746b464980f337f5669da3391695f31e285c11842661765089f999\" returns successfully" Feb 9 09:03:50.230239 env[1108]: time="2024-02-09T09:03:50.229300207Z" level=info msg="shim disconnected" id=e19abab8f3746b464980f337f5669da3391695f31e285c11842661765089f999 Feb 9 09:03:50.230239 env[1108]: time="2024-02-09T09:03:50.229371555Z" level=warning msg="cleaning up after shim disconnected" id=e19abab8f3746b464980f337f5669da3391695f31e285c11842661765089f999 namespace=k8s.io Feb 9 09:03:50.230239 env[1108]: time="2024-02-09T09:03:50.229383432Z" level=info msg="cleaning up dead shim" Feb 9 09:03:50.249217 env[1108]: time="2024-02-09T09:03:50.249163338Z" level=warning msg="cleanup warnings time=\"2024-02-09T09:03:50Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=1637 runtime=io.containerd.runc.v2\n" Feb 9 09:03:50.347254 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-fa22ea78685b767657fec5a8225818e4e3491ed5dfcac1158214eb78bac1e67d-rootfs.mount: Deactivated successfully. Feb 9 09:03:50.534333 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3273961761.mount: Deactivated successfully. Feb 9 09:03:50.803587 kubelet[1395]: E0209 09:03:50.803405 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:03:50.969397 kubelet[1395]: E0209 09:03:50.969367 1395 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Feb 9 09:03:50.971947 env[1108]: time="2024-02-09T09:03:50.971898451Z" level=info msg="CreateContainer within sandbox \"3f55ee1ffb7b986f1ae81b075696d310d94fa86e0069101af0741e798b3bc5e4\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Feb 9 09:03:50.994108 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount838571718.mount: Deactivated successfully. Feb 9 09:03:51.008128 env[1108]: time="2024-02-09T09:03:51.008064447Z" level=info msg="CreateContainer within sandbox \"3f55ee1ffb7b986f1ae81b075696d310d94fa86e0069101af0741e798b3bc5e4\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"82189475b3479daae734762dd9373163de30c083a7ed75ee7a8f2e9f15db74b0\"" Feb 9 09:03:51.009517 env[1108]: time="2024-02-09T09:03:51.009440020Z" level=info msg="StartContainer for \"82189475b3479daae734762dd9373163de30c083a7ed75ee7a8f2e9f15db74b0\"" Feb 9 09:03:51.043749 systemd[1]: Started cri-containerd-82189475b3479daae734762dd9373163de30c083a7ed75ee7a8f2e9f15db74b0.scope. Feb 9 09:03:51.105204 systemd[1]: cri-containerd-82189475b3479daae734762dd9373163de30c083a7ed75ee7a8f2e9f15db74b0.scope: Deactivated successfully. Feb 9 09:03:51.107898 env[1108]: time="2024-02-09T09:03:51.107844857Z" level=info msg="StartContainer for \"82189475b3479daae734762dd9373163de30c083a7ed75ee7a8f2e9f15db74b0\" returns successfully" Feb 9 09:03:51.196013 env[1108]: time="2024-02-09T09:03:51.195941570Z" level=info msg="shim disconnected" id=82189475b3479daae734762dd9373163de30c083a7ed75ee7a8f2e9f15db74b0 Feb 9 09:03:51.196013 env[1108]: time="2024-02-09T09:03:51.196002262Z" level=warning msg="cleaning up after shim disconnected" id=82189475b3479daae734762dd9373163de30c083a7ed75ee7a8f2e9f15db74b0 namespace=k8s.io Feb 9 09:03:51.196013 env[1108]: time="2024-02-09T09:03:51.196015817Z" level=info msg="cleaning up dead shim" Feb 9 09:03:51.231628 env[1108]: time="2024-02-09T09:03:51.231562535Z" level=warning msg="cleanup warnings time=\"2024-02-09T09:03:51Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=1699 runtime=io.containerd.runc.v2\n" Feb 9 09:03:51.346062 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1196319824.mount: Deactivated successfully. Feb 9 09:03:51.376703 env[1108]: time="2024-02-09T09:03:51.376526022Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy:v1.28.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 09:03:51.379568 env[1108]: time="2024-02-09T09:03:51.379518689Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:342a759d88156b4f56ba522a1aed0e3d32d72542545346b40877f6583bebe05f,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 09:03:51.381951 env[1108]: time="2024-02-09T09:03:51.381891128Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-proxy:v1.28.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 09:03:51.385264 env[1108]: time="2024-02-09T09:03:51.385224886Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy@sha256:3898a1671ae42be1cd3c2e777549bc7b5b306b8da3a224b747365f6679fb902a,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 09:03:51.386143 env[1108]: time="2024-02-09T09:03:51.386110565Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.28.6\" returns image reference \"sha256:342a759d88156b4f56ba522a1aed0e3d32d72542545346b40877f6583bebe05f\"" Feb 9 09:03:51.388714 env[1108]: time="2024-02-09T09:03:51.388676971Z" level=info msg="CreateContainer within sandbox \"12b0726dbea16a1b5a212140b2992b123b82bc45a0e3a27f90285e5163c7b086\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Feb 9 09:03:51.404900 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4275165014.mount: Deactivated successfully. Feb 9 09:03:51.409741 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1284064424.mount: Deactivated successfully. Feb 9 09:03:51.417148 env[1108]: time="2024-02-09T09:03:51.417072472Z" level=info msg="CreateContainer within sandbox \"12b0726dbea16a1b5a212140b2992b123b82bc45a0e3a27f90285e5163c7b086\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"067f47b4b243f87594a64ac42f58dbf9cc7f33ae558066db5587c4aacfe9843a\"" Feb 9 09:03:51.417925 env[1108]: time="2024-02-09T09:03:51.417889324Z" level=info msg="StartContainer for \"067f47b4b243f87594a64ac42f58dbf9cc7f33ae558066db5587c4aacfe9843a\"" Feb 9 09:03:51.444364 systemd[1]: Started cri-containerd-067f47b4b243f87594a64ac42f58dbf9cc7f33ae558066db5587c4aacfe9843a.scope. Feb 9 09:03:51.506363 env[1108]: time="2024-02-09T09:03:51.506302744Z" level=info msg="StartContainer for \"067f47b4b243f87594a64ac42f58dbf9cc7f33ae558066db5587c4aacfe9843a\" returns successfully" Feb 9 09:03:51.804395 kubelet[1395]: E0209 09:03:51.804327 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:03:51.973655 kubelet[1395]: E0209 09:03:51.973625 1395 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Feb 9 09:03:51.976304 kubelet[1395]: E0209 09:03:51.976275 1395 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Feb 9 09:03:51.976856 env[1108]: time="2024-02-09T09:03:51.976803489Z" level=info msg="CreateContainer within sandbox \"3f55ee1ffb7b986f1ae81b075696d310d94fa86e0069101af0741e798b3bc5e4\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Feb 9 09:03:51.995334 env[1108]: time="2024-02-09T09:03:51.995274584Z" level=info msg="CreateContainer within sandbox \"3f55ee1ffb7b986f1ae81b075696d310d94fa86e0069101af0741e798b3bc5e4\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"ae2f0999c1a3f2ff36597066153f98e00c426f1f85f40515b293ae9338d9c5ea\"" Feb 9 09:03:51.996701 env[1108]: time="2024-02-09T09:03:51.996654902Z" level=info msg="StartContainer for \"ae2f0999c1a3f2ff36597066153f98e00c426f1f85f40515b293ae9338d9c5ea\"" Feb 9 09:03:52.017919 kubelet[1395]: I0209 09:03:52.017860 1395 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-proxy-n2hp7" podStartSLOduration=3.506861688 podCreationTimestamp="2024-02-09 09:03:37 +0000 UTC" firstStartedPulling="2024-02-09 09:03:39.875597725 +0000 UTC m=+4.688999869" lastFinishedPulling="2024-02-09 09:03:51.386549407 +0000 UTC m=+16.199951555" observedRunningTime="2024-02-09 09:03:52.017587053 +0000 UTC m=+16.830989218" watchObservedRunningTime="2024-02-09 09:03:52.017813374 +0000 UTC m=+16.831215532" Feb 9 09:03:52.023884 systemd[1]: Started cri-containerd-ae2f0999c1a3f2ff36597066153f98e00c426f1f85f40515b293ae9338d9c5ea.scope. Feb 9 09:03:52.073148 systemd[1]: cri-containerd-ae2f0999c1a3f2ff36597066153f98e00c426f1f85f40515b293ae9338d9c5ea.scope: Deactivated successfully. Feb 9 09:03:52.078951 env[1108]: time="2024-02-09T09:03:52.078901922Z" level=info msg="StartContainer for \"ae2f0999c1a3f2ff36597066153f98e00c426f1f85f40515b293ae9338d9c5ea\" returns successfully" Feb 9 09:03:52.080737 env[1108]: time="2024-02-09T09:03:52.080646595Z" level=warning msg="error from *cgroupsv2.Manager.EventChan" error="failed to add inotify watch for \"/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd05f098b_a09d_4a49_8adf_e1321bc177c6.slice/cri-containerd-ae2f0999c1a3f2ff36597066153f98e00c426f1f85f40515b293ae9338d9c5ea.scope/memory.events\": no such file or directory" Feb 9 09:03:52.118101 env[1108]: time="2024-02-09T09:03:52.118020963Z" level=info msg="shim disconnected" id=ae2f0999c1a3f2ff36597066153f98e00c426f1f85f40515b293ae9338d9c5ea Feb 9 09:03:52.118101 env[1108]: time="2024-02-09T09:03:52.118094245Z" level=warning msg="cleaning up after shim disconnected" id=ae2f0999c1a3f2ff36597066153f98e00c426f1f85f40515b293ae9338d9c5ea namespace=k8s.io Feb 9 09:03:52.118101 env[1108]: time="2024-02-09T09:03:52.118110014Z" level=info msg="cleaning up dead shim" Feb 9 09:03:52.133161 env[1108]: time="2024-02-09T09:03:52.133096639Z" level=warning msg="cleanup warnings time=\"2024-02-09T09:03:52Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=1917 runtime=io.containerd.runc.v2\n" Feb 9 09:03:52.804834 kubelet[1395]: E0209 09:03:52.804773 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:03:52.982202 kubelet[1395]: E0209 09:03:52.981729 1395 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Feb 9 09:03:52.982202 kubelet[1395]: E0209 09:03:52.981915 1395 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Feb 9 09:03:52.985265 env[1108]: time="2024-02-09T09:03:52.985205866Z" level=info msg="CreateContainer within sandbox \"3f55ee1ffb7b986f1ae81b075696d310d94fa86e0069101af0741e798b3bc5e4\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Feb 9 09:03:53.014281 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1182601343.mount: Deactivated successfully. Feb 9 09:03:53.026705 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1795238280.mount: Deactivated successfully. Feb 9 09:03:53.031127 env[1108]: time="2024-02-09T09:03:53.030970991Z" level=info msg="CreateContainer within sandbox \"3f55ee1ffb7b986f1ae81b075696d310d94fa86e0069101af0741e798b3bc5e4\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"3054d463ceb9b9e2d733ac777f55d5bc633d9a45b71193841f8ad8b94161254a\"" Feb 9 09:03:53.032221 env[1108]: time="2024-02-09T09:03:53.032169027Z" level=info msg="StartContainer for \"3054d463ceb9b9e2d733ac777f55d5bc633d9a45b71193841f8ad8b94161254a\"" Feb 9 09:03:53.056569 systemd[1]: Started cri-containerd-3054d463ceb9b9e2d733ac777f55d5bc633d9a45b71193841f8ad8b94161254a.scope. Feb 9 09:03:53.115005 env[1108]: time="2024-02-09T09:03:53.114936401Z" level=info msg="StartContainer for \"3054d463ceb9b9e2d733ac777f55d5bc633d9a45b71193841f8ad8b94161254a\" returns successfully" Feb 9 09:03:53.285309 kubelet[1395]: I0209 09:03:53.285092 1395 kubelet_node_status.go:493] "Fast updating node status as it just became ready" Feb 9 09:03:53.580539 kernel: Initializing XFRM netlink socket Feb 9 09:03:53.805305 kubelet[1395]: E0209 09:03:53.805232 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:03:53.988114 kubelet[1395]: E0209 09:03:53.987989 1395 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Feb 9 09:03:54.132115 kubelet[1395]: I0209 09:03:54.132038 1395 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/cilium-mwqzv" podStartSLOduration=7.665956825 podCreationTimestamp="2024-02-09 09:03:37 +0000 UTC" firstStartedPulling="2024-02-09 09:03:39.866160683 +0000 UTC m=+4.679562828" lastFinishedPulling="2024-02-09 09:03:49.332175857 +0000 UTC m=+14.145578002" observedRunningTime="2024-02-09 09:03:54.130573685 +0000 UTC m=+18.943975852" watchObservedRunningTime="2024-02-09 09:03:54.131971999 +0000 UTC m=+18.945374157" Feb 9 09:03:54.806020 kubelet[1395]: E0209 09:03:54.805926 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:03:54.876360 kubelet[1395]: I0209 09:03:54.875686 1395 topology_manager.go:215] "Topology Admit Handler" podUID="d12a828f-0cd6-4d43-a1e9-83bf31ec6616" podNamespace="default" podName="nginx-deployment-6d5f899847-xw5zj" Feb 9 09:03:54.881267 systemd[1]: Created slice kubepods-besteffort-podd12a828f_0cd6_4d43_a1e9_83bf31ec6616.slice. Feb 9 09:03:54.981474 kubelet[1395]: I0209 09:03:54.981357 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bspn\" (UniqueName: \"kubernetes.io/projected/d12a828f-0cd6-4d43-a1e9-83bf31ec6616-kube-api-access-5bspn\") pod \"nginx-deployment-6d5f899847-xw5zj\" (UID: \"d12a828f-0cd6-4d43-a1e9-83bf31ec6616\") " pod="default/nginx-deployment-6d5f899847-xw5zj" Feb 9 09:03:54.990589 kubelet[1395]: E0209 09:03:54.990547 1395 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Feb 9 09:03:55.187597 env[1108]: time="2024-02-09T09:03:55.187448203Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-6d5f899847-xw5zj,Uid:d12a828f-0cd6-4d43-a1e9-83bf31ec6616,Namespace:default,Attempt:0,}" Feb 9 09:03:55.289120 systemd-networkd[1002]: cilium_host: Link UP Feb 9 09:03:55.289527 systemd-networkd[1002]: cilium_net: Link UP Feb 9 09:03:55.293758 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cilium_net: link becomes ready Feb 9 09:03:55.293899 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cilium_host: link becomes ready Feb 9 09:03:55.297886 systemd-networkd[1002]: cilium_net: Gained carrier Feb 9 09:03:55.298621 systemd-networkd[1002]: cilium_host: Gained carrier Feb 9 09:03:55.457371 systemd-networkd[1002]: cilium_vxlan: Link UP Feb 9 09:03:55.457380 systemd-networkd[1002]: cilium_vxlan: Gained carrier Feb 9 09:03:55.583702 systemd-networkd[1002]: cilium_host: Gained IPv6LL Feb 9 09:03:55.626805 systemd[1]: Started sshd@5-24.144.92.253:22-85.209.11.27:9618.service. Feb 9 09:03:55.726546 kernel: NET: Registered PF_ALG protocol family Feb 9 09:03:55.784183 systemd-networkd[1002]: cilium_net: Gained IPv6LL Feb 9 09:03:55.791763 kubelet[1395]: E0209 09:03:55.791687 1395 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:03:55.807139 kubelet[1395]: E0209 09:03:55.806958 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:03:55.992093 kubelet[1395]: E0209 09:03:55.992050 1395 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Feb 9 09:03:56.560404 systemd-networkd[1002]: lxc_health: Link UP Feb 9 09:03:56.560780 systemd-networkd[1002]: lxc_health: Gained carrier Feb 9 09:03:56.561590 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc_health: link becomes ready Feb 9 09:03:56.744299 systemd-networkd[1002]: lxc7e8f9f2f576d: Link UP Feb 9 09:03:56.754624 kernel: eth0: renamed from tmp9d714 Feb 9 09:03:56.763321 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc7e8f9f2f576d: link becomes ready Feb 9 09:03:56.762967 systemd-networkd[1002]: lxc7e8f9f2f576d: Gained carrier Feb 9 09:03:56.807288 kubelet[1395]: E0209 09:03:56.807217 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:03:57.032610 systemd-networkd[1002]: cilium_vxlan: Gained IPv6LL Feb 9 09:03:57.147129 kubelet[1395]: E0209 09:03:57.147084 1395 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Feb 9 09:03:57.299961 systemd[1]: Started sshd@6-24.144.92.253:22-159.223.209.81:44186.service. Feb 9 09:03:57.719673 systemd-networkd[1002]: lxc_health: Gained IPv6LL Feb 9 09:03:57.807889 kubelet[1395]: E0209 09:03:57.807824 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:03:58.027792 sshd[2419]: Invalid user minima from 159.223.209.81 port 44186 Feb 9 09:03:58.208859 sshd[2419]: pam_faillock(sshd:auth): User unknown Feb 9 09:03:58.209562 sshd[2419]: pam_unix(sshd:auth): check pass; user unknown Feb 9 09:03:58.209613 sshd[2419]: pam_unix(sshd:auth): authentication failure; logname= uid=0 euid=0 tty=ssh ruser= rhost=159.223.209.81 Feb 9 09:03:58.210194 sshd[2419]: pam_faillock(sshd:auth): User unknown Feb 9 09:03:58.359850 systemd-networkd[1002]: lxc7e8f9f2f576d: Gained IPv6LL Feb 9 09:03:58.808878 kubelet[1395]: E0209 09:03:58.808816 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:03:59.593293 sshd[2156]: pam_unix(sshd:auth): authentication failure; logname= uid=0 euid=0 tty=ssh ruser= rhost=85.209.11.27 user=root Feb 9 09:03:59.809908 kubelet[1395]: E0209 09:03:59.809861 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:03:59.947521 sshd[2419]: Failed password for invalid user minima from 159.223.209.81 port 44186 ssh2 Feb 9 09:04:00.565619 sshd[2419]: Connection closed by invalid user minima 159.223.209.81 port 44186 [preauth] Feb 9 09:04:00.566044 systemd[1]: sshd@6-24.144.92.253:22-159.223.209.81:44186.service: Deactivated successfully. Feb 9 09:04:00.811565 kubelet[1395]: E0209 09:04:00.811513 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:01.796341 env[1108]: time="2024-02-09T09:04:01.796241156Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 9 09:04:01.796875 env[1108]: time="2024-02-09T09:04:01.796832022Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 9 09:04:01.796995 env[1108]: time="2024-02-09T09:04:01.796969409Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 9 09:04:01.797279 env[1108]: time="2024-02-09T09:04:01.797244474Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/9d714792f4a51a49152bbea93d1e249d0857d62fbddb06be19f4c3768134b72f pid=2447 runtime=io.containerd.runc.v2 Feb 9 09:04:01.803826 sshd[2156]: Failed password for root from 85.209.11.27 port 9618 ssh2 Feb 9 09:04:01.812589 kubelet[1395]: E0209 09:04:01.812500 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:01.843677 systemd[1]: run-containerd-runc-k8s.io-9d714792f4a51a49152bbea93d1e249d0857d62fbddb06be19f4c3768134b72f-runc.8Xa5GE.mount: Deactivated successfully. Feb 9 09:04:01.856883 systemd[1]: Started cri-containerd-9d714792f4a51a49152bbea93d1e249d0857d62fbddb06be19f4c3768134b72f.scope. Feb 9 09:04:01.917704 env[1108]: time="2024-02-09T09:04:01.917635907Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-6d5f899847-xw5zj,Uid:d12a828f-0cd6-4d43-a1e9-83bf31ec6616,Namespace:default,Attempt:0,} returns sandbox id \"9d714792f4a51a49152bbea93d1e249d0857d62fbddb06be19f4c3768134b72f\"" Feb 9 09:04:01.922155 env[1108]: time="2024-02-09T09:04:01.922109733Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\"" Feb 9 09:04:02.813265 kubelet[1395]: E0209 09:04:02.813196 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:03.310765 sshd[2156]: Connection closed by authenticating user root 85.209.11.27 port 9618 [preauth] Feb 9 09:04:03.312634 systemd[1]: sshd@5-24.144.92.253:22-85.209.11.27:9618.service: Deactivated successfully. Feb 9 09:04:03.813872 kubelet[1395]: E0209 09:04:03.813815 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:04.815014 kubelet[1395]: E0209 09:04:04.814963 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:05.414436 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1210434653.mount: Deactivated successfully. Feb 9 09:04:05.815393 kubelet[1395]: E0209 09:04:05.815344 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:06.619623 env[1108]: time="2024-02-09T09:04:06.619547965Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/nginx:latest,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 09:04:06.624318 env[1108]: time="2024-02-09T09:04:06.624270128Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:3a8963c304a2f89d2bfa055e07403bae348b293c891b8ea01f7136642eaa277a,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 09:04:06.629991 env[1108]: time="2024-02-09T09:04:06.629942617Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/nginx:latest,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 09:04:06.634582 env[1108]: time="2024-02-09T09:04:06.634535707Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/nginx@sha256:e34a272f01984c973b1e034e197c02f77dda18981038e3a54e957554ada4fec6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 09:04:06.635977 env[1108]: time="2024-02-09T09:04:06.635932095Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\" returns image reference \"sha256:3a8963c304a2f89d2bfa055e07403bae348b293c891b8ea01f7136642eaa277a\"" Feb 9 09:04:06.638552 env[1108]: time="2024-02-09T09:04:06.638515940Z" level=info msg="CreateContainer within sandbox \"9d714792f4a51a49152bbea93d1e249d0857d62fbddb06be19f4c3768134b72f\" for container &ContainerMetadata{Name:nginx,Attempt:0,}" Feb 9 09:04:06.657026 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount284280987.mount: Deactivated successfully. Feb 9 09:04:06.667303 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1121948672.mount: Deactivated successfully. Feb 9 09:04:06.669765 env[1108]: time="2024-02-09T09:04:06.669707454Z" level=info msg="CreateContainer within sandbox \"9d714792f4a51a49152bbea93d1e249d0857d62fbddb06be19f4c3768134b72f\" for &ContainerMetadata{Name:nginx,Attempt:0,} returns container id \"eceff704cc00fd3507073056079b46cea7461116c16b786c2ad3047263a75cfc\"" Feb 9 09:04:06.670910 env[1108]: time="2024-02-09T09:04:06.670873829Z" level=info msg="StartContainer for \"eceff704cc00fd3507073056079b46cea7461116c16b786c2ad3047263a75cfc\"" Feb 9 09:04:06.697116 systemd[1]: Started cri-containerd-eceff704cc00fd3507073056079b46cea7461116c16b786c2ad3047263a75cfc.scope. Feb 9 09:04:06.747358 env[1108]: time="2024-02-09T09:04:06.747294440Z" level=info msg="StartContainer for \"eceff704cc00fd3507073056079b46cea7461116c16b786c2ad3047263a75cfc\" returns successfully" Feb 9 09:04:06.816336 kubelet[1395]: E0209 09:04:06.816273 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:07.043760 kubelet[1395]: I0209 09:04:07.043698 1395 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="default/nginx-deployment-6d5f899847-xw5zj" podStartSLOduration=8.328581694 podCreationTimestamp="2024-02-09 09:03:54 +0000 UTC" firstStartedPulling="2024-02-09 09:04:01.92135523 +0000 UTC m=+26.734757391" lastFinishedPulling="2024-02-09 09:04:06.636404357 +0000 UTC m=+31.449806516" observedRunningTime="2024-02-09 09:04:07.043156501 +0000 UTC m=+31.856558666" watchObservedRunningTime="2024-02-09 09:04:07.043630819 +0000 UTC m=+31.857032985" Feb 9 09:04:07.817411 kubelet[1395]: E0209 09:04:07.817364 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:08.413697 kubelet[1395]: I0209 09:04:08.413640 1395 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 9 09:04:08.415026 kubelet[1395]: E0209 09:04:08.414994 1395 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Feb 9 09:04:08.474586 update_engine[1099]: I0209 09:04:08.474465 1099 update_attempter.cc:509] Updating boot flags... Feb 9 09:04:08.819006 kubelet[1395]: E0209 09:04:08.818959 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:09.037045 kubelet[1395]: E0209 09:04:09.037014 1395 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Feb 9 09:04:09.820683 kubelet[1395]: E0209 09:04:09.820618 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:10.821793 kubelet[1395]: E0209 09:04:10.821727 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:11.822833 kubelet[1395]: E0209 09:04:11.822770 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:12.793887 kubelet[1395]: I0209 09:04:12.793841 1395 topology_manager.go:215] "Topology Admit Handler" podUID="3e10ae0f-f983-4fe3-8804-780360f545ed" podNamespace="default" podName="nfs-server-provisioner-0" Feb 9 09:04:12.803584 systemd[1]: Created slice kubepods-besteffort-pod3e10ae0f_f983_4fe3_8804_780360f545ed.slice. Feb 9 09:04:12.824008 kubelet[1395]: E0209 09:04:12.823943 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:12.907743 kubelet[1395]: I0209 09:04:12.907660 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data\" (UniqueName: \"kubernetes.io/empty-dir/3e10ae0f-f983-4fe3-8804-780360f545ed-data\") pod \"nfs-server-provisioner-0\" (UID: \"3e10ae0f-f983-4fe3-8804-780360f545ed\") " pod="default/nfs-server-provisioner-0" Feb 9 09:04:12.907743 kubelet[1395]: I0209 09:04:12.907761 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gg2xc\" (UniqueName: \"kubernetes.io/projected/3e10ae0f-f983-4fe3-8804-780360f545ed-kube-api-access-gg2xc\") pod \"nfs-server-provisioner-0\" (UID: \"3e10ae0f-f983-4fe3-8804-780360f545ed\") " pod="default/nfs-server-provisioner-0" Feb 9 09:04:13.108739 env[1108]: time="2024-02-09T09:04:13.108461926Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nfs-server-provisioner-0,Uid:3e10ae0f-f983-4fe3-8804-780360f545ed,Namespace:default,Attempt:0,}" Feb 9 09:04:13.185404 systemd-networkd[1002]: lxcee6f5d7619fd: Link UP Feb 9 09:04:13.198634 kernel: eth0: renamed from tmp9892e Feb 9 09:04:13.205297 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready Feb 9 09:04:13.205450 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxcee6f5d7619fd: link becomes ready Feb 9 09:04:13.205795 systemd-networkd[1002]: lxcee6f5d7619fd: Gained carrier Feb 9 09:04:13.560847 env[1108]: time="2024-02-09T09:04:13.560700798Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 9 09:04:13.560847 env[1108]: time="2024-02-09T09:04:13.560780497Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 9 09:04:13.561260 env[1108]: time="2024-02-09T09:04:13.560797964Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 9 09:04:13.561475 env[1108]: time="2024-02-09T09:04:13.561300306Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/9892ee5967d38491dfb3a8fbdddc01bcab9fff886a5c194bab99f720e3a3f0c1 pid=2589 runtime=io.containerd.runc.v2 Feb 9 09:04:13.591989 systemd[1]: Started cri-containerd-9892ee5967d38491dfb3a8fbdddc01bcab9fff886a5c194bab99f720e3a3f0c1.scope. Feb 9 09:04:13.667367 env[1108]: time="2024-02-09T09:04:13.667305198Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nfs-server-provisioner-0,Uid:3e10ae0f-f983-4fe3-8804-780360f545ed,Namespace:default,Attempt:0,} returns sandbox id \"9892ee5967d38491dfb3a8fbdddc01bcab9fff886a5c194bab99f720e3a3f0c1\"" Feb 9 09:04:13.670239 env[1108]: time="2024-02-09T09:04:13.670185493Z" level=info msg="PullImage \"registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8\"" Feb 9 09:04:13.825289 kubelet[1395]: E0209 09:04:13.824518 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:14.825667 kubelet[1395]: E0209 09:04:14.825586 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:15.255928 systemd-networkd[1002]: lxcee6f5d7619fd: Gained IPv6LL Feb 9 09:04:15.791409 kubelet[1395]: E0209 09:04:15.791334 1395 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:15.826629 kubelet[1395]: E0209 09:04:15.826551 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:16.826790 kubelet[1395]: E0209 09:04:16.826690 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:16.993214 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount336380903.mount: Deactivated successfully. Feb 9 09:04:17.827668 kubelet[1395]: E0209 09:04:17.827582 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:18.827833 kubelet[1395]: E0209 09:04:18.827741 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:19.828477 kubelet[1395]: E0209 09:04:19.828413 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:19.868939 env[1108]: time="2024-02-09T09:04:19.868881588Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 09:04:19.871774 env[1108]: time="2024-02-09T09:04:19.871726445Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:fd0b16f70b66b72bcb2f91d556fa33eba02729c44ffc5f2c16130e7f9fbed3c4,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 09:04:19.875172 env[1108]: time="2024-02-09T09:04:19.875125863Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 09:04:19.878472 env[1108]: time="2024-02-09T09:04:19.878389281Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/sig-storage/nfs-provisioner@sha256:c825f3d5e28bde099bd7a3daace28772d412c9157ad47fa752a9ad0baafc118d,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 09:04:19.880180 env[1108]: time="2024-02-09T09:04:19.880131272Z" level=info msg="PullImage \"registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8\" returns image reference \"sha256:fd0b16f70b66b72bcb2f91d556fa33eba02729c44ffc5f2c16130e7f9fbed3c4\"" Feb 9 09:04:19.883764 env[1108]: time="2024-02-09T09:04:19.883720929Z" level=info msg="CreateContainer within sandbox \"9892ee5967d38491dfb3a8fbdddc01bcab9fff886a5c194bab99f720e3a3f0c1\" for container &ContainerMetadata{Name:nfs-server-provisioner,Attempt:0,}" Feb 9 09:04:19.901173 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2630338615.mount: Deactivated successfully. Feb 9 09:04:19.912644 env[1108]: time="2024-02-09T09:04:19.912541436Z" level=info msg="CreateContainer within sandbox \"9892ee5967d38491dfb3a8fbdddc01bcab9fff886a5c194bab99f720e3a3f0c1\" for &ContainerMetadata{Name:nfs-server-provisioner,Attempt:0,} returns container id \"769bba9ed4a94e7bc9c35411aac1868708adfb7173a52d08c28dbc967357cc4b\"" Feb 9 09:04:19.913398 env[1108]: time="2024-02-09T09:04:19.913346681Z" level=info msg="StartContainer for \"769bba9ed4a94e7bc9c35411aac1868708adfb7173a52d08c28dbc967357cc4b\"" Feb 9 09:04:19.951048 systemd[1]: Started cri-containerd-769bba9ed4a94e7bc9c35411aac1868708adfb7173a52d08c28dbc967357cc4b.scope. Feb 9 09:04:19.991103 env[1108]: time="2024-02-09T09:04:19.991032679Z" level=info msg="StartContainer for \"769bba9ed4a94e7bc9c35411aac1868708adfb7173a52d08c28dbc967357cc4b\" returns successfully" Feb 9 09:04:20.077297 kubelet[1395]: I0209 09:04:20.077145 1395 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="default/nfs-server-provisioner-0" podStartSLOduration=1.865712581 podCreationTimestamp="2024-02-09 09:04:12 +0000 UTC" firstStartedPulling="2024-02-09 09:04:13.669410194 +0000 UTC m=+38.482812338" lastFinishedPulling="2024-02-09 09:04:19.880798106 +0000 UTC m=+44.694200255" observedRunningTime="2024-02-09 09:04:20.07661659 +0000 UTC m=+44.890018757" watchObservedRunningTime="2024-02-09 09:04:20.077100498 +0000 UTC m=+44.890502665" Feb 9 09:04:20.829003 kubelet[1395]: E0209 09:04:20.828939 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:21.829684 kubelet[1395]: E0209 09:04:21.829588 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:22.830448 kubelet[1395]: E0209 09:04:22.830382 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:23.831451 kubelet[1395]: E0209 09:04:23.831384 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:24.831702 kubelet[1395]: E0209 09:04:24.831655 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:25.833283 kubelet[1395]: E0209 09:04:25.833189 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:26.834431 kubelet[1395]: E0209 09:04:26.834382 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:27.836029 kubelet[1395]: E0209 09:04:27.835983 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:28.836831 kubelet[1395]: E0209 09:04:28.836783 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:29.290531 kubelet[1395]: I0209 09:04:29.290449 1395 topology_manager.go:215] "Topology Admit Handler" podUID="4bea8d8a-08b5-4519-8716-3ef9462070b5" podNamespace="default" podName="test-pod-1" Feb 9 09:04:29.299903 systemd[1]: Created slice kubepods-besteffort-pod4bea8d8a_08b5_4519_8716_3ef9462070b5.slice. Feb 9 09:04:29.426811 kubelet[1395]: I0209 09:04:29.426722 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-2c476634-b123-463a-89e3-24485672e54f\" (UniqueName: \"kubernetes.io/nfs/4bea8d8a-08b5-4519-8716-3ef9462070b5-pvc-2c476634-b123-463a-89e3-24485672e54f\") pod \"test-pod-1\" (UID: \"4bea8d8a-08b5-4519-8716-3ef9462070b5\") " pod="default/test-pod-1" Feb 9 09:04:29.426811 kubelet[1395]: I0209 09:04:29.426797 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcr4l\" (UniqueName: \"kubernetes.io/projected/4bea8d8a-08b5-4519-8716-3ef9462070b5-kube-api-access-qcr4l\") pod \"test-pod-1\" (UID: \"4bea8d8a-08b5-4519-8716-3ef9462070b5\") " pod="default/test-pod-1" Feb 9 09:04:29.582534 kernel: FS-Cache: Loaded Feb 9 09:04:29.641416 kernel: RPC: Registered named UNIX socket transport module. Feb 9 09:04:29.641632 kernel: RPC: Registered udp transport module. Feb 9 09:04:29.641681 kernel: RPC: Registered tcp transport module. Feb 9 09:04:29.642751 kernel: RPC: Registered tcp NFSv4.1 backchannel transport module. Feb 9 09:04:29.722556 kernel: FS-Cache: Netfs 'nfs' registered for caching Feb 9 09:04:29.841092 kubelet[1395]: E0209 09:04:29.838383 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:29.954350 kernel: NFS: Registering the id_resolver key type Feb 9 09:04:29.954528 kernel: Key type id_resolver registered Feb 9 09:04:29.956534 kernel: Key type id_legacy registered Feb 9 09:04:30.839476 kubelet[1395]: E0209 09:04:30.839418 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:31.840804 kubelet[1395]: E0209 09:04:31.840739 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:32.841893 kubelet[1395]: E0209 09:04:32.841844 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:33.842935 kubelet[1395]: E0209 09:04:33.842842 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:34.843849 kubelet[1395]: E0209 09:04:34.843778 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:35.791094 kubelet[1395]: E0209 09:04:35.791035 1395 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:35.844846 kubelet[1395]: E0209 09:04:35.844797 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:36.123550 nfsidmap[2705]: nss_getpwnam: name 'root@nfs-server-provisioner.default.svc.cluster.local' does not map into domain '3.2-6-d69396b42b' Feb 9 09:04:36.846164 kubelet[1395]: E0209 09:04:36.846077 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:37.846689 kubelet[1395]: E0209 09:04:37.846640 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:38.847659 kubelet[1395]: E0209 09:04:38.847615 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:39.849035 kubelet[1395]: E0209 09:04:39.848973 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:40.849818 kubelet[1395]: E0209 09:04:40.849745 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:41.850033 kubelet[1395]: E0209 09:04:41.849956 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:42.267933 nfsidmap[2710]: nss_name_to_gid: name 'root@nfs-server-provisioner.default.svc.cluster.local' does not map into domain '3.2-6-d69396b42b' Feb 9 09:04:42.508168 env[1108]: time="2024-02-09T09:04:42.507623079Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:test-pod-1,Uid:4bea8d8a-08b5-4519-8716-3ef9462070b5,Namespace:default,Attempt:0,}" Feb 9 09:04:42.560647 systemd-networkd[1002]: lxcb19816d1930e: Link UP Feb 9 09:04:42.569552 kernel: eth0: renamed from tmp6b944 Feb 9 09:04:42.577433 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready Feb 9 09:04:42.577597 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxcb19816d1930e: link becomes ready Feb 9 09:04:42.577827 systemd-networkd[1002]: lxcb19816d1930e: Gained carrier Feb 9 09:04:42.850551 kubelet[1395]: E0209 09:04:42.850367 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:42.880059 env[1108]: time="2024-02-09T09:04:42.879931609Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 9 09:04:42.880059 env[1108]: time="2024-02-09T09:04:42.880006602Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 9 09:04:42.880390 env[1108]: time="2024-02-09T09:04:42.880022769Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 9 09:04:42.880518 env[1108]: time="2024-02-09T09:04:42.880404790Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/6b944cd333978b287260b4697a5a984a1a9126f0e6335f585dca5966eb097697 pid=2740 runtime=io.containerd.runc.v2 Feb 9 09:04:42.900456 systemd[1]: Started cri-containerd-6b944cd333978b287260b4697a5a984a1a9126f0e6335f585dca5966eb097697.scope. Feb 9 09:04:42.974719 env[1108]: time="2024-02-09T09:04:42.974311124Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:test-pod-1,Uid:4bea8d8a-08b5-4519-8716-3ef9462070b5,Namespace:default,Attempt:0,} returns sandbox id \"6b944cd333978b287260b4697a5a984a1a9126f0e6335f585dca5966eb097697\"" Feb 9 09:04:42.981811 env[1108]: time="2024-02-09T09:04:42.981660243Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\"" Feb 9 09:04:43.552956 env[1108]: time="2024-02-09T09:04:43.552899477Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/nginx:latest,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 09:04:43.560362 env[1108]: time="2024-02-09T09:04:43.559518715Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:3a8963c304a2f89d2bfa055e07403bae348b293c891b8ea01f7136642eaa277a,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 09:04:43.566322 env[1108]: time="2024-02-09T09:04:43.566263320Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/nginx:latest,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 09:04:43.573450 env[1108]: time="2024-02-09T09:04:43.573388424Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/nginx@sha256:e34a272f01984c973b1e034e197c02f77dda18981038e3a54e957554ada4fec6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 09:04:43.574649 env[1108]: time="2024-02-09T09:04:43.574561273Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\" returns image reference \"sha256:3a8963c304a2f89d2bfa055e07403bae348b293c891b8ea01f7136642eaa277a\"" Feb 9 09:04:43.577904 env[1108]: time="2024-02-09T09:04:43.577843500Z" level=info msg="CreateContainer within sandbox \"6b944cd333978b287260b4697a5a984a1a9126f0e6335f585dca5966eb097697\" for container &ContainerMetadata{Name:test,Attempt:0,}" Feb 9 09:04:43.603932 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1693813191.mount: Deactivated successfully. Feb 9 09:04:43.620974 env[1108]: time="2024-02-09T09:04:43.620910725Z" level=info msg="CreateContainer within sandbox \"6b944cd333978b287260b4697a5a984a1a9126f0e6335f585dca5966eb097697\" for &ContainerMetadata{Name:test,Attempt:0,} returns container id \"89c2c9fb54f8a0d3e3e6acbc3ad95f4cea676e75f3344bd1cd42b1d1aeb6cddb\"" Feb 9 09:04:43.622288 env[1108]: time="2024-02-09T09:04:43.622228300Z" level=info msg="StartContainer for \"89c2c9fb54f8a0d3e3e6acbc3ad95f4cea676e75f3344bd1cd42b1d1aeb6cddb\"" Feb 9 09:04:43.658349 systemd[1]: Started cri-containerd-89c2c9fb54f8a0d3e3e6acbc3ad95f4cea676e75f3344bd1cd42b1d1aeb6cddb.scope. Feb 9 09:04:43.713754 env[1108]: time="2024-02-09T09:04:43.713684229Z" level=info msg="StartContainer for \"89c2c9fb54f8a0d3e3e6acbc3ad95f4cea676e75f3344bd1cd42b1d1aeb6cddb\" returns successfully" Feb 9 09:04:43.851394 kubelet[1395]: E0209 09:04:43.851228 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:44.136163 kubelet[1395]: I0209 09:04:44.136001 1395 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="default/test-pod-1" podStartSLOduration=30.541451299 podCreationTimestamp="2024-02-09 09:04:13 +0000 UTC" firstStartedPulling="2024-02-09 09:04:42.980695893 +0000 UTC m=+67.794098116" lastFinishedPulling="2024-02-09 09:04:43.575180957 +0000 UTC m=+68.388583105" observedRunningTime="2024-02-09 09:04:44.135451099 +0000 UTC m=+68.948853272" watchObservedRunningTime="2024-02-09 09:04:44.135936288 +0000 UTC m=+68.949338510" Feb 9 09:04:44.440702 systemd-networkd[1002]: lxcb19816d1930e: Gained IPv6LL Feb 9 09:04:44.852250 kubelet[1395]: E0209 09:04:44.852177 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:45.852800 kubelet[1395]: E0209 09:04:45.852737 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:46.853417 kubelet[1395]: E0209 09:04:46.853359 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:47.385484 systemd[1]: run-containerd-runc-k8s.io-3054d463ceb9b9e2d733ac777f55d5bc633d9a45b71193841f8ad8b94161254a-runc.JbpQbp.mount: Deactivated successfully. Feb 9 09:04:47.411162 env[1108]: time="2024-02-09T09:04:47.410997705Z" level=error msg="failed to reload cni configuration after receiving fs change event(\"/etc/cni/net.d/05-cilium.conf\": REMOVE)" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Feb 9 09:04:47.418903 env[1108]: time="2024-02-09T09:04:47.418853270Z" level=info msg="StopContainer for \"3054d463ceb9b9e2d733ac777f55d5bc633d9a45b71193841f8ad8b94161254a\" with timeout 2 (s)" Feb 9 09:04:47.419326 env[1108]: time="2024-02-09T09:04:47.419262414Z" level=info msg="Stop container \"3054d463ceb9b9e2d733ac777f55d5bc633d9a45b71193841f8ad8b94161254a\" with signal terminated" Feb 9 09:04:47.427648 systemd-networkd[1002]: lxc_health: Link DOWN Feb 9 09:04:47.427662 systemd-networkd[1002]: lxc_health: Lost carrier Feb 9 09:04:47.469048 systemd[1]: cri-containerd-3054d463ceb9b9e2d733ac777f55d5bc633d9a45b71193841f8ad8b94161254a.scope: Deactivated successfully. Feb 9 09:04:47.469405 systemd[1]: cri-containerd-3054d463ceb9b9e2d733ac777f55d5bc633d9a45b71193841f8ad8b94161254a.scope: Consumed 9.039s CPU time. Feb 9 09:04:47.495073 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-3054d463ceb9b9e2d733ac777f55d5bc633d9a45b71193841f8ad8b94161254a-rootfs.mount: Deactivated successfully. Feb 9 09:04:47.514191 env[1108]: time="2024-02-09T09:04:47.514130482Z" level=info msg="shim disconnected" id=3054d463ceb9b9e2d733ac777f55d5bc633d9a45b71193841f8ad8b94161254a Feb 9 09:04:47.514191 env[1108]: time="2024-02-09T09:04:47.514185481Z" level=warning msg="cleaning up after shim disconnected" id=3054d463ceb9b9e2d733ac777f55d5bc633d9a45b71193841f8ad8b94161254a namespace=k8s.io Feb 9 09:04:47.514191 env[1108]: time="2024-02-09T09:04:47.514197981Z" level=info msg="cleaning up dead shim" Feb 9 09:04:47.526165 env[1108]: time="2024-02-09T09:04:47.526070492Z" level=warning msg="cleanup warnings time=\"2024-02-09T09:04:47Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2870 runtime=io.containerd.runc.v2\n" Feb 9 09:04:47.529483 env[1108]: time="2024-02-09T09:04:47.529412756Z" level=info msg="StopContainer for \"3054d463ceb9b9e2d733ac777f55d5bc633d9a45b71193841f8ad8b94161254a\" returns successfully" Feb 9 09:04:47.530223 env[1108]: time="2024-02-09T09:04:47.530188557Z" level=info msg="StopPodSandbox for \"3f55ee1ffb7b986f1ae81b075696d310d94fa86e0069101af0741e798b3bc5e4\"" Feb 9 09:04:47.530419 env[1108]: time="2024-02-09T09:04:47.530395077Z" level=info msg="Container to stop \"ae2f0999c1a3f2ff36597066153f98e00c426f1f85f40515b293ae9338d9c5ea\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Feb 9 09:04:47.530534 env[1108]: time="2024-02-09T09:04:47.530481847Z" level=info msg="Container to stop \"3054d463ceb9b9e2d733ac777f55d5bc633d9a45b71193841f8ad8b94161254a\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Feb 9 09:04:47.530717 env[1108]: time="2024-02-09T09:04:47.530683689Z" level=info msg="Container to stop \"fa22ea78685b767657fec5a8225818e4e3491ed5dfcac1158214eb78bac1e67d\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Feb 9 09:04:47.530838 env[1108]: time="2024-02-09T09:04:47.530808827Z" level=info msg="Container to stop \"e19abab8f3746b464980f337f5669da3391695f31e285c11842661765089f999\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Feb 9 09:04:47.530955 env[1108]: time="2024-02-09T09:04:47.530925293Z" level=info msg="Container to stop \"82189475b3479daae734762dd9373163de30c083a7ed75ee7a8f2e9f15db74b0\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Feb 9 09:04:47.533623 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-3f55ee1ffb7b986f1ae81b075696d310d94fa86e0069101af0741e798b3bc5e4-shm.mount: Deactivated successfully. Feb 9 09:04:47.542912 systemd[1]: cri-containerd-3f55ee1ffb7b986f1ae81b075696d310d94fa86e0069101af0741e798b3bc5e4.scope: Deactivated successfully. Feb 9 09:04:47.574810 env[1108]: time="2024-02-09T09:04:47.574735882Z" level=info msg="shim disconnected" id=3f55ee1ffb7b986f1ae81b075696d310d94fa86e0069101af0741e798b3bc5e4 Feb 9 09:04:47.574810 env[1108]: time="2024-02-09T09:04:47.574809593Z" level=warning msg="cleaning up after shim disconnected" id=3f55ee1ffb7b986f1ae81b075696d310d94fa86e0069101af0741e798b3bc5e4 namespace=k8s.io Feb 9 09:04:47.575365 env[1108]: time="2024-02-09T09:04:47.574828578Z" level=info msg="cleaning up dead shim" Feb 9 09:04:47.587212 env[1108]: time="2024-02-09T09:04:47.587159206Z" level=warning msg="cleanup warnings time=\"2024-02-09T09:04:47Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2901 runtime=io.containerd.runc.v2\n" Feb 9 09:04:47.587811 env[1108]: time="2024-02-09T09:04:47.587776174Z" level=info msg="TearDown network for sandbox \"3f55ee1ffb7b986f1ae81b075696d310d94fa86e0069101af0741e798b3bc5e4\" successfully" Feb 9 09:04:47.587944 env[1108]: time="2024-02-09T09:04:47.587922414Z" level=info msg="StopPodSandbox for \"3f55ee1ffb7b986f1ae81b075696d310d94fa86e0069101af0741e798b3bc5e4\" returns successfully" Feb 9 09:04:47.668993 kubelet[1395]: I0209 09:04:47.668866 1395 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-cni-path\") pod \"d05f098b-a09d-4a49-8adf-e1321bc177c6\" (UID: \"d05f098b-a09d-4a49-8adf-e1321bc177c6\") " Feb 9 09:04:47.669869 kubelet[1395]: I0209 09:04:47.669772 1395 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-cni-path" (OuterVolumeSpecName: "cni-path") pod "d05f098b-a09d-4a49-8adf-e1321bc177c6" (UID: "d05f098b-a09d-4a49-8adf-e1321bc177c6"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:04:47.669966 kubelet[1395]: I0209 09:04:47.669936 1395 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "d05f098b-a09d-4a49-8adf-e1321bc177c6" (UID: "d05f098b-a09d-4a49-8adf-e1321bc177c6"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:04:47.670032 kubelet[1395]: I0209 09:04:47.669965 1395 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-lib-modules\") pod \"d05f098b-a09d-4a49-8adf-e1321bc177c6\" (UID: \"d05f098b-a09d-4a49-8adf-e1321bc177c6\") " Feb 9 09:04:47.670032 kubelet[1395]: I0209 09:04:47.670013 1395 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/d05f098b-a09d-4a49-8adf-e1321bc177c6-clustermesh-secrets\") pod \"d05f098b-a09d-4a49-8adf-e1321bc177c6\" (UID: \"d05f098b-a09d-4a49-8adf-e1321bc177c6\") " Feb 9 09:04:47.670158 kubelet[1395]: I0209 09:04:47.670034 1395 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-host-proc-sys-net\") pod \"d05f098b-a09d-4a49-8adf-e1321bc177c6\" (UID: \"d05f098b-a09d-4a49-8adf-e1321bc177c6\") " Feb 9 09:04:47.670469 kubelet[1395]: I0209 09:04:47.670448 1395 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-cilium-cgroup\") pod \"d05f098b-a09d-4a49-8adf-e1321bc177c6\" (UID: \"d05f098b-a09d-4a49-8adf-e1321bc177c6\") " Feb 9 09:04:47.670565 kubelet[1395]: I0209 09:04:47.670514 1395 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djcfj\" (UniqueName: \"kubernetes.io/projected/d05f098b-a09d-4a49-8adf-e1321bc177c6-kube-api-access-djcfj\") pod \"d05f098b-a09d-4a49-8adf-e1321bc177c6\" (UID: \"d05f098b-a09d-4a49-8adf-e1321bc177c6\") " Feb 9 09:04:47.670565 kubelet[1395]: I0209 09:04:47.670537 1395 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-hostproc\") pod \"d05f098b-a09d-4a49-8adf-e1321bc177c6\" (UID: \"d05f098b-a09d-4a49-8adf-e1321bc177c6\") " Feb 9 09:04:47.670565 kubelet[1395]: I0209 09:04:47.670555 1395 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-xtables-lock\") pod \"d05f098b-a09d-4a49-8adf-e1321bc177c6\" (UID: \"d05f098b-a09d-4a49-8adf-e1321bc177c6\") " Feb 9 09:04:47.670728 kubelet[1395]: I0209 09:04:47.670575 1395 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/d05f098b-a09d-4a49-8adf-e1321bc177c6-cilium-config-path\") pod \"d05f098b-a09d-4a49-8adf-e1321bc177c6\" (UID: \"d05f098b-a09d-4a49-8adf-e1321bc177c6\") " Feb 9 09:04:47.670728 kubelet[1395]: I0209 09:04:47.670604 1395 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-bpf-maps\") pod \"d05f098b-a09d-4a49-8adf-e1321bc177c6\" (UID: \"d05f098b-a09d-4a49-8adf-e1321bc177c6\") " Feb 9 09:04:47.670728 kubelet[1395]: I0209 09:04:47.670625 1395 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-host-proc-sys-kernel\") pod \"d05f098b-a09d-4a49-8adf-e1321bc177c6\" (UID: \"d05f098b-a09d-4a49-8adf-e1321bc177c6\") " Feb 9 09:04:47.670728 kubelet[1395]: I0209 09:04:47.670648 1395 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/d05f098b-a09d-4a49-8adf-e1321bc177c6-hubble-tls\") pod \"d05f098b-a09d-4a49-8adf-e1321bc177c6\" (UID: \"d05f098b-a09d-4a49-8adf-e1321bc177c6\") " Feb 9 09:04:47.670728 kubelet[1395]: I0209 09:04:47.670679 1395 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-etc-cni-netd\") pod \"d05f098b-a09d-4a49-8adf-e1321bc177c6\" (UID: \"d05f098b-a09d-4a49-8adf-e1321bc177c6\") " Feb 9 09:04:47.670728 kubelet[1395]: I0209 09:04:47.670700 1395 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-cilium-run\") pod \"d05f098b-a09d-4a49-8adf-e1321bc177c6\" (UID: \"d05f098b-a09d-4a49-8adf-e1321bc177c6\") " Feb 9 09:04:47.671088 kubelet[1395]: I0209 09:04:47.670750 1395 reconciler_common.go:300] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-cni-path\") on node \"24.144.92.253\" DevicePath \"\"" Feb 9 09:04:47.671088 kubelet[1395]: I0209 09:04:47.670762 1395 reconciler_common.go:300] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-lib-modules\") on node \"24.144.92.253\" DevicePath \"\"" Feb 9 09:04:47.671088 kubelet[1395]: I0209 09:04:47.670783 1395 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "d05f098b-a09d-4a49-8adf-e1321bc177c6" (UID: "d05f098b-a09d-4a49-8adf-e1321bc177c6"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:04:47.671088 kubelet[1395]: I0209 09:04:47.670805 1395 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "d05f098b-a09d-4a49-8adf-e1321bc177c6" (UID: "d05f098b-a09d-4a49-8adf-e1321bc177c6"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:04:47.671088 kubelet[1395]: I0209 09:04:47.670836 1395 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "d05f098b-a09d-4a49-8adf-e1321bc177c6" (UID: "d05f098b-a09d-4a49-8adf-e1321bc177c6"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:04:47.671347 kubelet[1395]: I0209 09:04:47.671184 1395 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-hostproc" (OuterVolumeSpecName: "hostproc") pod "d05f098b-a09d-4a49-8adf-e1321bc177c6" (UID: "d05f098b-a09d-4a49-8adf-e1321bc177c6"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:04:47.671347 kubelet[1395]: I0209 09:04:47.671211 1395 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "d05f098b-a09d-4a49-8adf-e1321bc177c6" (UID: "d05f098b-a09d-4a49-8adf-e1321bc177c6"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:04:47.673892 kubelet[1395]: I0209 09:04:47.673843 1395 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d05f098b-a09d-4a49-8adf-e1321bc177c6-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "d05f098b-a09d-4a49-8adf-e1321bc177c6" (UID: "d05f098b-a09d-4a49-8adf-e1321bc177c6"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 9 09:04:47.674067 kubelet[1395]: I0209 09:04:47.673925 1395 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "d05f098b-a09d-4a49-8adf-e1321bc177c6" (UID: "d05f098b-a09d-4a49-8adf-e1321bc177c6"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:04:47.674067 kubelet[1395]: I0209 09:04:47.673944 1395 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "d05f098b-a09d-4a49-8adf-e1321bc177c6" (UID: "d05f098b-a09d-4a49-8adf-e1321bc177c6"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:04:47.674215 kubelet[1395]: I0209 09:04:47.674193 1395 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "d05f098b-a09d-4a49-8adf-e1321bc177c6" (UID: "d05f098b-a09d-4a49-8adf-e1321bc177c6"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:04:47.675903 kubelet[1395]: I0209 09:04:47.675842 1395 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d05f098b-a09d-4a49-8adf-e1321bc177c6-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "d05f098b-a09d-4a49-8adf-e1321bc177c6" (UID: "d05f098b-a09d-4a49-8adf-e1321bc177c6"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 9 09:04:47.678721 kubelet[1395]: I0209 09:04:47.678677 1395 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d05f098b-a09d-4a49-8adf-e1321bc177c6-kube-api-access-djcfj" (OuterVolumeSpecName: "kube-api-access-djcfj") pod "d05f098b-a09d-4a49-8adf-e1321bc177c6" (UID: "d05f098b-a09d-4a49-8adf-e1321bc177c6"). InnerVolumeSpecName "kube-api-access-djcfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 9 09:04:47.679682 kubelet[1395]: I0209 09:04:47.679631 1395 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d05f098b-a09d-4a49-8adf-e1321bc177c6-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "d05f098b-a09d-4a49-8adf-e1321bc177c6" (UID: "d05f098b-a09d-4a49-8adf-e1321bc177c6"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 9 09:04:47.771216 kubelet[1395]: I0209 09:04:47.771126 1395 reconciler_common.go:300] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-etc-cni-netd\") on node \"24.144.92.253\" DevicePath \"\"" Feb 9 09:04:47.771216 kubelet[1395]: I0209 09:04:47.771178 1395 reconciler_common.go:300] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/d05f098b-a09d-4a49-8adf-e1321bc177c6-hubble-tls\") on node \"24.144.92.253\" DevicePath \"\"" Feb 9 09:04:47.771216 kubelet[1395]: I0209 09:04:47.771189 1395 reconciler_common.go:300] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-cilium-run\") on node \"24.144.92.253\" DevicePath \"\"" Feb 9 09:04:47.771216 kubelet[1395]: I0209 09:04:47.771201 1395 reconciler_common.go:300] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/d05f098b-a09d-4a49-8adf-e1321bc177c6-clustermesh-secrets\") on node \"24.144.92.253\" DevicePath \"\"" Feb 9 09:04:47.771216 kubelet[1395]: I0209 09:04:47.771210 1395 reconciler_common.go:300] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-host-proc-sys-net\") on node \"24.144.92.253\" DevicePath \"\"" Feb 9 09:04:47.771216 kubelet[1395]: I0209 09:04:47.771221 1395 reconciler_common.go:300] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-cilium-cgroup\") on node \"24.144.92.253\" DevicePath \"\"" Feb 9 09:04:47.771216 kubelet[1395]: I0209 09:04:47.771234 1395 reconciler_common.go:300] "Volume detached for volume \"kube-api-access-djcfj\" (UniqueName: \"kubernetes.io/projected/d05f098b-a09d-4a49-8adf-e1321bc177c6-kube-api-access-djcfj\") on node \"24.144.92.253\" DevicePath \"\"" Feb 9 09:04:47.771216 kubelet[1395]: I0209 09:04:47.771243 1395 reconciler_common.go:300] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-hostproc\") on node \"24.144.92.253\" DevicePath \"\"" Feb 9 09:04:47.771887 kubelet[1395]: I0209 09:04:47.771253 1395 reconciler_common.go:300] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/d05f098b-a09d-4a49-8adf-e1321bc177c6-cilium-config-path\") on node \"24.144.92.253\" DevicePath \"\"" Feb 9 09:04:47.771887 kubelet[1395]: I0209 09:04:47.771262 1395 reconciler_common.go:300] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-bpf-maps\") on node \"24.144.92.253\" DevicePath \"\"" Feb 9 09:04:47.771887 kubelet[1395]: I0209 09:04:47.771271 1395 reconciler_common.go:300] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-host-proc-sys-kernel\") on node \"24.144.92.253\" DevicePath \"\"" Feb 9 09:04:47.771887 kubelet[1395]: I0209 09:04:47.771283 1395 reconciler_common.go:300] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/d05f098b-a09d-4a49-8adf-e1321bc177c6-xtables-lock\") on node \"24.144.92.253\" DevicePath \"\"" Feb 9 09:04:47.854650 kubelet[1395]: E0209 09:04:47.854566 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:47.942185 systemd[1]: Removed slice kubepods-burstable-podd05f098b_a09d_4a49_8adf_e1321bc177c6.slice. Feb 9 09:04:47.942362 systemd[1]: kubepods-burstable-podd05f098b_a09d_4a49_8adf_e1321bc177c6.slice: Consumed 9.230s CPU time. Feb 9 09:04:48.134521 kubelet[1395]: I0209 09:04:48.134465 1395 scope.go:117] "RemoveContainer" containerID="3054d463ceb9b9e2d733ac777f55d5bc633d9a45b71193841f8ad8b94161254a" Feb 9 09:04:48.138667 env[1108]: time="2024-02-09T09:04:48.138610905Z" level=info msg="RemoveContainer for \"3054d463ceb9b9e2d733ac777f55d5bc633d9a45b71193841f8ad8b94161254a\"" Feb 9 09:04:48.143658 env[1108]: time="2024-02-09T09:04:48.143601392Z" level=info msg="RemoveContainer for \"3054d463ceb9b9e2d733ac777f55d5bc633d9a45b71193841f8ad8b94161254a\" returns successfully" Feb 9 09:04:48.144927 kubelet[1395]: I0209 09:04:48.144749 1395 scope.go:117] "RemoveContainer" containerID="ae2f0999c1a3f2ff36597066153f98e00c426f1f85f40515b293ae9338d9c5ea" Feb 9 09:04:48.146383 env[1108]: time="2024-02-09T09:04:48.146347644Z" level=info msg="RemoveContainer for \"ae2f0999c1a3f2ff36597066153f98e00c426f1f85f40515b293ae9338d9c5ea\"" Feb 9 09:04:48.157485 env[1108]: time="2024-02-09T09:04:48.157422988Z" level=info msg="RemoveContainer for \"ae2f0999c1a3f2ff36597066153f98e00c426f1f85f40515b293ae9338d9c5ea\" returns successfully" Feb 9 09:04:48.158108 kubelet[1395]: I0209 09:04:48.157965 1395 scope.go:117] "RemoveContainer" containerID="82189475b3479daae734762dd9373163de30c083a7ed75ee7a8f2e9f15db74b0" Feb 9 09:04:48.159609 env[1108]: time="2024-02-09T09:04:48.159564684Z" level=info msg="RemoveContainer for \"82189475b3479daae734762dd9373163de30c083a7ed75ee7a8f2e9f15db74b0\"" Feb 9 09:04:48.163974 env[1108]: time="2024-02-09T09:04:48.163920951Z" level=info msg="RemoveContainer for \"82189475b3479daae734762dd9373163de30c083a7ed75ee7a8f2e9f15db74b0\" returns successfully" Feb 9 09:04:48.164531 kubelet[1395]: I0209 09:04:48.164375 1395 scope.go:117] "RemoveContainer" containerID="e19abab8f3746b464980f337f5669da3391695f31e285c11842661765089f999" Feb 9 09:04:48.165681 env[1108]: time="2024-02-09T09:04:48.165648338Z" level=info msg="RemoveContainer for \"e19abab8f3746b464980f337f5669da3391695f31e285c11842661765089f999\"" Feb 9 09:04:48.170286 env[1108]: time="2024-02-09T09:04:48.170234686Z" level=info msg="RemoveContainer for \"e19abab8f3746b464980f337f5669da3391695f31e285c11842661765089f999\" returns successfully" Feb 9 09:04:48.170630 kubelet[1395]: I0209 09:04:48.170594 1395 scope.go:117] "RemoveContainer" containerID="fa22ea78685b767657fec5a8225818e4e3491ed5dfcac1158214eb78bac1e67d" Feb 9 09:04:48.172037 env[1108]: time="2024-02-09T09:04:48.171999057Z" level=info msg="RemoveContainer for \"fa22ea78685b767657fec5a8225818e4e3491ed5dfcac1158214eb78bac1e67d\"" Feb 9 09:04:48.176429 env[1108]: time="2024-02-09T09:04:48.176377907Z" level=info msg="RemoveContainer for \"fa22ea78685b767657fec5a8225818e4e3491ed5dfcac1158214eb78bac1e67d\" returns successfully" Feb 9 09:04:48.176645 kubelet[1395]: I0209 09:04:48.176621 1395 scope.go:117] "RemoveContainer" containerID="3054d463ceb9b9e2d733ac777f55d5bc633d9a45b71193841f8ad8b94161254a" Feb 9 09:04:48.176933 env[1108]: time="2024-02-09T09:04:48.176850948Z" level=error msg="ContainerStatus for \"3054d463ceb9b9e2d733ac777f55d5bc633d9a45b71193841f8ad8b94161254a\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"3054d463ceb9b9e2d733ac777f55d5bc633d9a45b71193841f8ad8b94161254a\": not found" Feb 9 09:04:48.177431 kubelet[1395]: E0209 09:04:48.177209 1395 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"3054d463ceb9b9e2d733ac777f55d5bc633d9a45b71193841f8ad8b94161254a\": not found" containerID="3054d463ceb9b9e2d733ac777f55d5bc633d9a45b71193841f8ad8b94161254a" Feb 9 09:04:48.177431 kubelet[1395]: I0209 09:04:48.177333 1395 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"3054d463ceb9b9e2d733ac777f55d5bc633d9a45b71193841f8ad8b94161254a"} err="failed to get container status \"3054d463ceb9b9e2d733ac777f55d5bc633d9a45b71193841f8ad8b94161254a\": rpc error: code = NotFound desc = an error occurred when try to find container \"3054d463ceb9b9e2d733ac777f55d5bc633d9a45b71193841f8ad8b94161254a\": not found" Feb 9 09:04:48.177431 kubelet[1395]: I0209 09:04:48.177347 1395 scope.go:117] "RemoveContainer" containerID="ae2f0999c1a3f2ff36597066153f98e00c426f1f85f40515b293ae9338d9c5ea" Feb 9 09:04:48.177661 env[1108]: time="2024-02-09T09:04:48.177606472Z" level=error msg="ContainerStatus for \"ae2f0999c1a3f2ff36597066153f98e00c426f1f85f40515b293ae9338d9c5ea\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"ae2f0999c1a3f2ff36597066153f98e00c426f1f85f40515b293ae9338d9c5ea\": not found" Feb 9 09:04:48.177801 kubelet[1395]: E0209 09:04:48.177769 1395 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"ae2f0999c1a3f2ff36597066153f98e00c426f1f85f40515b293ae9338d9c5ea\": not found" containerID="ae2f0999c1a3f2ff36597066153f98e00c426f1f85f40515b293ae9338d9c5ea" Feb 9 09:04:48.177854 kubelet[1395]: I0209 09:04:48.177805 1395 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"ae2f0999c1a3f2ff36597066153f98e00c426f1f85f40515b293ae9338d9c5ea"} err="failed to get container status \"ae2f0999c1a3f2ff36597066153f98e00c426f1f85f40515b293ae9338d9c5ea\": rpc error: code = NotFound desc = an error occurred when try to find container \"ae2f0999c1a3f2ff36597066153f98e00c426f1f85f40515b293ae9338d9c5ea\": not found" Feb 9 09:04:48.177854 kubelet[1395]: I0209 09:04:48.177816 1395 scope.go:117] "RemoveContainer" containerID="82189475b3479daae734762dd9373163de30c083a7ed75ee7a8f2e9f15db74b0" Feb 9 09:04:48.178011 env[1108]: time="2024-02-09T09:04:48.177961584Z" level=error msg="ContainerStatus for \"82189475b3479daae734762dd9373163de30c083a7ed75ee7a8f2e9f15db74b0\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"82189475b3479daae734762dd9373163de30c083a7ed75ee7a8f2e9f15db74b0\": not found" Feb 9 09:04:48.178126 kubelet[1395]: E0209 09:04:48.178095 1395 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"82189475b3479daae734762dd9373163de30c083a7ed75ee7a8f2e9f15db74b0\": not found" containerID="82189475b3479daae734762dd9373163de30c083a7ed75ee7a8f2e9f15db74b0" Feb 9 09:04:48.178126 kubelet[1395]: I0209 09:04:48.178121 1395 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"82189475b3479daae734762dd9373163de30c083a7ed75ee7a8f2e9f15db74b0"} err="failed to get container status \"82189475b3479daae734762dd9373163de30c083a7ed75ee7a8f2e9f15db74b0\": rpc error: code = NotFound desc = an error occurred when try to find container \"82189475b3479daae734762dd9373163de30c083a7ed75ee7a8f2e9f15db74b0\": not found" Feb 9 09:04:48.178226 kubelet[1395]: I0209 09:04:48.178132 1395 scope.go:117] "RemoveContainer" containerID="e19abab8f3746b464980f337f5669da3391695f31e285c11842661765089f999" Feb 9 09:04:48.178315 env[1108]: time="2024-02-09T09:04:48.178271063Z" level=error msg="ContainerStatus for \"e19abab8f3746b464980f337f5669da3391695f31e285c11842661765089f999\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"e19abab8f3746b464980f337f5669da3391695f31e285c11842661765089f999\": not found" Feb 9 09:04:48.178604 kubelet[1395]: E0209 09:04:48.178467 1395 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"e19abab8f3746b464980f337f5669da3391695f31e285c11842661765089f999\": not found" containerID="e19abab8f3746b464980f337f5669da3391695f31e285c11842661765089f999" Feb 9 09:04:48.178604 kubelet[1395]: I0209 09:04:48.178519 1395 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"e19abab8f3746b464980f337f5669da3391695f31e285c11842661765089f999"} err="failed to get container status \"e19abab8f3746b464980f337f5669da3391695f31e285c11842661765089f999\": rpc error: code = NotFound desc = an error occurred when try to find container \"e19abab8f3746b464980f337f5669da3391695f31e285c11842661765089f999\": not found" Feb 9 09:04:48.178604 kubelet[1395]: I0209 09:04:48.178533 1395 scope.go:117] "RemoveContainer" containerID="fa22ea78685b767657fec5a8225818e4e3491ed5dfcac1158214eb78bac1e67d" Feb 9 09:04:48.178750 env[1108]: time="2024-02-09T09:04:48.178678730Z" level=error msg="ContainerStatus for \"fa22ea78685b767657fec5a8225818e4e3491ed5dfcac1158214eb78bac1e67d\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"fa22ea78685b767657fec5a8225818e4e3491ed5dfcac1158214eb78bac1e67d\": not found" Feb 9 09:04:48.178817 kubelet[1395]: E0209 09:04:48.178799 1395 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"fa22ea78685b767657fec5a8225818e4e3491ed5dfcac1158214eb78bac1e67d\": not found" containerID="fa22ea78685b767657fec5a8225818e4e3491ed5dfcac1158214eb78bac1e67d" Feb 9 09:04:48.178872 kubelet[1395]: I0209 09:04:48.178827 1395 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"fa22ea78685b767657fec5a8225818e4e3491ed5dfcac1158214eb78bac1e67d"} err="failed to get container status \"fa22ea78685b767657fec5a8225818e4e3491ed5dfcac1158214eb78bac1e67d\": rpc error: code = NotFound desc = an error occurred when try to find container \"fa22ea78685b767657fec5a8225818e4e3491ed5dfcac1158214eb78bac1e67d\": not found" Feb 9 09:04:48.379074 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-3f55ee1ffb7b986f1ae81b075696d310d94fa86e0069101af0741e798b3bc5e4-rootfs.mount: Deactivated successfully. Feb 9 09:04:48.379251 systemd[1]: var-lib-kubelet-pods-d05f098b\x2da09d\x2d4a49\x2d8adf\x2de1321bc177c6-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2ddjcfj.mount: Deactivated successfully. Feb 9 09:04:48.379367 systemd[1]: var-lib-kubelet-pods-d05f098b\x2da09d\x2d4a49\x2d8adf\x2de1321bc177c6-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Feb 9 09:04:48.379477 systemd[1]: var-lib-kubelet-pods-d05f098b\x2da09d\x2d4a49\x2d8adf\x2de1321bc177c6-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Feb 9 09:04:48.855083 kubelet[1395]: E0209 09:04:48.854973 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:49.855331 kubelet[1395]: E0209 09:04:49.855277 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:49.938238 kubelet[1395]: I0209 09:04:49.938200 1395 kubelet_volumes.go:161] "Cleaned up orphaned pod volumes dir" podUID="d05f098b-a09d-4a49-8adf-e1321bc177c6" path="/var/lib/kubelet/pods/d05f098b-a09d-4a49-8adf-e1321bc177c6/volumes" Feb 9 09:04:50.396330 kubelet[1395]: I0209 09:04:50.396245 1395 topology_manager.go:215] "Topology Admit Handler" podUID="e6ab66a4-cfd7-4756-9de7-aff42053eb74" podNamespace="kube-system" podName="cilium-operator-6bc8ccdb58-pxqvs" Feb 9 09:04:50.396330 kubelet[1395]: E0209 09:04:50.396321 1395 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="d05f098b-a09d-4a49-8adf-e1321bc177c6" containerName="clean-cilium-state" Feb 9 09:04:50.396330 kubelet[1395]: E0209 09:04:50.396337 1395 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="d05f098b-a09d-4a49-8adf-e1321bc177c6" containerName="cilium-agent" Feb 9 09:04:50.396330 kubelet[1395]: E0209 09:04:50.396347 1395 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="d05f098b-a09d-4a49-8adf-e1321bc177c6" containerName="mount-cgroup" Feb 9 09:04:50.396778 kubelet[1395]: E0209 09:04:50.396358 1395 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="d05f098b-a09d-4a49-8adf-e1321bc177c6" containerName="apply-sysctl-overwrites" Feb 9 09:04:50.396778 kubelet[1395]: E0209 09:04:50.396369 1395 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="d05f098b-a09d-4a49-8adf-e1321bc177c6" containerName="mount-bpf-fs" Feb 9 09:04:50.396778 kubelet[1395]: I0209 09:04:50.396394 1395 memory_manager.go:346] "RemoveStaleState removing state" podUID="d05f098b-a09d-4a49-8adf-e1321bc177c6" containerName="cilium-agent" Feb 9 09:04:50.403359 systemd[1]: Created slice kubepods-besteffort-pode6ab66a4_cfd7_4756_9de7_aff42053eb74.slice. Feb 9 09:04:50.415176 kubelet[1395]: W0209 09:04:50.415121 1395 reflector.go:535] object-"kube-system"/"cilium-config": failed to list *v1.ConfigMap: configmaps "cilium-config" is forbidden: User "system:node:24.144.92.253" cannot list resource "configmaps" in API group "" in the namespace "kube-system": no relationship found between node '24.144.92.253' and this object Feb 9 09:04:50.415398 kubelet[1395]: E0209 09:04:50.415196 1395 reflector.go:147] object-"kube-system"/"cilium-config": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps "cilium-config" is forbidden: User "system:node:24.144.92.253" cannot list resource "configmaps" in API group "" in the namespace "kube-system": no relationship found between node '24.144.92.253' and this object Feb 9 09:04:50.417131 kubelet[1395]: I0209 09:04:50.417081 1395 topology_manager.go:215] "Topology Admit Handler" podUID="14558848-333d-4982-be68-b07522bacf32" podNamespace="kube-system" podName="cilium-vhhz6" Feb 9 09:04:50.424475 systemd[1]: Created slice kubepods-burstable-pod14558848_333d_4982_be68_b07522bacf32.slice. Feb 9 09:04:50.491434 kubelet[1395]: I0209 09:04:50.491378 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-hostproc\") pod \"cilium-vhhz6\" (UID: \"14558848-333d-4982-be68-b07522bacf32\") " pod="kube-system/cilium-vhhz6" Feb 9 09:04:50.491755 kubelet[1395]: I0209 09:04:50.491724 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-etc-cni-netd\") pod \"cilium-vhhz6\" (UID: \"14558848-333d-4982-be68-b07522bacf32\") " pod="kube-system/cilium-vhhz6" Feb 9 09:04:50.491905 kubelet[1395]: I0209 09:04:50.491891 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/14558848-333d-4982-be68-b07522bacf32-clustermesh-secrets\") pod \"cilium-vhhz6\" (UID: \"14558848-333d-4982-be68-b07522bacf32\") " pod="kube-system/cilium-vhhz6" Feb 9 09:04:50.492024 kubelet[1395]: I0209 09:04:50.492011 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/14558848-333d-4982-be68-b07522bacf32-cilium-config-path\") pod \"cilium-vhhz6\" (UID: \"14558848-333d-4982-be68-b07522bacf32\") " pod="kube-system/cilium-vhhz6" Feb 9 09:04:50.492151 kubelet[1395]: I0209 09:04:50.492138 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-bpf-maps\") pod \"cilium-vhhz6\" (UID: \"14558848-333d-4982-be68-b07522bacf32\") " pod="kube-system/cilium-vhhz6" Feb 9 09:04:50.492268 kubelet[1395]: I0209 09:04:50.492256 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-cilium-cgroup\") pod \"cilium-vhhz6\" (UID: \"14558848-333d-4982-be68-b07522bacf32\") " pod="kube-system/cilium-vhhz6" Feb 9 09:04:50.492411 kubelet[1395]: I0209 09:04:50.492392 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-lib-modules\") pod \"cilium-vhhz6\" (UID: \"14558848-333d-4982-be68-b07522bacf32\") " pod="kube-system/cilium-vhhz6" Feb 9 09:04:50.492613 kubelet[1395]: I0209 09:04:50.492598 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/14558848-333d-4982-be68-b07522bacf32-cilium-ipsec-secrets\") pod \"cilium-vhhz6\" (UID: \"14558848-333d-4982-be68-b07522bacf32\") " pod="kube-system/cilium-vhhz6" Feb 9 09:04:50.492794 kubelet[1395]: I0209 09:04:50.492773 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-host-proc-sys-net\") pod \"cilium-vhhz6\" (UID: \"14558848-333d-4982-be68-b07522bacf32\") " pod="kube-system/cilium-vhhz6" Feb 9 09:04:50.492983 kubelet[1395]: I0209 09:04:50.492969 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-host-proc-sys-kernel\") pod \"cilium-vhhz6\" (UID: \"14558848-333d-4982-be68-b07522bacf32\") " pod="kube-system/cilium-vhhz6" Feb 9 09:04:50.493121 kubelet[1395]: I0209 09:04:50.493107 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-cni-path\") pod \"cilium-vhhz6\" (UID: \"14558848-333d-4982-be68-b07522bacf32\") " pod="kube-system/cilium-vhhz6" Feb 9 09:04:50.493276 kubelet[1395]: I0209 09:04:50.493260 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-xtables-lock\") pod \"cilium-vhhz6\" (UID: \"14558848-333d-4982-be68-b07522bacf32\") " pod="kube-system/cilium-vhhz6" Feb 9 09:04:50.493439 kubelet[1395]: I0209 09:04:50.493425 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/e6ab66a4-cfd7-4756-9de7-aff42053eb74-cilium-config-path\") pod \"cilium-operator-6bc8ccdb58-pxqvs\" (UID: \"e6ab66a4-cfd7-4756-9de7-aff42053eb74\") " pod="kube-system/cilium-operator-6bc8ccdb58-pxqvs" Feb 9 09:04:50.493597 kubelet[1395]: I0209 09:04:50.493583 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spmqr\" (UniqueName: \"kubernetes.io/projected/e6ab66a4-cfd7-4756-9de7-aff42053eb74-kube-api-access-spmqr\") pod \"cilium-operator-6bc8ccdb58-pxqvs\" (UID: \"e6ab66a4-cfd7-4756-9de7-aff42053eb74\") " pod="kube-system/cilium-operator-6bc8ccdb58-pxqvs" Feb 9 09:04:50.493730 kubelet[1395]: I0209 09:04:50.493717 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-cilium-run\") pod \"cilium-vhhz6\" (UID: \"14558848-333d-4982-be68-b07522bacf32\") " pod="kube-system/cilium-vhhz6" Feb 9 09:04:50.493858 kubelet[1395]: I0209 09:04:50.493845 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/14558848-333d-4982-be68-b07522bacf32-hubble-tls\") pod \"cilium-vhhz6\" (UID: \"14558848-333d-4982-be68-b07522bacf32\") " pod="kube-system/cilium-vhhz6" Feb 9 09:04:50.493998 kubelet[1395]: I0209 09:04:50.493984 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pccj\" (UniqueName: \"kubernetes.io/projected/14558848-333d-4982-be68-b07522bacf32-kube-api-access-9pccj\") pod \"cilium-vhhz6\" (UID: \"14558848-333d-4982-be68-b07522bacf32\") " pod="kube-system/cilium-vhhz6" Feb 9 09:04:50.625145 kubelet[1395]: E0209 09:04:50.623414 1395 pod_workers.go:1300] "Error syncing pod, skipping" err="unmounted volumes=[cilium-config-path cilium-ipsec-secrets], unattached volumes=[], failed to process volumes=[]: context canceled" pod="kube-system/cilium-vhhz6" podUID="14558848-333d-4982-be68-b07522bacf32" Feb 9 09:04:50.856162 kubelet[1395]: E0209 09:04:50.856027 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:50.910249 kubelet[1395]: E0209 09:04:50.910200 1395 kubelet.go:2855] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Feb 9 09:04:51.200590 kubelet[1395]: I0209 09:04:51.200428 1395 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-xtables-lock\") pod \"14558848-333d-4982-be68-b07522bacf32\" (UID: \"14558848-333d-4982-be68-b07522bacf32\") " Feb 9 09:04:51.200980 kubelet[1395]: I0209 09:04:51.200817 1395 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-cilium-run\") pod \"14558848-333d-4982-be68-b07522bacf32\" (UID: \"14558848-333d-4982-be68-b07522bacf32\") " Feb 9 09:04:51.201189 kubelet[1395]: I0209 09:04:51.201169 1395 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/14558848-333d-4982-be68-b07522bacf32-hubble-tls\") pod \"14558848-333d-4982-be68-b07522bacf32\" (UID: \"14558848-333d-4982-be68-b07522bacf32\") " Feb 9 09:04:51.201323 kubelet[1395]: I0209 09:04:51.201309 1395 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-etc-cni-netd\") pod \"14558848-333d-4982-be68-b07522bacf32\" (UID: \"14558848-333d-4982-be68-b07522bacf32\") " Feb 9 09:04:51.201436 kubelet[1395]: I0209 09:04:51.201423 1395 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-bpf-maps\") pod \"14558848-333d-4982-be68-b07522bacf32\" (UID: \"14558848-333d-4982-be68-b07522bacf32\") " Feb 9 09:04:51.201587 kubelet[1395]: I0209 09:04:51.201572 1395 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9pccj\" (UniqueName: \"kubernetes.io/projected/14558848-333d-4982-be68-b07522bacf32-kube-api-access-9pccj\") pod \"14558848-333d-4982-be68-b07522bacf32\" (UID: \"14558848-333d-4982-be68-b07522bacf32\") " Feb 9 09:04:51.201710 kubelet[1395]: I0209 09:04:51.201696 1395 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-lib-modules\") pod \"14558848-333d-4982-be68-b07522bacf32\" (UID: \"14558848-333d-4982-be68-b07522bacf32\") " Feb 9 09:04:51.201904 kubelet[1395]: I0209 09:04:51.201888 1395 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/14558848-333d-4982-be68-b07522bacf32-clustermesh-secrets\") pod \"14558848-333d-4982-be68-b07522bacf32\" (UID: \"14558848-333d-4982-be68-b07522bacf32\") " Feb 9 09:04:51.202050 kubelet[1395]: I0209 09:04:51.202028 1395 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-cilium-cgroup\") pod \"14558848-333d-4982-be68-b07522bacf32\" (UID: \"14558848-333d-4982-be68-b07522bacf32\") " Feb 9 09:04:51.202186 kubelet[1395]: I0209 09:04:51.202170 1395 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-host-proc-sys-net\") pod \"14558848-333d-4982-be68-b07522bacf32\" (UID: \"14558848-333d-4982-be68-b07522bacf32\") " Feb 9 09:04:51.202322 kubelet[1395]: I0209 09:04:51.202306 1395 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-host-proc-sys-kernel\") pod \"14558848-333d-4982-be68-b07522bacf32\" (UID: \"14558848-333d-4982-be68-b07522bacf32\") " Feb 9 09:04:51.202456 kubelet[1395]: I0209 09:04:51.202443 1395 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/14558848-333d-4982-be68-b07522bacf32-cilium-ipsec-secrets\") pod \"14558848-333d-4982-be68-b07522bacf32\" (UID: \"14558848-333d-4982-be68-b07522bacf32\") " Feb 9 09:04:51.202618 kubelet[1395]: I0209 09:04:51.202601 1395 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-cni-path\") pod \"14558848-333d-4982-be68-b07522bacf32\" (UID: \"14558848-333d-4982-be68-b07522bacf32\") " Feb 9 09:04:51.202817 kubelet[1395]: I0209 09:04:51.202799 1395 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-hostproc\") pod \"14558848-333d-4982-be68-b07522bacf32\" (UID: \"14558848-333d-4982-be68-b07522bacf32\") " Feb 9 09:04:51.203160 kubelet[1395]: I0209 09:04:51.203129 1395 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-hostproc" (OuterVolumeSpecName: "hostproc") pod "14558848-333d-4982-be68-b07522bacf32" (UID: "14558848-333d-4982-be68-b07522bacf32"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:04:51.203286 kubelet[1395]: I0209 09:04:51.200906 1395 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "14558848-333d-4982-be68-b07522bacf32" (UID: "14558848-333d-4982-be68-b07522bacf32"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:04:51.203379 kubelet[1395]: I0209 09:04:51.200944 1395 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "14558848-333d-4982-be68-b07522bacf32" (UID: "14558848-333d-4982-be68-b07522bacf32"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:04:51.203521 kubelet[1395]: I0209 09:04:51.203481 1395 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "14558848-333d-4982-be68-b07522bacf32" (UID: "14558848-333d-4982-be68-b07522bacf32"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:04:51.203636 kubelet[1395]: I0209 09:04:51.203618 1395 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "14558848-333d-4982-be68-b07522bacf32" (UID: "14558848-333d-4982-be68-b07522bacf32"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:04:51.205525 kubelet[1395]: I0209 09:04:51.205461 1395 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14558848-333d-4982-be68-b07522bacf32-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "14558848-333d-4982-be68-b07522bacf32" (UID: "14558848-333d-4982-be68-b07522bacf32"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 9 09:04:51.205722 kubelet[1395]: I0209 09:04:51.205538 1395 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "14558848-333d-4982-be68-b07522bacf32" (UID: "14558848-333d-4982-be68-b07522bacf32"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:04:51.205722 kubelet[1395]: I0209 09:04:51.205565 1395 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "14558848-333d-4982-be68-b07522bacf32" (UID: "14558848-333d-4982-be68-b07522bacf32"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:04:51.207995 kubelet[1395]: I0209 09:04:51.207952 1395 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14558848-333d-4982-be68-b07522bacf32-kube-api-access-9pccj" (OuterVolumeSpecName: "kube-api-access-9pccj") pod "14558848-333d-4982-be68-b07522bacf32" (UID: "14558848-333d-4982-be68-b07522bacf32"). InnerVolumeSpecName "kube-api-access-9pccj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 9 09:04:51.208245 kubelet[1395]: I0209 09:04:51.208225 1395 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "14558848-333d-4982-be68-b07522bacf32" (UID: "14558848-333d-4982-be68-b07522bacf32"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:04:51.209153 kubelet[1395]: I0209 09:04:51.209109 1395 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14558848-333d-4982-be68-b07522bacf32-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "14558848-333d-4982-be68-b07522bacf32" (UID: "14558848-333d-4982-be68-b07522bacf32"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 9 09:04:51.209245 kubelet[1395]: I0209 09:04:51.209169 1395 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "14558848-333d-4982-be68-b07522bacf32" (UID: "14558848-333d-4982-be68-b07522bacf32"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:04:51.209245 kubelet[1395]: I0209 09:04:51.209196 1395 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-cni-path" (OuterVolumeSpecName: "cni-path") pod "14558848-333d-4982-be68-b07522bacf32" (UID: "14558848-333d-4982-be68-b07522bacf32"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:04:51.212194 kubelet[1395]: I0209 09:04:51.212148 1395 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14558848-333d-4982-be68-b07522bacf32-cilium-ipsec-secrets" (OuterVolumeSpecName: "cilium-ipsec-secrets") pod "14558848-333d-4982-be68-b07522bacf32" (UID: "14558848-333d-4982-be68-b07522bacf32"). InnerVolumeSpecName "cilium-ipsec-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 9 09:04:51.303690 kubelet[1395]: I0209 09:04:51.303642 1395 reconciler_common.go:300] "Volume detached for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/14558848-333d-4982-be68-b07522bacf32-cilium-ipsec-secrets\") on node \"24.144.92.253\" DevicePath \"\"" Feb 9 09:04:51.303998 kubelet[1395]: I0209 09:04:51.303977 1395 reconciler_common.go:300] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-cni-path\") on node \"24.144.92.253\" DevicePath \"\"" Feb 9 09:04:51.304121 kubelet[1395]: I0209 09:04:51.304107 1395 reconciler_common.go:300] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-hostproc\") on node \"24.144.92.253\" DevicePath \"\"" Feb 9 09:04:51.304224 kubelet[1395]: I0209 09:04:51.304212 1395 reconciler_common.go:300] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-xtables-lock\") on node \"24.144.92.253\" DevicePath \"\"" Feb 9 09:04:51.304321 kubelet[1395]: I0209 09:04:51.304309 1395 reconciler_common.go:300] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-cilium-run\") on node \"24.144.92.253\" DevicePath \"\"" Feb 9 09:04:51.304415 kubelet[1395]: I0209 09:04:51.304404 1395 reconciler_common.go:300] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/14558848-333d-4982-be68-b07522bacf32-hubble-tls\") on node \"24.144.92.253\" DevicePath \"\"" Feb 9 09:04:51.304544 kubelet[1395]: I0209 09:04:51.304530 1395 reconciler_common.go:300] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-etc-cni-netd\") on node \"24.144.92.253\" DevicePath \"\"" Feb 9 09:04:51.304661 kubelet[1395]: I0209 09:04:51.304649 1395 reconciler_common.go:300] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-bpf-maps\") on node \"24.144.92.253\" DevicePath \"\"" Feb 9 09:04:51.304768 kubelet[1395]: I0209 09:04:51.304751 1395 reconciler_common.go:300] "Volume detached for volume \"kube-api-access-9pccj\" (UniqueName: \"kubernetes.io/projected/14558848-333d-4982-be68-b07522bacf32-kube-api-access-9pccj\") on node \"24.144.92.253\" DevicePath \"\"" Feb 9 09:04:51.304897 kubelet[1395]: I0209 09:04:51.304881 1395 reconciler_common.go:300] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-lib-modules\") on node \"24.144.92.253\" DevicePath \"\"" Feb 9 09:04:51.305012 kubelet[1395]: I0209 09:04:51.304999 1395 reconciler_common.go:300] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/14558848-333d-4982-be68-b07522bacf32-clustermesh-secrets\") on node \"24.144.92.253\" DevicePath \"\"" Feb 9 09:04:51.305122 kubelet[1395]: I0209 09:04:51.305107 1395 reconciler_common.go:300] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-cilium-cgroup\") on node \"24.144.92.253\" DevicePath \"\"" Feb 9 09:04:51.305242 kubelet[1395]: I0209 09:04:51.305227 1395 reconciler_common.go:300] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-host-proc-sys-net\") on node \"24.144.92.253\" DevicePath \"\"" Feb 9 09:04:51.305355 kubelet[1395]: I0209 09:04:51.305341 1395 reconciler_common.go:300] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/14558848-333d-4982-be68-b07522bacf32-host-proc-sys-kernel\") on node \"24.144.92.253\" DevicePath \"\"" Feb 9 09:04:51.601909 systemd[1]: var-lib-kubelet-pods-14558848\x2d333d\x2d4982\x2dbe68\x2db07522bacf32-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d9pccj.mount: Deactivated successfully. Feb 9 09:04:51.602090 systemd[1]: var-lib-kubelet-pods-14558848\x2d333d\x2d4982\x2dbe68\x2db07522bacf32-volumes-kubernetes.io\x7esecret-cilium\x2dipsec\x2dsecrets.mount: Deactivated successfully. Feb 9 09:04:51.602191 systemd[1]: var-lib-kubelet-pods-14558848\x2d333d\x2d4982\x2dbe68\x2db07522bacf32-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Feb 9 09:04:51.602285 systemd[1]: var-lib-kubelet-pods-14558848\x2d333d\x2d4982\x2dbe68\x2db07522bacf32-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Feb 9 09:04:51.607538 kubelet[1395]: I0209 09:04:51.607470 1395 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/14558848-333d-4982-be68-b07522bacf32-cilium-config-path\") pod \"14558848-333d-4982-be68-b07522bacf32\" (UID: \"14558848-333d-4982-be68-b07522bacf32\") " Feb 9 09:04:51.607879 kubelet[1395]: E0209 09:04:51.607860 1395 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Feb 9 09:04:51.608841 env[1108]: time="2024-02-09T09:04:51.608789900Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6bc8ccdb58-pxqvs,Uid:e6ab66a4-cfd7-4756-9de7-aff42053eb74,Namespace:kube-system,Attempt:0,}" Feb 9 09:04:51.612202 kubelet[1395]: I0209 09:04:51.612146 1395 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14558848-333d-4982-be68-b07522bacf32-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "14558848-333d-4982-be68-b07522bacf32" (UID: "14558848-333d-4982-be68-b07522bacf32"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 9 09:04:51.631800 env[1108]: time="2024-02-09T09:04:51.631573219Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 9 09:04:51.631800 env[1108]: time="2024-02-09T09:04:51.631619413Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 9 09:04:51.631800 env[1108]: time="2024-02-09T09:04:51.631634817Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 9 09:04:51.632267 env[1108]: time="2024-02-09T09:04:51.632199416Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/710c6dcda0e88938421bb50278f2e66d15fb3f589cccdeef4f0054a4e0f62a8a pid=2937 runtime=io.containerd.runc.v2 Feb 9 09:04:51.652001 systemd[1]: Started cri-containerd-710c6dcda0e88938421bb50278f2e66d15fb3f589cccdeef4f0054a4e0f62a8a.scope. Feb 9 09:04:51.708585 kubelet[1395]: I0209 09:04:51.708523 1395 reconciler_common.go:300] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/14558848-333d-4982-be68-b07522bacf32-cilium-config-path\") on node \"24.144.92.253\" DevicePath \"\"" Feb 9 09:04:51.712903 env[1108]: time="2024-02-09T09:04:51.712854818Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6bc8ccdb58-pxqvs,Uid:e6ab66a4-cfd7-4756-9de7-aff42053eb74,Namespace:kube-system,Attempt:0,} returns sandbox id \"710c6dcda0e88938421bb50278f2e66d15fb3f589cccdeef4f0054a4e0f62a8a\"" Feb 9 09:04:51.714552 kubelet[1395]: E0209 09:04:51.713886 1395 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Feb 9 09:04:51.715447 env[1108]: time="2024-02-09T09:04:51.715406329Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Feb 9 09:04:51.857900 kubelet[1395]: E0209 09:04:51.857761 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:51.942038 systemd[1]: Removed slice kubepods-burstable-pod14558848_333d_4982_be68_b07522bacf32.slice. Feb 9 09:04:52.205420 kubelet[1395]: I0209 09:04:52.205278 1395 topology_manager.go:215] "Topology Admit Handler" podUID="3c5d1870-3829-4524-b439-fc42df3c5511" podNamespace="kube-system" podName="cilium-qwp8f" Feb 9 09:04:52.212652 systemd[1]: Created slice kubepods-burstable-pod3c5d1870_3829_4524_b439_fc42df3c5511.slice. Feb 9 09:04:52.312763 kubelet[1395]: I0209 09:04:52.312695 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/3c5d1870-3829-4524-b439-fc42df3c5511-host-proc-sys-net\") pod \"cilium-qwp8f\" (UID: \"3c5d1870-3829-4524-b439-fc42df3c5511\") " pod="kube-system/cilium-qwp8f" Feb 9 09:04:52.313008 kubelet[1395]: I0209 09:04:52.312801 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jh5k4\" (UniqueName: \"kubernetes.io/projected/3c5d1870-3829-4524-b439-fc42df3c5511-kube-api-access-jh5k4\") pod \"cilium-qwp8f\" (UID: \"3c5d1870-3829-4524-b439-fc42df3c5511\") " pod="kube-system/cilium-qwp8f" Feb 9 09:04:52.313008 kubelet[1395]: I0209 09:04:52.312835 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/3c5d1870-3829-4524-b439-fc42df3c5511-cni-path\") pod \"cilium-qwp8f\" (UID: \"3c5d1870-3829-4524-b439-fc42df3c5511\") " pod="kube-system/cilium-qwp8f" Feb 9 09:04:52.313008 kubelet[1395]: I0209 09:04:52.312863 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/3c5d1870-3829-4524-b439-fc42df3c5511-cilium-ipsec-secrets\") pod \"cilium-qwp8f\" (UID: \"3c5d1870-3829-4524-b439-fc42df3c5511\") " pod="kube-system/cilium-qwp8f" Feb 9 09:04:52.313008 kubelet[1395]: I0209 09:04:52.312895 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/3c5d1870-3829-4524-b439-fc42df3c5511-bpf-maps\") pod \"cilium-qwp8f\" (UID: \"3c5d1870-3829-4524-b439-fc42df3c5511\") " pod="kube-system/cilium-qwp8f" Feb 9 09:04:52.313008 kubelet[1395]: I0209 09:04:52.312923 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/3c5d1870-3829-4524-b439-fc42df3c5511-hostproc\") pod \"cilium-qwp8f\" (UID: \"3c5d1870-3829-4524-b439-fc42df3c5511\") " pod="kube-system/cilium-qwp8f" Feb 9 09:04:52.313008 kubelet[1395]: I0209 09:04:52.312954 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/3c5d1870-3829-4524-b439-fc42df3c5511-cilium-cgroup\") pod \"cilium-qwp8f\" (UID: \"3c5d1870-3829-4524-b439-fc42df3c5511\") " pod="kube-system/cilium-qwp8f" Feb 9 09:04:52.313297 kubelet[1395]: I0209 09:04:52.312985 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3c5d1870-3829-4524-b439-fc42df3c5511-etc-cni-netd\") pod \"cilium-qwp8f\" (UID: \"3c5d1870-3829-4524-b439-fc42df3c5511\") " pod="kube-system/cilium-qwp8f" Feb 9 09:04:52.313297 kubelet[1395]: I0209 09:04:52.313014 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/3c5d1870-3829-4524-b439-fc42df3c5511-cilium-run\") pod \"cilium-qwp8f\" (UID: \"3c5d1870-3829-4524-b439-fc42df3c5511\") " pod="kube-system/cilium-qwp8f" Feb 9 09:04:52.313297 kubelet[1395]: I0209 09:04:52.313044 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/3c5d1870-3829-4524-b439-fc42df3c5511-host-proc-sys-kernel\") pod \"cilium-qwp8f\" (UID: \"3c5d1870-3829-4524-b439-fc42df3c5511\") " pod="kube-system/cilium-qwp8f" Feb 9 09:04:52.313297 kubelet[1395]: I0209 09:04:52.313077 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/3c5d1870-3829-4524-b439-fc42df3c5511-hubble-tls\") pod \"cilium-qwp8f\" (UID: \"3c5d1870-3829-4524-b439-fc42df3c5511\") " pod="kube-system/cilium-qwp8f" Feb 9 09:04:52.313297 kubelet[1395]: I0209 09:04:52.313110 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/3c5d1870-3829-4524-b439-fc42df3c5511-xtables-lock\") pod \"cilium-qwp8f\" (UID: \"3c5d1870-3829-4524-b439-fc42df3c5511\") " pod="kube-system/cilium-qwp8f" Feb 9 09:04:52.313297 kubelet[1395]: I0209 09:04:52.313147 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/3c5d1870-3829-4524-b439-fc42df3c5511-clustermesh-secrets\") pod \"cilium-qwp8f\" (UID: \"3c5d1870-3829-4524-b439-fc42df3c5511\") " pod="kube-system/cilium-qwp8f" Feb 9 09:04:52.313662 kubelet[1395]: I0209 09:04:52.313186 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/3c5d1870-3829-4524-b439-fc42df3c5511-cilium-config-path\") pod \"cilium-qwp8f\" (UID: \"3c5d1870-3829-4524-b439-fc42df3c5511\") " pod="kube-system/cilium-qwp8f" Feb 9 09:04:52.313662 kubelet[1395]: I0209 09:04:52.313227 1395 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3c5d1870-3829-4524-b439-fc42df3c5511-lib-modules\") pod \"cilium-qwp8f\" (UID: \"3c5d1870-3829-4524-b439-fc42df3c5511\") " pod="kube-system/cilium-qwp8f" Feb 9 09:04:52.523015 kubelet[1395]: E0209 09:04:52.522867 1395 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Feb 9 09:04:52.523670 env[1108]: time="2024-02-09T09:04:52.523622019Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-qwp8f,Uid:3c5d1870-3829-4524-b439-fc42df3c5511,Namespace:kube-system,Attempt:0,}" Feb 9 09:04:52.545206 env[1108]: time="2024-02-09T09:04:52.544920338Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 9 09:04:52.545206 env[1108]: time="2024-02-09T09:04:52.544992308Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 9 09:04:52.545206 env[1108]: time="2024-02-09T09:04:52.545011349Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 9 09:04:52.545711 env[1108]: time="2024-02-09T09:04:52.545631401Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/9bd2ffd6018b1636eb1e16c0c3ded73591290da03e1bf5647e080664267a92b2 pid=2982 runtime=io.containerd.runc.v2 Feb 9 09:04:52.562054 systemd[1]: Started cri-containerd-9bd2ffd6018b1636eb1e16c0c3ded73591290da03e1bf5647e080664267a92b2.scope. Feb 9 09:04:52.626191 env[1108]: time="2024-02-09T09:04:52.626139674Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-qwp8f,Uid:3c5d1870-3829-4524-b439-fc42df3c5511,Namespace:kube-system,Attempt:0,} returns sandbox id \"9bd2ffd6018b1636eb1e16c0c3ded73591290da03e1bf5647e080664267a92b2\"" Feb 9 09:04:52.627833 kubelet[1395]: E0209 09:04:52.627748 1395 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Feb 9 09:04:52.631132 env[1108]: time="2024-02-09T09:04:52.631078693Z" level=info msg="CreateContainer within sandbox \"9bd2ffd6018b1636eb1e16c0c3ded73591290da03e1bf5647e080664267a92b2\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Feb 9 09:04:52.647951 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1913726011.mount: Deactivated successfully. Feb 9 09:04:52.658726 env[1108]: time="2024-02-09T09:04:52.658661355Z" level=info msg="CreateContainer within sandbox \"9bd2ffd6018b1636eb1e16c0c3ded73591290da03e1bf5647e080664267a92b2\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"f37b1becae265c01043285927d1ee00bd5484dbb4045a59e1cce1b272507fdce\"" Feb 9 09:04:52.660000 env[1108]: time="2024-02-09T09:04:52.659963314Z" level=info msg="StartContainer for \"f37b1becae265c01043285927d1ee00bd5484dbb4045a59e1cce1b272507fdce\"" Feb 9 09:04:52.690388 systemd[1]: Started cri-containerd-f37b1becae265c01043285927d1ee00bd5484dbb4045a59e1cce1b272507fdce.scope. Feb 9 09:04:52.739382 env[1108]: time="2024-02-09T09:04:52.739322660Z" level=info msg="StartContainer for \"f37b1becae265c01043285927d1ee00bd5484dbb4045a59e1cce1b272507fdce\" returns successfully" Feb 9 09:04:52.754073 systemd[1]: cri-containerd-f37b1becae265c01043285927d1ee00bd5484dbb4045a59e1cce1b272507fdce.scope: Deactivated successfully. Feb 9 09:04:52.803141 env[1108]: time="2024-02-09T09:04:52.802871364Z" level=info msg="shim disconnected" id=f37b1becae265c01043285927d1ee00bd5484dbb4045a59e1cce1b272507fdce Feb 9 09:04:52.803871 env[1108]: time="2024-02-09T09:04:52.803826777Z" level=warning msg="cleaning up after shim disconnected" id=f37b1becae265c01043285927d1ee00bd5484dbb4045a59e1cce1b272507fdce namespace=k8s.io Feb 9 09:04:52.804223 env[1108]: time="2024-02-09T09:04:52.804196462Z" level=info msg="cleaning up dead shim" Feb 9 09:04:52.817583 env[1108]: time="2024-02-09T09:04:52.817520620Z" level=warning msg="cleanup warnings time=\"2024-02-09T09:04:52Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3067 runtime=io.containerd.runc.v2\n" Feb 9 09:04:52.858135 kubelet[1395]: E0209 09:04:52.858069 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:53.149040 kubelet[1395]: E0209 09:04:53.148910 1395 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Feb 9 09:04:53.157630 env[1108]: time="2024-02-09T09:04:53.157567838Z" level=info msg="CreateContainer within sandbox \"9bd2ffd6018b1636eb1e16c0c3ded73591290da03e1bf5647e080664267a92b2\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Feb 9 09:04:53.172943 env[1108]: time="2024-02-09T09:04:53.172871546Z" level=info msg="CreateContainer within sandbox \"9bd2ffd6018b1636eb1e16c0c3ded73591290da03e1bf5647e080664267a92b2\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"f4d8c69f407e3dc79cefd045e9a4ed5d11acf9555d0ee9a67f13e1b1ff2c3999\"" Feb 9 09:04:53.173888 env[1108]: time="2024-02-09T09:04:53.173835094Z" level=info msg="StartContainer for \"f4d8c69f407e3dc79cefd045e9a4ed5d11acf9555d0ee9a67f13e1b1ff2c3999\"" Feb 9 09:04:53.211033 systemd[1]: Started cri-containerd-f4d8c69f407e3dc79cefd045e9a4ed5d11acf9555d0ee9a67f13e1b1ff2c3999.scope. Feb 9 09:04:53.272473 env[1108]: time="2024-02-09T09:04:53.272400821Z" level=info msg="StartContainer for \"f4d8c69f407e3dc79cefd045e9a4ed5d11acf9555d0ee9a67f13e1b1ff2c3999\" returns successfully" Feb 9 09:04:53.280300 systemd[1]: cri-containerd-f4d8c69f407e3dc79cefd045e9a4ed5d11acf9555d0ee9a67f13e1b1ff2c3999.scope: Deactivated successfully. Feb 9 09:04:53.323917 env[1108]: time="2024-02-09T09:04:53.323841534Z" level=info msg="shim disconnected" id=f4d8c69f407e3dc79cefd045e9a4ed5d11acf9555d0ee9a67f13e1b1ff2c3999 Feb 9 09:04:53.323917 env[1108]: time="2024-02-09T09:04:53.323903698Z" level=warning msg="cleaning up after shim disconnected" id=f4d8c69f407e3dc79cefd045e9a4ed5d11acf9555d0ee9a67f13e1b1ff2c3999 namespace=k8s.io Feb 9 09:04:53.323917 env[1108]: time="2024-02-09T09:04:53.323916249Z" level=info msg="cleaning up dead shim" Feb 9 09:04:53.337522 env[1108]: time="2024-02-09T09:04:53.337463241Z" level=warning msg="cleanup warnings time=\"2024-02-09T09:04:53Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3130 runtime=io.containerd.runc.v2\n" Feb 9 09:04:53.602110 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-f37b1becae265c01043285927d1ee00bd5484dbb4045a59e1cce1b272507fdce-rootfs.mount: Deactivated successfully. Feb 9 09:04:53.858558 kubelet[1395]: E0209 09:04:53.858401 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:53.938081 kubelet[1395]: I0209 09:04:53.938039 1395 kubelet_volumes.go:161] "Cleaned up orphaned pod volumes dir" podUID="14558848-333d-4982-be68-b07522bacf32" path="/var/lib/kubelet/pods/14558848-333d-4982-be68-b07522bacf32/volumes" Feb 9 09:04:54.139757 env[1108]: time="2024-02-09T09:04:54.139571766Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 09:04:54.143247 env[1108]: time="2024-02-09T09:04:54.143183842Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 09:04:54.146925 env[1108]: time="2024-02-09T09:04:54.146868358Z" level=info msg="ImageUpdate event &ImageUpdate{Name:quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 09:04:54.148115 env[1108]: time="2024-02-09T09:04:54.148070964Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"" Feb 9 09:04:54.151370 env[1108]: time="2024-02-09T09:04:54.151310560Z" level=info msg="CreateContainer within sandbox \"710c6dcda0e88938421bb50278f2e66d15fb3f589cccdeef4f0054a4e0f62a8a\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Feb 9 09:04:54.154304 kubelet[1395]: E0209 09:04:54.153527 1395 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Feb 9 09:04:54.156383 env[1108]: time="2024-02-09T09:04:54.156340589Z" level=info msg="CreateContainer within sandbox \"9bd2ffd6018b1636eb1e16c0c3ded73591290da03e1bf5647e080664267a92b2\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Feb 9 09:04:54.190693 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3452754743.mount: Deactivated successfully. Feb 9 09:04:54.195462 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1022142231.mount: Deactivated successfully. Feb 9 09:04:54.210078 env[1108]: time="2024-02-09T09:04:54.209994495Z" level=info msg="CreateContainer within sandbox \"710c6dcda0e88938421bb50278f2e66d15fb3f589cccdeef4f0054a4e0f62a8a\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"16cb5df085f2896f8f1c5afb673999558587498973dcd1fd00ec7b64c9f2359a\"" Feb 9 09:04:54.211152 env[1108]: time="2024-02-09T09:04:54.211101966Z" level=info msg="StartContainer for \"16cb5df085f2896f8f1c5afb673999558587498973dcd1fd00ec7b64c9f2359a\"" Feb 9 09:04:54.215059 env[1108]: time="2024-02-09T09:04:54.215000876Z" level=info msg="CreateContainer within sandbox \"9bd2ffd6018b1636eb1e16c0c3ded73591290da03e1bf5647e080664267a92b2\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"e362c9905651c4b2d01d2c7561b844f9e0a81f9dc74f93737d6d43a5be64bc0f\"" Feb 9 09:04:54.216053 env[1108]: time="2024-02-09T09:04:54.216013704Z" level=info msg="StartContainer for \"e362c9905651c4b2d01d2c7561b844f9e0a81f9dc74f93737d6d43a5be64bc0f\"" Feb 9 09:04:54.243787 systemd[1]: Started cri-containerd-16cb5df085f2896f8f1c5afb673999558587498973dcd1fd00ec7b64c9f2359a.scope. Feb 9 09:04:54.267615 systemd[1]: Started cri-containerd-e362c9905651c4b2d01d2c7561b844f9e0a81f9dc74f93737d6d43a5be64bc0f.scope. Feb 9 09:04:54.316991 env[1108]: time="2024-02-09T09:04:54.316921152Z" level=info msg="StartContainer for \"16cb5df085f2896f8f1c5afb673999558587498973dcd1fd00ec7b64c9f2359a\" returns successfully" Feb 9 09:04:54.337583 env[1108]: time="2024-02-09T09:04:54.337524963Z" level=info msg="StartContainer for \"e362c9905651c4b2d01d2c7561b844f9e0a81f9dc74f93737d6d43a5be64bc0f\" returns successfully" Feb 9 09:04:54.347083 systemd[1]: cri-containerd-e362c9905651c4b2d01d2c7561b844f9e0a81f9dc74f93737d6d43a5be64bc0f.scope: Deactivated successfully. Feb 9 09:04:54.445370 env[1108]: time="2024-02-09T09:04:54.445201805Z" level=info msg="shim disconnected" id=e362c9905651c4b2d01d2c7561b844f9e0a81f9dc74f93737d6d43a5be64bc0f Feb 9 09:04:54.445753 env[1108]: time="2024-02-09T09:04:54.445717240Z" level=warning msg="cleaning up after shim disconnected" id=e362c9905651c4b2d01d2c7561b844f9e0a81f9dc74f93737d6d43a5be64bc0f namespace=k8s.io Feb 9 09:04:54.445912 env[1108]: time="2024-02-09T09:04:54.445888385Z" level=info msg="cleaning up dead shim" Feb 9 09:04:54.464971 env[1108]: time="2024-02-09T09:04:54.464781003Z" level=warning msg="cleanup warnings time=\"2024-02-09T09:04:54Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3227 runtime=io.containerd.runc.v2\n" Feb 9 09:04:54.859157 kubelet[1395]: E0209 09:04:54.859076 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:55.158556 kubelet[1395]: E0209 09:04:55.157586 1395 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Feb 9 09:04:55.161559 kubelet[1395]: E0209 09:04:55.161475 1395 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Feb 9 09:04:55.164995 env[1108]: time="2024-02-09T09:04:55.164933518Z" level=info msg="CreateContainer within sandbox \"9bd2ffd6018b1636eb1e16c0c3ded73591290da03e1bf5647e080664267a92b2\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Feb 9 09:04:55.188875 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3772909262.mount: Deactivated successfully. Feb 9 09:04:55.197595 env[1108]: time="2024-02-09T09:04:55.197464205Z" level=info msg="CreateContainer within sandbox \"9bd2ffd6018b1636eb1e16c0c3ded73591290da03e1bf5647e080664267a92b2\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"5ceae40451e4c27e86d2c8b0d637d115ca4e86d2a3261341396aeb7ef46a73e2\"" Feb 9 09:04:55.199121 env[1108]: time="2024-02-09T09:04:55.199076048Z" level=info msg="StartContainer for \"5ceae40451e4c27e86d2c8b0d637d115ca4e86d2a3261341396aeb7ef46a73e2\"" Feb 9 09:04:55.200725 kubelet[1395]: I0209 09:04:55.200687 1395 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/cilium-operator-6bc8ccdb58-pxqvs" podStartSLOduration=2.766895344 podCreationTimestamp="2024-02-09 09:04:50 +0000 UTC" firstStartedPulling="2024-02-09 09:04:51.714919682 +0000 UTC m=+76.528321831" lastFinishedPulling="2024-02-09 09:04:54.148646606 +0000 UTC m=+78.962048964" observedRunningTime="2024-02-09 09:04:55.170554463 +0000 UTC m=+79.983956630" watchObservedRunningTime="2024-02-09 09:04:55.200622477 +0000 UTC m=+80.014024643" Feb 9 09:04:55.230360 systemd[1]: Started cri-containerd-5ceae40451e4c27e86d2c8b0d637d115ca4e86d2a3261341396aeb7ef46a73e2.scope. Feb 9 09:04:55.282745 systemd[1]: cri-containerd-5ceae40451e4c27e86d2c8b0d637d115ca4e86d2a3261341396aeb7ef46a73e2.scope: Deactivated successfully. Feb 9 09:04:55.287445 env[1108]: time="2024-02-09T09:04:55.287396760Z" level=info msg="StartContainer for \"5ceae40451e4c27e86d2c8b0d637d115ca4e86d2a3261341396aeb7ef46a73e2\" returns successfully" Feb 9 09:04:55.315747 env[1108]: time="2024-02-09T09:04:55.315674605Z" level=info msg="shim disconnected" id=5ceae40451e4c27e86d2c8b0d637d115ca4e86d2a3261341396aeb7ef46a73e2 Feb 9 09:04:55.315747 env[1108]: time="2024-02-09T09:04:55.315739839Z" level=warning msg="cleaning up after shim disconnected" id=5ceae40451e4c27e86d2c8b0d637d115ca4e86d2a3261341396aeb7ef46a73e2 namespace=k8s.io Feb 9 09:04:55.316089 env[1108]: time="2024-02-09T09:04:55.315757873Z" level=info msg="cleaning up dead shim" Feb 9 09:04:55.326545 env[1108]: time="2024-02-09T09:04:55.326474733Z" level=warning msg="cleanup warnings time=\"2024-02-09T09:04:55Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3282 runtime=io.containerd.runc.v2\n" Feb 9 09:04:55.602293 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-5ceae40451e4c27e86d2c8b0d637d115ca4e86d2a3261341396aeb7ef46a73e2-rootfs.mount: Deactivated successfully. Feb 9 09:04:55.791263 kubelet[1395]: E0209 09:04:55.791143 1395 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:55.860412 kubelet[1395]: E0209 09:04:55.860243 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:55.911646 kubelet[1395]: E0209 09:04:55.911602 1395 kubelet.go:2855] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Feb 9 09:04:56.166664 kubelet[1395]: E0209 09:04:56.166541 1395 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Feb 9 09:04:56.166864 kubelet[1395]: E0209 09:04:56.166550 1395 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Feb 9 09:04:56.169332 env[1108]: time="2024-02-09T09:04:56.169276334Z" level=info msg="CreateContainer within sandbox \"9bd2ffd6018b1636eb1e16c0c3ded73591290da03e1bf5647e080664267a92b2\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Feb 9 09:04:56.197555 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2669070433.mount: Deactivated successfully. Feb 9 09:04:56.204937 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2971149487.mount: Deactivated successfully. Feb 9 09:04:56.211536 env[1108]: time="2024-02-09T09:04:56.211439576Z" level=info msg="CreateContainer within sandbox \"9bd2ffd6018b1636eb1e16c0c3ded73591290da03e1bf5647e080664267a92b2\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"9381823b963ca880077348da2a14027a5ae832b5819a41f23d13a07a1261ca77\"" Feb 9 09:04:56.212761 env[1108]: time="2024-02-09T09:04:56.212716553Z" level=info msg="StartContainer for \"9381823b963ca880077348da2a14027a5ae832b5819a41f23d13a07a1261ca77\"" Feb 9 09:04:56.236259 systemd[1]: Started cri-containerd-9381823b963ca880077348da2a14027a5ae832b5819a41f23d13a07a1261ca77.scope. Feb 9 09:04:56.290446 env[1108]: time="2024-02-09T09:04:56.290384299Z" level=info msg="StartContainer for \"9381823b963ca880077348da2a14027a5ae832b5819a41f23d13a07a1261ca77\" returns successfully" Feb 9 09:04:56.840546 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aesni)) Feb 9 09:04:56.861093 kubelet[1395]: E0209 09:04:56.861031 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:57.174563 kubelet[1395]: E0209 09:04:57.173547 1395 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Feb 9 09:04:57.861698 kubelet[1395]: E0209 09:04:57.861643 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:58.526089 kubelet[1395]: E0209 09:04:58.526036 1395 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Feb 9 09:04:58.714192 kubelet[1395]: I0209 09:04:58.714157 1395 setters.go:552] "Node became not ready" node="24.144.92.253" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2024-02-09T09:04:58Z","lastTransitionTime":"2024-02-09T09:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} Feb 9 09:04:58.863568 kubelet[1395]: E0209 09:04:58.863403 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:59.069207 systemd[1]: run-containerd-runc-k8s.io-9381823b963ca880077348da2a14027a5ae832b5819a41f23d13a07a1261ca77-runc.mB5EWb.mount: Deactivated successfully. Feb 9 09:04:59.864701 kubelet[1395]: E0209 09:04:59.864661 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:04:59.933430 systemd-networkd[1002]: lxc_health: Link UP Feb 9 09:04:59.959751 systemd-networkd[1002]: lxc_health: Gained carrier Feb 9 09:04:59.960613 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc_health: link becomes ready Feb 9 09:05:00.527382 kubelet[1395]: E0209 09:05:00.527331 1395 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Feb 9 09:05:00.553111 kubelet[1395]: I0209 09:05:00.553047 1395 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/cilium-qwp8f" podStartSLOduration=8.552965781 podCreationTimestamp="2024-02-09 09:04:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-02-09 09:04:57.195289283 +0000 UTC m=+82.008691467" watchObservedRunningTime="2024-02-09 09:05:00.552965781 +0000 UTC m=+85.366367946" Feb 9 09:05:00.865649 kubelet[1395]: E0209 09:05:00.865437 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:05:01.187409 kubelet[1395]: E0209 09:05:01.187266 1395 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Feb 9 09:05:01.302457 systemd[1]: run-containerd-runc-k8s.io-9381823b963ca880077348da2a14027a5ae832b5819a41f23d13a07a1261ca77-runc.DcB2ch.mount: Deactivated successfully. Feb 9 09:05:01.591784 systemd-networkd[1002]: lxc_health: Gained IPv6LL Feb 9 09:05:01.866733 kubelet[1395]: E0209 09:05:01.866578 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:05:02.189903 kubelet[1395]: E0209 09:05:02.189770 1395 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Feb 9 09:05:02.867972 kubelet[1395]: E0209 09:05:02.867924 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:05:03.697753 systemd[1]: run-containerd-runc-k8s.io-9381823b963ca880077348da2a14027a5ae832b5819a41f23d13a07a1261ca77-runc.E0JD9o.mount: Deactivated successfully. Feb 9 09:05:03.869405 kubelet[1395]: E0209 09:05:03.869298 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:05:04.870299 kubelet[1395]: E0209 09:05:04.870237 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:05:05.871272 kubelet[1395]: E0209 09:05:05.871208 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:05:05.980424 systemd[1]: run-containerd-runc-k8s.io-9381823b963ca880077348da2a14027a5ae832b5819a41f23d13a07a1261ca77-runc.4Tf91a.mount: Deactivated successfully. Feb 9 09:05:06.872068 kubelet[1395]: E0209 09:05:06.872009 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 9 09:05:07.873223 kubelet[1395]: E0209 09:05:07.873167 1395 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests"