Dec 13 02:40:01.097518 kernel: Linux version 6.6.65-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Thu Dec 12 23:15:00 -00 2024 Dec 13 02:40:01.097561 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=openstack flatcar.autologin verity.usrhash=2fdbba50b59d8c8a9877a81151806ddc16f473fe99b9ba0d8825997d654583ff Dec 13 02:40:01.097573 kernel: BIOS-provided physical RAM map: Dec 13 02:40:01.097580 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable Dec 13 02:40:01.097587 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved Dec 13 02:40:01.097594 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Dec 13 02:40:01.097602 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000007ffdcfff] usable Dec 13 02:40:01.097610 kernel: BIOS-e820: [mem 0x000000007ffdd000-0x000000007fffffff] reserved Dec 13 02:40:01.097617 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Dec 13 02:40:01.097626 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Dec 13 02:40:01.097633 kernel: NX (Execute Disable) protection: active Dec 13 02:40:01.097640 kernel: APIC: Static calls initialized Dec 13 02:40:01.097647 kernel: SMBIOS 2.8 present. Dec 13 02:40:01.097655 kernel: DMI: OpenStack Foundation OpenStack Nova, BIOS 1.15.0-1 04/01/2014 Dec 13 02:40:01.097664 kernel: Hypervisor detected: KVM Dec 13 02:40:01.097674 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Dec 13 02:40:01.097682 kernel: kvm-clock: using sched offset of 6583792520 cycles Dec 13 02:40:01.097689 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Dec 13 02:40:01.097697 kernel: tsc: Detected 1996.249 MHz processor Dec 13 02:40:01.097705 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Dec 13 02:40:01.097713 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Dec 13 02:40:01.097721 kernel: last_pfn = 0x7ffdd max_arch_pfn = 0x400000000 Dec 13 02:40:01.097729 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs Dec 13 02:40:01.097737 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Dec 13 02:40:01.097747 kernel: ACPI: Early table checksum verification disabled Dec 13 02:40:01.097755 kernel: ACPI: RSDP 0x00000000000F5930 000014 (v00 BOCHS ) Dec 13 02:40:01.097763 kernel: ACPI: RSDT 0x000000007FFE1848 000030 (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 02:40:01.097771 kernel: ACPI: FACP 0x000000007FFE172C 000074 (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 02:40:01.097779 kernel: ACPI: DSDT 0x000000007FFE0040 0016EC (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 02:40:01.097786 kernel: ACPI: FACS 0x000000007FFE0000 000040 Dec 13 02:40:01.097794 kernel: ACPI: APIC 0x000000007FFE17A0 000080 (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 02:40:01.097802 kernel: ACPI: WAET 0x000000007FFE1820 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 02:40:01.097810 kernel: ACPI: Reserving FACP table memory at [mem 0x7ffe172c-0x7ffe179f] Dec 13 02:40:01.097820 kernel: ACPI: Reserving DSDT table memory at [mem 0x7ffe0040-0x7ffe172b] Dec 13 02:40:01.097827 kernel: ACPI: Reserving FACS table memory at [mem 0x7ffe0000-0x7ffe003f] Dec 13 02:40:01.097835 kernel: ACPI: Reserving APIC table memory at [mem 0x7ffe17a0-0x7ffe181f] Dec 13 02:40:01.097843 kernel: ACPI: Reserving WAET table memory at [mem 0x7ffe1820-0x7ffe1847] Dec 13 02:40:01.097850 kernel: No NUMA configuration found Dec 13 02:40:01.097858 kernel: Faking a node at [mem 0x0000000000000000-0x000000007ffdcfff] Dec 13 02:40:01.097866 kernel: NODE_DATA(0) allocated [mem 0x7ffd7000-0x7ffdcfff] Dec 13 02:40:01.097877 kernel: Zone ranges: Dec 13 02:40:01.097887 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Dec 13 02:40:01.099919 kernel: DMA32 [mem 0x0000000001000000-0x000000007ffdcfff] Dec 13 02:40:01.099933 kernel: Normal empty Dec 13 02:40:01.099956 kernel: Movable zone start for each node Dec 13 02:40:01.099966 kernel: Early memory node ranges Dec 13 02:40:01.099974 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Dec 13 02:40:01.099983 kernel: node 0: [mem 0x0000000000100000-0x000000007ffdcfff] Dec 13 02:40:01.099996 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000007ffdcfff] Dec 13 02:40:01.100005 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Dec 13 02:40:01.100013 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Dec 13 02:40:01.100022 kernel: On node 0, zone DMA32: 35 pages in unavailable ranges Dec 13 02:40:01.100031 kernel: ACPI: PM-Timer IO Port: 0x608 Dec 13 02:40:01.100039 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Dec 13 02:40:01.100048 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Dec 13 02:40:01.100057 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Dec 13 02:40:01.100065 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Dec 13 02:40:01.100076 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Dec 13 02:40:01.100085 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Dec 13 02:40:01.100093 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Dec 13 02:40:01.100102 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Dec 13 02:40:01.100111 kernel: smpboot: Allowing 2 CPUs, 0 hotplug CPUs Dec 13 02:40:01.100120 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Dec 13 02:40:01.100128 kernel: [mem 0x80000000-0xfeffbfff] available for PCI devices Dec 13 02:40:01.100137 kernel: Booting paravirtualized kernel on KVM Dec 13 02:40:01.100146 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Dec 13 02:40:01.100157 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1 Dec 13 02:40:01.100166 kernel: percpu: Embedded 58 pages/cpu s197032 r8192 d32344 u1048576 Dec 13 02:40:01.100175 kernel: pcpu-alloc: s197032 r8192 d32344 u1048576 alloc=1*2097152 Dec 13 02:40:01.100183 kernel: pcpu-alloc: [0] 0 1 Dec 13 02:40:01.100192 kernel: kvm-guest: PV spinlocks disabled, no host support Dec 13 02:40:01.100202 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=openstack flatcar.autologin verity.usrhash=2fdbba50b59d8c8a9877a81151806ddc16f473fe99b9ba0d8825997d654583ff Dec 13 02:40:01.100212 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Dec 13 02:40:01.100221 kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Dec 13 02:40:01.100232 kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear) Dec 13 02:40:01.100241 kernel: Fallback order for Node 0: 0 Dec 13 02:40:01.100249 kernel: Built 1 zonelists, mobility grouping on. Total pages: 515805 Dec 13 02:40:01.100258 kernel: Policy zone: DMA32 Dec 13 02:40:01.100267 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Dec 13 02:40:01.100276 kernel: Memory: 1971212K/2096620K available (12288K kernel code, 2299K rwdata, 22724K rodata, 42844K init, 2348K bss, 125148K reserved, 0K cma-reserved) Dec 13 02:40:01.100284 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Dec 13 02:40:01.100293 kernel: ftrace: allocating 37902 entries in 149 pages Dec 13 02:40:01.100304 kernel: ftrace: allocated 149 pages with 4 groups Dec 13 02:40:01.100312 kernel: Dynamic Preempt: voluntary Dec 13 02:40:01.100321 kernel: rcu: Preemptible hierarchical RCU implementation. Dec 13 02:40:01.100331 kernel: rcu: RCU event tracing is enabled. Dec 13 02:40:01.100340 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Dec 13 02:40:01.100349 kernel: Trampoline variant of Tasks RCU enabled. Dec 13 02:40:01.100358 kernel: Rude variant of Tasks RCU enabled. Dec 13 02:40:01.100367 kernel: Tracing variant of Tasks RCU enabled. Dec 13 02:40:01.100376 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Dec 13 02:40:01.100384 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Dec 13 02:40:01.100395 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 Dec 13 02:40:01.100404 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Dec 13 02:40:01.100412 kernel: Console: colour VGA+ 80x25 Dec 13 02:40:01.100421 kernel: printk: console [tty0] enabled Dec 13 02:40:01.100430 kernel: printk: console [ttyS0] enabled Dec 13 02:40:01.100438 kernel: ACPI: Core revision 20230628 Dec 13 02:40:01.100447 kernel: APIC: Switch to symmetric I/O mode setup Dec 13 02:40:01.100456 kernel: x2apic enabled Dec 13 02:40:01.100464 kernel: APIC: Switched APIC routing to: physical x2apic Dec 13 02:40:01.100475 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Dec 13 02:40:01.100484 kernel: tsc: Marking TSC unstable due to TSCs unsynchronized Dec 13 02:40:01.100493 kernel: Calibrating delay loop (skipped) preset value.. 3992.49 BogoMIPS (lpj=1996249) Dec 13 02:40:01.100502 kernel: Last level iTLB entries: 4KB 0, 2MB 0, 4MB 0 Dec 13 02:40:01.100510 kernel: Last level dTLB entries: 4KB 0, 2MB 0, 4MB 0, 1GB 0 Dec 13 02:40:01.100520 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Dec 13 02:40:01.100528 kernel: Spectre V2 : Mitigation: Retpolines Dec 13 02:40:01.100537 kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch Dec 13 02:40:01.100546 kernel: Spectre V2 : Spectre v2 / SpectreRSB : Filling RSB on VMEXIT Dec 13 02:40:01.100557 kernel: Speculative Store Bypass: Vulnerable Dec 13 02:40:01.100566 kernel: x86/fpu: x87 FPU will use FXSAVE Dec 13 02:40:01.100574 kernel: Freeing SMP alternatives memory: 32K Dec 13 02:40:01.100583 kernel: pid_max: default: 32768 minimum: 301 Dec 13 02:40:01.100592 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Dec 13 02:40:01.100601 kernel: landlock: Up and running. Dec 13 02:40:01.100609 kernel: SELinux: Initializing. Dec 13 02:40:01.100619 kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Dec 13 02:40:01.100635 kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Dec 13 02:40:01.100645 kernel: smpboot: CPU0: AMD Intel Core i7 9xx (Nehalem Class Core i7) (family: 0x6, model: 0x1a, stepping: 0x3) Dec 13 02:40:01.100654 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Dec 13 02:40:01.100665 kernel: RCU Tasks Rude: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Dec 13 02:40:01.100675 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Dec 13 02:40:01.100684 kernel: Performance Events: AMD PMU driver. Dec 13 02:40:01.100693 kernel: ... version: 0 Dec 13 02:40:01.100703 kernel: ... bit width: 48 Dec 13 02:40:01.100713 kernel: ... generic registers: 4 Dec 13 02:40:01.100722 kernel: ... value mask: 0000ffffffffffff Dec 13 02:40:01.100732 kernel: ... max period: 00007fffffffffff Dec 13 02:40:01.100741 kernel: ... fixed-purpose events: 0 Dec 13 02:40:01.100750 kernel: ... event mask: 000000000000000f Dec 13 02:40:01.100759 kernel: signal: max sigframe size: 1440 Dec 13 02:40:01.100768 kernel: rcu: Hierarchical SRCU implementation. Dec 13 02:40:01.100778 kernel: rcu: Max phase no-delay instances is 400. Dec 13 02:40:01.100787 kernel: smp: Bringing up secondary CPUs ... Dec 13 02:40:01.100796 kernel: smpboot: x86: Booting SMP configuration: Dec 13 02:40:01.100807 kernel: .... node #0, CPUs: #1 Dec 13 02:40:01.100816 kernel: smp: Brought up 1 node, 2 CPUs Dec 13 02:40:01.100825 kernel: smpboot: Max logical packages: 2 Dec 13 02:40:01.100834 kernel: smpboot: Total of 2 processors activated (7984.99 BogoMIPS) Dec 13 02:40:01.100843 kernel: devtmpfs: initialized Dec 13 02:40:01.100852 kernel: x86/mm: Memory block size: 128MB Dec 13 02:40:01.100862 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Dec 13 02:40:01.100871 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Dec 13 02:40:01.100881 kernel: pinctrl core: initialized pinctrl subsystem Dec 13 02:40:01.100892 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Dec 13 02:40:01.100920 kernel: audit: initializing netlink subsys (disabled) Dec 13 02:40:01.100929 kernel: thermal_sys: Registered thermal governor 'step_wise' Dec 13 02:40:01.100938 kernel: thermal_sys: Registered thermal governor 'user_space' Dec 13 02:40:01.101933 kernel: audit: type=2000 audit(1734057599.563:1): state=initialized audit_enabled=0 res=1 Dec 13 02:40:01.101945 kernel: cpuidle: using governor menu Dec 13 02:40:01.101953 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Dec 13 02:40:01.101962 kernel: dca service started, version 1.12.1 Dec 13 02:40:01.101971 kernel: PCI: Using configuration type 1 for base access Dec 13 02:40:01.101984 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Dec 13 02:40:01.101993 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Dec 13 02:40:01.102001 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Dec 13 02:40:01.102010 kernel: ACPI: Added _OSI(Module Device) Dec 13 02:40:01.102018 kernel: ACPI: Added _OSI(Processor Device) Dec 13 02:40:01.102027 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Dec 13 02:40:01.102036 kernel: ACPI: Added _OSI(Processor Aggregator Device) Dec 13 02:40:01.102048 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Dec 13 02:40:01.102070 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Dec 13 02:40:01.102089 kernel: ACPI: Interpreter enabled Dec 13 02:40:01.102101 kernel: ACPI: PM: (supports S0 S3 S5) Dec 13 02:40:01.102112 kernel: ACPI: Using IOAPIC for interrupt routing Dec 13 02:40:01.102127 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Dec 13 02:40:01.102140 kernel: PCI: Using E820 reservations for host bridge windows Dec 13 02:40:01.102153 kernel: ACPI: Enabled 2 GPEs in block 00 to 0F Dec 13 02:40:01.102164 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Dec 13 02:40:01.102420 kernel: acpi PNP0A03:00: _OSC: OS supports [ASPM ClockPM Segments MSI HPX-Type3] Dec 13 02:40:01.102569 kernel: acpi PNP0A03:00: _OSC: not requesting OS control; OS requires [ExtendedConfig ASPM ClockPM MSI] Dec 13 02:40:01.102700 kernel: acpi PNP0A03:00: fail to add MMCONFIG information, can't access extended configuration space under this bridge Dec 13 02:40:01.102721 kernel: acpiphp: Slot [3] registered Dec 13 02:40:01.102736 kernel: acpiphp: Slot [4] registered Dec 13 02:40:01.102748 kernel: acpiphp: Slot [5] registered Dec 13 02:40:01.102760 kernel: acpiphp: Slot [6] registered Dec 13 02:40:01.102773 kernel: acpiphp: Slot [7] registered Dec 13 02:40:01.102786 kernel: acpiphp: Slot [8] registered Dec 13 02:40:01.102804 kernel: acpiphp: Slot [9] registered Dec 13 02:40:01.102818 kernel: acpiphp: Slot [10] registered Dec 13 02:40:01.102831 kernel: acpiphp: Slot [11] registered Dec 13 02:40:01.102845 kernel: acpiphp: Slot [12] registered Dec 13 02:40:01.102858 kernel: acpiphp: Slot [13] registered Dec 13 02:40:01.102871 kernel: acpiphp: Slot [14] registered Dec 13 02:40:01.102885 kernel: acpiphp: Slot [15] registered Dec 13 02:40:01.103044 kernel: acpiphp: Slot [16] registered Dec 13 02:40:01.103061 kernel: acpiphp: Slot [17] registered Dec 13 02:40:01.103075 kernel: acpiphp: Slot [18] registered Dec 13 02:40:01.103089 kernel: acpiphp: Slot [19] registered Dec 13 02:40:01.103102 kernel: acpiphp: Slot [20] registered Dec 13 02:40:01.103116 kernel: acpiphp: Slot [21] registered Dec 13 02:40:01.103129 kernel: acpiphp: Slot [22] registered Dec 13 02:40:01.103143 kernel: acpiphp: Slot [23] registered Dec 13 02:40:01.103157 kernel: acpiphp: Slot [24] registered Dec 13 02:40:01.103170 kernel: acpiphp: Slot [25] registered Dec 13 02:40:01.103201 kernel: acpiphp: Slot [26] registered Dec 13 02:40:01.103214 kernel: acpiphp: Slot [27] registered Dec 13 02:40:01.103232 kernel: acpiphp: Slot [28] registered Dec 13 02:40:01.103243 kernel: acpiphp: Slot [29] registered Dec 13 02:40:01.103256 kernel: acpiphp: Slot [30] registered Dec 13 02:40:01.103269 kernel: acpiphp: Slot [31] registered Dec 13 02:40:01.103283 kernel: PCI host bridge to bus 0000:00 Dec 13 02:40:01.103440 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Dec 13 02:40:01.103577 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Dec 13 02:40:01.103709 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Dec 13 02:40:01.105975 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xfebfffff window] Dec 13 02:40:01.106075 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x17fffffff window] Dec 13 02:40:01.106157 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Dec 13 02:40:01.106277 kernel: pci 0000:00:00.0: [8086:1237] type 00 class 0x060000 Dec 13 02:40:01.106381 kernel: pci 0000:00:01.0: [8086:7000] type 00 class 0x060100 Dec 13 02:40:01.106482 kernel: pci 0000:00:01.1: [8086:7010] type 00 class 0x010180 Dec 13 02:40:01.106580 kernel: pci 0000:00:01.1: reg 0x20: [io 0xc120-0xc12f] Dec 13 02:40:01.106670 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x10: [io 0x01f0-0x01f7] Dec 13 02:40:01.106759 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x14: [io 0x03f6] Dec 13 02:40:01.106850 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x18: [io 0x0170-0x0177] Dec 13 02:40:01.106992 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x1c: [io 0x0376] Dec 13 02:40:01.107098 kernel: pci 0000:00:01.3: [8086:7113] type 00 class 0x068000 Dec 13 02:40:01.107253 kernel: pci 0000:00:01.3: quirk: [io 0x0600-0x063f] claimed by PIIX4 ACPI Dec 13 02:40:01.107354 kernel: pci 0000:00:01.3: quirk: [io 0x0700-0x070f] claimed by PIIX4 SMB Dec 13 02:40:01.107459 kernel: pci 0000:00:02.0: [1af4:1050] type 00 class 0x030000 Dec 13 02:40:01.107553 kernel: pci 0000:00:02.0: reg 0x10: [mem 0xfe000000-0xfe7fffff pref] Dec 13 02:40:01.107647 kernel: pci 0000:00:02.0: reg 0x18: [mem 0xfe800000-0xfe803fff 64bit pref] Dec 13 02:40:01.107742 kernel: pci 0000:00:02.0: reg 0x20: [mem 0xfeb90000-0xfeb90fff] Dec 13 02:40:01.107841 kernel: pci 0000:00:02.0: reg 0x30: [mem 0xfeb80000-0xfeb8ffff pref] Dec 13 02:40:01.109952 kernel: pci 0000:00:02.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Dec 13 02:40:01.110068 kernel: pci 0000:00:03.0: [1af4:1000] type 00 class 0x020000 Dec 13 02:40:01.110160 kernel: pci 0000:00:03.0: reg 0x10: [io 0xc080-0xc0bf] Dec 13 02:40:01.110251 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xfeb91000-0xfeb91fff] Dec 13 02:40:01.110340 kernel: pci 0000:00:03.0: reg 0x20: [mem 0xfe804000-0xfe807fff 64bit pref] Dec 13 02:40:01.110428 kernel: pci 0000:00:03.0: reg 0x30: [mem 0xfeb00000-0xfeb7ffff pref] Dec 13 02:40:01.110525 kernel: pci 0000:00:04.0: [1af4:1001] type 00 class 0x010000 Dec 13 02:40:01.110621 kernel: pci 0000:00:04.0: reg 0x10: [io 0xc000-0xc07f] Dec 13 02:40:01.110710 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xfeb92000-0xfeb92fff] Dec 13 02:40:01.110798 kernel: pci 0000:00:04.0: reg 0x20: [mem 0xfe808000-0xfe80bfff 64bit pref] Dec 13 02:40:01.110911 kernel: pci 0000:00:05.0: [1af4:1002] type 00 class 0x00ff00 Dec 13 02:40:01.111030 kernel: pci 0000:00:05.0: reg 0x10: [io 0xc0c0-0xc0ff] Dec 13 02:40:01.111121 kernel: pci 0000:00:05.0: reg 0x20: [mem 0xfe80c000-0xfe80ffff 64bit pref] Dec 13 02:40:01.111234 kernel: pci 0000:00:06.0: [1af4:1005] type 00 class 0x00ff00 Dec 13 02:40:01.111335 kernel: pci 0000:00:06.0: reg 0x10: [io 0xc100-0xc11f] Dec 13 02:40:01.111424 kernel: pci 0000:00:06.0: reg 0x20: [mem 0xfe810000-0xfe813fff 64bit pref] Dec 13 02:40:01.111437 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Dec 13 02:40:01.111446 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Dec 13 02:40:01.111455 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Dec 13 02:40:01.111463 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Dec 13 02:40:01.111472 kernel: ACPI: PCI: Interrupt link LNKS configured for IRQ 9 Dec 13 02:40:01.111481 kernel: iommu: Default domain type: Translated Dec 13 02:40:01.111490 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Dec 13 02:40:01.111502 kernel: PCI: Using ACPI for IRQ routing Dec 13 02:40:01.111511 kernel: PCI: pci_cache_line_size set to 64 bytes Dec 13 02:40:01.111519 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] Dec 13 02:40:01.111529 kernel: e820: reserve RAM buffer [mem 0x7ffdd000-0x7fffffff] Dec 13 02:40:01.111617 kernel: pci 0000:00:02.0: vgaarb: setting as boot VGA device Dec 13 02:40:01.111708 kernel: pci 0000:00:02.0: vgaarb: bridge control possible Dec 13 02:40:01.111796 kernel: pci 0000:00:02.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Dec 13 02:40:01.111809 kernel: vgaarb: loaded Dec 13 02:40:01.111818 kernel: clocksource: Switched to clocksource kvm-clock Dec 13 02:40:01.111832 kernel: VFS: Disk quotas dquot_6.6.0 Dec 13 02:40:01.111840 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Dec 13 02:40:01.111849 kernel: pnp: PnP ACPI init Dec 13 02:40:01.113004 kernel: pnp 00:03: [dma 2] Dec 13 02:40:01.113022 kernel: pnp: PnP ACPI: found 5 devices Dec 13 02:40:01.113032 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Dec 13 02:40:01.113040 kernel: NET: Registered PF_INET protocol family Dec 13 02:40:01.113049 kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear) Dec 13 02:40:01.113062 kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear) Dec 13 02:40:01.113071 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Dec 13 02:40:01.113080 kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear) Dec 13 02:40:01.113088 kernel: TCP bind hash table entries: 16384 (order: 7, 524288 bytes, linear) Dec 13 02:40:01.113098 kernel: TCP: Hash tables configured (established 16384 bind 16384) Dec 13 02:40:01.113106 kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear) Dec 13 02:40:01.113115 kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear) Dec 13 02:40:01.113124 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Dec 13 02:40:01.113133 kernel: NET: Registered PF_XDP protocol family Dec 13 02:40:01.113216 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Dec 13 02:40:01.113296 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Dec 13 02:40:01.113373 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Dec 13 02:40:01.113449 kernel: pci_bus 0000:00: resource 7 [mem 0x80000000-0xfebfffff window] Dec 13 02:40:01.113524 kernel: pci_bus 0000:00: resource 8 [mem 0x100000000-0x17fffffff window] Dec 13 02:40:01.113614 kernel: pci 0000:00:01.0: PIIX3: Enabling Passive Release Dec 13 02:40:01.113703 kernel: pci 0000:00:00.0: Limiting direct PCI/PCI transfers Dec 13 02:40:01.113716 kernel: PCI: CLS 0 bytes, default 64 Dec 13 02:40:01.113729 kernel: Initialise system trusted keyrings Dec 13 02:40:01.113738 kernel: workingset: timestamp_bits=39 max_order=19 bucket_order=0 Dec 13 02:40:01.113747 kernel: Key type asymmetric registered Dec 13 02:40:01.113756 kernel: Asymmetric key parser 'x509' registered Dec 13 02:40:01.113764 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Dec 13 02:40:01.113773 kernel: io scheduler mq-deadline registered Dec 13 02:40:01.113781 kernel: io scheduler kyber registered Dec 13 02:40:01.113790 kernel: io scheduler bfq registered Dec 13 02:40:01.113799 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Dec 13 02:40:01.113810 kernel: ACPI: \_SB_.LNKB: Enabled at IRQ 10 Dec 13 02:40:01.113819 kernel: ACPI: \_SB_.LNKC: Enabled at IRQ 11 Dec 13 02:40:01.113827 kernel: ACPI: \_SB_.LNKD: Enabled at IRQ 11 Dec 13 02:40:01.113836 kernel: ACPI: \_SB_.LNKA: Enabled at IRQ 10 Dec 13 02:40:01.113845 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Dec 13 02:40:01.113853 kernel: 00:00: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Dec 13 02:40:01.113863 kernel: random: crng init done Dec 13 02:40:01.113871 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Dec 13 02:40:01.113880 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Dec 13 02:40:01.113890 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Dec 13 02:40:01.114940 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Dec 13 02:40:01.115061 kernel: rtc_cmos 00:04: RTC can wake from S4 Dec 13 02:40:01.115157 kernel: rtc_cmos 00:04: registered as rtc0 Dec 13 02:40:01.115259 kernel: rtc_cmos 00:04: setting system clock to 2024-12-13T02:40:00 UTC (1734057600) Dec 13 02:40:01.115343 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram Dec 13 02:40:01.115355 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Dec 13 02:40:01.115369 kernel: NET: Registered PF_INET6 protocol family Dec 13 02:40:01.115377 kernel: Segment Routing with IPv6 Dec 13 02:40:01.115386 kernel: In-situ OAM (IOAM) with IPv6 Dec 13 02:40:01.115395 kernel: NET: Registered PF_PACKET protocol family Dec 13 02:40:01.115404 kernel: Key type dns_resolver registered Dec 13 02:40:01.115412 kernel: IPI shorthand broadcast: enabled Dec 13 02:40:01.115421 kernel: sched_clock: Marking stable (947008141, 123305096)->(1074836395, -4523158) Dec 13 02:40:01.115430 kernel: registered taskstats version 1 Dec 13 02:40:01.115439 kernel: Loading compiled-in X.509 certificates Dec 13 02:40:01.115447 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.65-flatcar: c82d546f528d79a5758dcebbc47fb6daf92836a0' Dec 13 02:40:01.115458 kernel: Key type .fscrypt registered Dec 13 02:40:01.115466 kernel: Key type fscrypt-provisioning registered Dec 13 02:40:01.115475 kernel: ima: No TPM chip found, activating TPM-bypass! Dec 13 02:40:01.115484 kernel: ima: Allocated hash algorithm: sha1 Dec 13 02:40:01.115492 kernel: ima: No architecture policies found Dec 13 02:40:01.115501 kernel: clk: Disabling unused clocks Dec 13 02:40:01.115509 kernel: Freeing unused kernel image (initmem) memory: 42844K Dec 13 02:40:01.115518 kernel: Write protecting the kernel read-only data: 36864k Dec 13 02:40:01.115529 kernel: Freeing unused kernel image (rodata/data gap) memory: 1852K Dec 13 02:40:01.115537 kernel: Run /init as init process Dec 13 02:40:01.115546 kernel: with arguments: Dec 13 02:40:01.115554 kernel: /init Dec 13 02:40:01.115563 kernel: with environment: Dec 13 02:40:01.115571 kernel: HOME=/ Dec 13 02:40:01.115580 kernel: TERM=linux Dec 13 02:40:01.115588 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Dec 13 02:40:01.115606 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Dec 13 02:40:01.115620 systemd[1]: Detected virtualization kvm. Dec 13 02:40:01.115630 systemd[1]: Detected architecture x86-64. Dec 13 02:40:01.115639 systemd[1]: Running in initrd. Dec 13 02:40:01.115648 systemd[1]: No hostname configured, using default hostname. Dec 13 02:40:01.115658 systemd[1]: Hostname set to . Dec 13 02:40:01.115667 systemd[1]: Initializing machine ID from VM UUID. Dec 13 02:40:01.115677 systemd[1]: Queued start job for default target initrd.target. Dec 13 02:40:01.115689 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Dec 13 02:40:01.115698 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Dec 13 02:40:01.115708 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Dec 13 02:40:01.115718 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Dec 13 02:40:01.115727 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Dec 13 02:40:01.115737 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Dec 13 02:40:01.115748 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Dec 13 02:40:01.115760 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Dec 13 02:40:01.115770 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Dec 13 02:40:01.115779 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Dec 13 02:40:01.115789 systemd[1]: Reached target paths.target - Path Units. Dec 13 02:40:01.115809 systemd[1]: Reached target slices.target - Slice Units. Dec 13 02:40:01.115821 systemd[1]: Reached target swap.target - Swaps. Dec 13 02:40:01.115832 systemd[1]: Reached target timers.target - Timer Units. Dec 13 02:40:01.115842 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Dec 13 02:40:01.115851 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Dec 13 02:40:01.115861 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Dec 13 02:40:01.115871 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Dec 13 02:40:01.115881 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Dec 13 02:40:01.115890 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Dec 13 02:40:01.116934 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Dec 13 02:40:01.116950 systemd[1]: Reached target sockets.target - Socket Units. Dec 13 02:40:01.116959 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Dec 13 02:40:01.116969 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Dec 13 02:40:01.116979 systemd[1]: Finished network-cleanup.service - Network Cleanup. Dec 13 02:40:01.116988 systemd[1]: Starting systemd-fsck-usr.service... Dec 13 02:40:01.116998 systemd[1]: Starting systemd-journald.service - Journal Service... Dec 13 02:40:01.117007 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Dec 13 02:40:01.117017 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Dec 13 02:40:01.117027 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Dec 13 02:40:01.117040 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Dec 13 02:40:01.117072 systemd-journald[185]: Collecting audit messages is disabled. Dec 13 02:40:01.117095 systemd[1]: Finished systemd-fsck-usr.service. Dec 13 02:40:01.117109 systemd-journald[185]: Journal started Dec 13 02:40:01.117131 systemd-journald[185]: Runtime Journal (/run/log/journal/2abc6322425547259962dc510fbd41be) is 4.9M, max 39.3M, 34.4M free. Dec 13 02:40:01.121763 systemd-modules-load[186]: Inserted module 'overlay' Dec 13 02:40:01.157540 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Dec 13 02:40:01.157612 systemd[1]: Started systemd-journald.service - Journal Service. Dec 13 02:40:01.157638 kernel: Bridge firewalling registered Dec 13 02:40:01.158651 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Dec 13 02:40:01.160994 systemd-modules-load[186]: Inserted module 'br_netfilter' Dec 13 02:40:01.162444 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Dec 13 02:40:01.168056 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Dec 13 02:40:01.169922 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Dec 13 02:40:01.183107 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Dec 13 02:40:01.187238 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Dec 13 02:40:01.194872 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Dec 13 02:40:01.197393 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Dec 13 02:40:01.206152 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Dec 13 02:40:01.216099 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Dec 13 02:40:01.216926 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Dec 13 02:40:01.221453 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Dec 13 02:40:01.224053 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Dec 13 02:40:01.227034 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Dec 13 02:40:01.241770 dracut-cmdline[218]: dracut-dracut-053 Dec 13 02:40:01.246454 dracut-cmdline[218]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=openstack flatcar.autologin verity.usrhash=2fdbba50b59d8c8a9877a81151806ddc16f473fe99b9ba0d8825997d654583ff Dec 13 02:40:01.303279 systemd-resolved[219]: Positive Trust Anchors: Dec 13 02:40:01.303312 systemd-resolved[219]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Dec 13 02:40:01.303354 systemd-resolved[219]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Dec 13 02:40:01.306230 systemd-resolved[219]: Defaulting to hostname 'linux'. Dec 13 02:40:01.307926 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Dec 13 02:40:01.311955 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Dec 13 02:40:01.388930 kernel: SCSI subsystem initialized Dec 13 02:40:01.399058 kernel: Loading iSCSI transport class v2.0-870. Dec 13 02:40:01.412003 kernel: iscsi: registered transport (tcp) Dec 13 02:40:01.438196 kernel: iscsi: registered transport (qla4xxx) Dec 13 02:40:01.438265 kernel: QLogic iSCSI HBA Driver Dec 13 02:40:01.502650 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Dec 13 02:40:01.512220 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Dec 13 02:40:01.568923 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Dec 13 02:40:01.569021 kernel: device-mapper: uevent: version 1.0.3 Dec 13 02:40:01.571923 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Dec 13 02:40:01.618955 kernel: raid6: sse2x4 gen() 12051 MB/s Dec 13 02:40:01.635950 kernel: raid6: sse2x2 gen() 14163 MB/s Dec 13 02:40:01.653135 kernel: raid6: sse2x1 gen() 9622 MB/s Dec 13 02:40:01.653216 kernel: raid6: using algorithm sse2x2 gen() 14163 MB/s Dec 13 02:40:01.671219 kernel: raid6: .... xor() 8517 MB/s, rmw enabled Dec 13 02:40:01.671336 kernel: raid6: using ssse3x2 recovery algorithm Dec 13 02:40:01.696176 kernel: xor: measuring software checksum speed Dec 13 02:40:01.696314 kernel: prefetch64-sse : 16332 MB/sec Dec 13 02:40:01.696344 kernel: generic_sse : 15143 MB/sec Dec 13 02:40:01.697254 kernel: xor: using function: prefetch64-sse (16332 MB/sec) Dec 13 02:40:01.887973 kernel: Btrfs loaded, zoned=no, fsverity=no Dec 13 02:40:01.906537 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Dec 13 02:40:01.918240 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Dec 13 02:40:01.931183 systemd-udevd[403]: Using default interface naming scheme 'v255'. Dec 13 02:40:01.935948 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Dec 13 02:40:01.948433 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Dec 13 02:40:01.969996 dracut-pre-trigger[413]: rd.md=0: removing MD RAID activation Dec 13 02:40:02.022640 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Dec 13 02:40:02.033052 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Dec 13 02:40:02.078491 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Dec 13 02:40:02.092400 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Dec 13 02:40:02.136835 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Dec 13 02:40:02.139358 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Dec 13 02:40:02.141815 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Dec 13 02:40:02.143293 systemd[1]: Reached target remote-fs.target - Remote File Systems. Dec 13 02:40:02.150381 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Dec 13 02:40:02.179809 kernel: virtio_blk virtio2: 2/0/0 default/read/poll queues Dec 13 02:40:02.217704 kernel: virtio_blk virtio2: [vda] 41943040 512-byte logical blocks (21.5 GB/20.0 GiB) Dec 13 02:40:02.217873 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Dec 13 02:40:02.217921 kernel: GPT:17805311 != 41943039 Dec 13 02:40:02.217936 kernel: GPT:Alternate GPT header not at the end of the disk. Dec 13 02:40:02.217949 kernel: GPT:17805311 != 41943039 Dec 13 02:40:02.217961 kernel: GPT: Use GNU Parted to correct GPT errors. Dec 13 02:40:02.217972 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Dec 13 02:40:02.181815 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Dec 13 02:40:02.223127 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Dec 13 02:40:02.223486 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Dec 13 02:40:02.227872 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Dec 13 02:40:02.228783 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Dec 13 02:40:02.229073 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Dec 13 02:40:02.233850 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Dec 13 02:40:02.246273 kernel: libata version 3.00 loaded. Dec 13 02:40:02.246266 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Dec 13 02:40:02.253227 kernel: ata_piix 0000:00:01.1: version 2.13 Dec 13 02:40:02.258628 kernel: scsi host0: ata_piix Dec 13 02:40:02.258758 kernel: scsi host1: ata_piix Dec 13 02:40:02.258868 kernel: ata1: PATA max MWDMA2 cmd 0x1f0 ctl 0x3f6 bmdma 0xc120 irq 14 Dec 13 02:40:02.258882 kernel: ata2: PATA max MWDMA2 cmd 0x170 ctl 0x376 bmdma 0xc128 irq 15 Dec 13 02:40:02.336151 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Dec 13 02:40:02.345208 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Dec 13 02:40:02.370612 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Dec 13 02:40:02.375346 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/vda6 scanned by (udev-worker) (463) Dec 13 02:40:02.399507 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Dec 13 02:40:02.471952 kernel: BTRFS: device fsid c3b72f8a-27ca-4d37-9d0e-1ec3c4bdc3be devid 1 transid 41 /dev/vda3 scanned by (udev-worker) (452) Dec 13 02:40:02.499287 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Dec 13 02:40:02.500979 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Dec 13 02:40:02.512223 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Dec 13 02:40:02.537560 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Dec 13 02:40:02.555600 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Dec 13 02:40:02.780035 disk-uuid[508]: Primary Header is updated. Dec 13 02:40:02.780035 disk-uuid[508]: Secondary Entries is updated. Dec 13 02:40:02.780035 disk-uuid[508]: Secondary Header is updated. Dec 13 02:40:02.794963 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Dec 13 02:40:02.807031 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Dec 13 02:40:03.829340 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Dec 13 02:40:03.831888 disk-uuid[509]: The operation has completed successfully. Dec 13 02:40:03.915340 systemd[1]: disk-uuid.service: Deactivated successfully. Dec 13 02:40:03.915652 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Dec 13 02:40:03.941099 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Dec 13 02:40:03.947566 sh[523]: Success Dec 13 02:40:03.964960 kernel: device-mapper: verity: sha256 using implementation "sha256-ssse3" Dec 13 02:40:04.033716 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Dec 13 02:40:04.044199 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Dec 13 02:40:04.046959 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Dec 13 02:40:04.098967 kernel: BTRFS info (device dm-0): first mount of filesystem c3b72f8a-27ca-4d37-9d0e-1ec3c4bdc3be Dec 13 02:40:04.099103 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Dec 13 02:40:04.102143 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Dec 13 02:40:04.105617 kernel: BTRFS info (device dm-0): disabling log replay at mount time Dec 13 02:40:04.108323 kernel: BTRFS info (device dm-0): using free space tree Dec 13 02:40:04.133726 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Dec 13 02:40:04.136237 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Dec 13 02:40:04.152417 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Dec 13 02:40:04.159297 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Dec 13 02:40:04.187964 kernel: BTRFS info (device vda6): first mount of filesystem db063747-cac8-4176-8963-c216c1b11dcb Dec 13 02:40:04.192720 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Dec 13 02:40:04.192786 kernel: BTRFS info (device vda6): using free space tree Dec 13 02:40:04.383258 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Dec 13 02:40:04.397301 systemd[1]: Starting systemd-networkd.service - Network Configuration... Dec 13 02:40:04.444646 systemd-networkd[694]: lo: Link UP Dec 13 02:40:04.444658 systemd-networkd[694]: lo: Gained carrier Dec 13 02:40:04.446183 systemd-networkd[694]: Enumeration completed Dec 13 02:40:04.446809 systemd[1]: Started systemd-networkd.service - Network Configuration. Dec 13 02:40:04.447520 systemd[1]: Reached target network.target - Network. Dec 13 02:40:04.448093 systemd-networkd[694]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 13 02:40:04.448096 systemd-networkd[694]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Dec 13 02:40:04.449555 systemd-networkd[694]: eth0: Link UP Dec 13 02:40:04.449559 systemd-networkd[694]: eth0: Gained carrier Dec 13 02:40:04.449567 systemd-networkd[694]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 13 02:40:04.458984 systemd-networkd[694]: eth0: DHCPv4 address 172.24.4.241/24, gateway 172.24.4.1 acquired from 172.24.4.1 Dec 13 02:40:04.473998 kernel: BTRFS info (device vda6): auto enabling async discard Dec 13 02:40:04.613509 systemd[1]: mnt-oem.mount: Deactivated successfully. Dec 13 02:40:04.618980 kernel: BTRFS info (device vda6): last unmount of filesystem db063747-cac8-4176-8963-c216c1b11dcb Dec 13 02:40:05.021705 systemd[1]: Finished ignition-setup.service - Ignition (setup). Dec 13 02:40:05.031373 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Dec 13 02:40:05.256944 ignition[707]: Ignition 2.19.0 Dec 13 02:40:05.256959 ignition[707]: Stage: fetch-offline Dec 13 02:40:05.257004 ignition[707]: no configs at "/usr/lib/ignition/base.d" Dec 13 02:40:05.258774 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Dec 13 02:40:05.257014 ignition[707]: no config dir at "/usr/lib/ignition/base.platform.d/openstack" Dec 13 02:40:05.257126 ignition[707]: parsed url from cmdline: "" Dec 13 02:40:05.257141 ignition[707]: no config URL provided Dec 13 02:40:05.257148 ignition[707]: reading system config file "/usr/lib/ignition/user.ign" Dec 13 02:40:05.257157 ignition[707]: no config at "/usr/lib/ignition/user.ign" Dec 13 02:40:05.257162 ignition[707]: failed to fetch config: resource requires networking Dec 13 02:40:05.257388 ignition[707]: Ignition finished successfully Dec 13 02:40:05.267154 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Dec 13 02:40:05.287333 ignition[715]: Ignition 2.19.0 Dec 13 02:40:05.287355 ignition[715]: Stage: fetch Dec 13 02:40:05.287760 ignition[715]: no configs at "/usr/lib/ignition/base.d" Dec 13 02:40:05.287786 ignition[715]: no config dir at "/usr/lib/ignition/base.platform.d/openstack" Dec 13 02:40:05.288062 ignition[715]: parsed url from cmdline: "" Dec 13 02:40:05.288072 ignition[715]: no config URL provided Dec 13 02:40:05.288085 ignition[715]: reading system config file "/usr/lib/ignition/user.ign" Dec 13 02:40:05.288104 ignition[715]: no config at "/usr/lib/ignition/user.ign" Dec 13 02:40:05.288384 ignition[715]: config drive ("/dev/disk/by-label/config-2") not found. Waiting... Dec 13 02:40:05.288415 ignition[715]: config drive ("/dev/disk/by-label/CONFIG-2") not found. Waiting... Dec 13 02:40:05.288466 ignition[715]: GET http://169.254.169.254/openstack/latest/user_data: attempt #1 Dec 13 02:40:05.501493 ignition[715]: GET result: OK Dec 13 02:40:05.501649 ignition[715]: parsing config with SHA512: 7285d375166e7faf03a39b83cb28bb18cde411dd66a620301eba53e53f480570762a20cf0e04858d3699f515ba9313574e214bfc8a82c4d05a6b948af5d31a0f Dec 13 02:40:05.509780 unknown[715]: fetched base config from "system" Dec 13 02:40:05.509808 unknown[715]: fetched base config from "system" Dec 13 02:40:05.510452 ignition[715]: fetch: fetch complete Dec 13 02:40:05.509823 unknown[715]: fetched user config from "openstack" Dec 13 02:40:05.510464 ignition[715]: fetch: fetch passed Dec 13 02:40:05.514400 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Dec 13 02:40:05.510567 ignition[715]: Ignition finished successfully Dec 13 02:40:05.528826 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Dec 13 02:40:05.559065 ignition[722]: Ignition 2.19.0 Dec 13 02:40:05.559095 ignition[722]: Stage: kargs Dec 13 02:40:05.559720 ignition[722]: no configs at "/usr/lib/ignition/base.d" Dec 13 02:40:05.559748 ignition[722]: no config dir at "/usr/lib/ignition/base.platform.d/openstack" Dec 13 02:40:05.564690 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Dec 13 02:40:05.561611 ignition[722]: kargs: kargs passed Dec 13 02:40:05.561715 ignition[722]: Ignition finished successfully Dec 13 02:40:05.575405 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Dec 13 02:40:05.614608 ignition[728]: Ignition 2.19.0 Dec 13 02:40:05.614638 ignition[728]: Stage: disks Dec 13 02:40:05.615259 ignition[728]: no configs at "/usr/lib/ignition/base.d" Dec 13 02:40:05.615290 ignition[728]: no config dir at "/usr/lib/ignition/base.platform.d/openstack" Dec 13 02:40:05.619828 systemd[1]: Finished ignition-disks.service - Ignition (disks). Dec 13 02:40:05.617347 ignition[728]: disks: disks passed Dec 13 02:40:05.623249 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Dec 13 02:40:05.617459 ignition[728]: Ignition finished successfully Dec 13 02:40:05.625361 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Dec 13 02:40:05.627069 systemd[1]: Reached target local-fs.target - Local File Systems. Dec 13 02:40:05.629176 systemd[1]: Reached target sysinit.target - System Initialization. Dec 13 02:40:05.630923 systemd[1]: Reached target basic.target - Basic System. Dec 13 02:40:05.646350 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Dec 13 02:40:05.675477 systemd-fsck[736]: ROOT: clean, 14/1628000 files, 120691/1617920 blocks Dec 13 02:40:05.692371 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Dec 13 02:40:05.701270 systemd[1]: Mounting sysroot.mount - /sysroot... Dec 13 02:40:05.849076 kernel: EXT4-fs (vda9): mounted filesystem 390119fa-ab9c-4f50-b046-3b5c76c46193 r/w with ordered data mode. Quota mode: none. Dec 13 02:40:05.850619 systemd[1]: Mounted sysroot.mount - /sysroot. Dec 13 02:40:05.853021 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Dec 13 02:40:05.869180 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Dec 13 02:40:05.873149 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Dec 13 02:40:05.876111 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Dec 13 02:40:05.879799 systemd[1]: Starting flatcar-openstack-hostname.service - Flatcar OpenStack Metadata Hostname Agent... Dec 13 02:40:05.882867 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Dec 13 02:40:05.891572 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/vda6 scanned by mount (745) Dec 13 02:40:05.891598 kernel: BTRFS info (device vda6): first mount of filesystem db063747-cac8-4176-8963-c216c1b11dcb Dec 13 02:40:05.891612 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Dec 13 02:40:05.891623 kernel: BTRFS info (device vda6): using free space tree Dec 13 02:40:05.882982 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Dec 13 02:40:05.895395 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Dec 13 02:40:05.900929 kernel: BTRFS info (device vda6): auto enabling async discard Dec 13 02:40:05.904072 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Dec 13 02:40:05.905749 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Dec 13 02:40:06.116163 initrd-setup-root[773]: cut: /sysroot/etc/passwd: No such file or directory Dec 13 02:40:06.131276 initrd-setup-root[780]: cut: /sysroot/etc/group: No such file or directory Dec 13 02:40:06.180266 initrd-setup-root[787]: cut: /sysroot/etc/shadow: No such file or directory Dec 13 02:40:06.248656 initrd-setup-root[794]: cut: /sysroot/etc/gshadow: No such file or directory Dec 13 02:40:06.364529 systemd-networkd[694]: eth0: Gained IPv6LL Dec 13 02:40:06.697168 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Dec 13 02:40:06.714130 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Dec 13 02:40:06.719300 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Dec 13 02:40:06.736589 systemd[1]: sysroot-oem.mount: Deactivated successfully. Dec 13 02:40:06.742950 kernel: BTRFS info (device vda6): last unmount of filesystem db063747-cac8-4176-8963-c216c1b11dcb Dec 13 02:40:06.780580 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Dec 13 02:40:06.801187 ignition[862]: INFO : Ignition 2.19.0 Dec 13 02:40:06.801187 ignition[862]: INFO : Stage: mount Dec 13 02:40:06.802334 ignition[862]: INFO : no configs at "/usr/lib/ignition/base.d" Dec 13 02:40:06.802334 ignition[862]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/openstack" Dec 13 02:40:06.803686 ignition[862]: INFO : mount: mount passed Dec 13 02:40:06.804266 ignition[862]: INFO : Ignition finished successfully Dec 13 02:40:06.806103 systemd[1]: Finished ignition-mount.service - Ignition (mount). Dec 13 02:40:13.623453 coreos-metadata[747]: Dec 13 02:40:13.623 WARN failed to locate config-drive, using the metadata service API instead Dec 13 02:40:13.663939 coreos-metadata[747]: Dec 13 02:40:13.663 INFO Fetching http://169.254.169.254/latest/meta-data/hostname: Attempt #1 Dec 13 02:40:13.678300 coreos-metadata[747]: Dec 13 02:40:13.678 INFO Fetch successful Dec 13 02:40:13.679771 coreos-metadata[747]: Dec 13 02:40:13.679 INFO wrote hostname ci-4081-2-1-b-3218de15a5.novalocal to /sysroot/etc/hostname Dec 13 02:40:13.682600 systemd[1]: flatcar-openstack-hostname.service: Deactivated successfully. Dec 13 02:40:13.682858 systemd[1]: Finished flatcar-openstack-hostname.service - Flatcar OpenStack Metadata Hostname Agent. Dec 13 02:40:13.698228 systemd[1]: Starting ignition-files.service - Ignition (files)... Dec 13 02:40:13.728332 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Dec 13 02:40:13.743955 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/vda6 scanned by mount (879) Dec 13 02:40:13.749997 kernel: BTRFS info (device vda6): first mount of filesystem db063747-cac8-4176-8963-c216c1b11dcb Dec 13 02:40:13.750072 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Dec 13 02:40:13.754414 kernel: BTRFS info (device vda6): using free space tree Dec 13 02:40:13.761949 kernel: BTRFS info (device vda6): auto enabling async discard Dec 13 02:40:13.768204 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Dec 13 02:40:13.813474 ignition[897]: INFO : Ignition 2.19.0 Dec 13 02:40:13.815272 ignition[897]: INFO : Stage: files Dec 13 02:40:13.815272 ignition[897]: INFO : no configs at "/usr/lib/ignition/base.d" Dec 13 02:40:13.815272 ignition[897]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/openstack" Dec 13 02:40:13.819585 ignition[897]: DEBUG : files: compiled without relabeling support, skipping Dec 13 02:40:13.819585 ignition[897]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Dec 13 02:40:13.819585 ignition[897]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Dec 13 02:40:13.825304 ignition[897]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Dec 13 02:40:13.826021 ignition[897]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Dec 13 02:40:13.826795 ignition[897]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Dec 13 02:40:13.826355 unknown[897]: wrote ssh authorized keys file for user: core Dec 13 02:40:13.830123 ignition[897]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/home/core/install.sh" Dec 13 02:40:13.831018 ignition[897]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/home/core/install.sh" Dec 13 02:40:13.831018 ignition[897]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/etc/flatcar/update.conf" Dec 13 02:40:13.832712 ignition[897]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/etc/flatcar/update.conf" Dec 13 02:40:13.832712 ignition[897]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.29.2-x86-64.raw" Dec 13 02:40:13.832712 ignition[897]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.29.2-x86-64.raw" Dec 13 02:40:13.832712 ignition[897]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.29.2-x86-64.raw" Dec 13 02:40:13.832712 ignition[897]: INFO : files: createFilesystemsFiles: createFiles: op(6): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.29.2-x86-64.raw: attempt #1 Dec 13 02:40:14.262376 ignition[897]: INFO : files: createFilesystemsFiles: createFiles: op(6): GET result: OK Dec 13 02:40:15.902628 ignition[897]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.29.2-x86-64.raw" Dec 13 02:40:15.904663 ignition[897]: INFO : files: createResultFile: createFiles: op(7): [started] writing file "/sysroot/etc/.ignition-result.json" Dec 13 02:40:15.904663 ignition[897]: INFO : files: createResultFile: createFiles: op(7): [finished] writing file "/sysroot/etc/.ignition-result.json" Dec 13 02:40:15.904663 ignition[897]: INFO : files: files passed Dec 13 02:40:15.904663 ignition[897]: INFO : Ignition finished successfully Dec 13 02:40:15.908434 systemd[1]: Finished ignition-files.service - Ignition (files). Dec 13 02:40:15.916198 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Dec 13 02:40:15.921008 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Dec 13 02:40:15.952573 systemd[1]: ignition-quench.service: Deactivated successfully. Dec 13 02:40:15.952739 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Dec 13 02:40:15.970942 initrd-setup-root-after-ignition[926]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Dec 13 02:40:15.970942 initrd-setup-root-after-ignition[926]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Dec 13 02:40:15.975734 initrd-setup-root-after-ignition[930]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Dec 13 02:40:15.976259 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Dec 13 02:40:15.983843 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Dec 13 02:40:15.994651 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Dec 13 02:40:16.063756 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Dec 13 02:40:16.064162 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Dec 13 02:40:16.067426 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Dec 13 02:40:16.070000 systemd[1]: Reached target initrd.target - Initrd Default Target. Dec 13 02:40:16.072937 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Dec 13 02:40:16.080219 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Dec 13 02:40:16.116783 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Dec 13 02:40:16.130174 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Dec 13 02:40:16.158615 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Dec 13 02:40:16.160368 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Dec 13 02:40:16.163400 systemd[1]: Stopped target timers.target - Timer Units. Dec 13 02:40:16.166131 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Dec 13 02:40:16.166431 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Dec 13 02:40:16.169732 systemd[1]: Stopped target initrd.target - Initrd Default Target. Dec 13 02:40:16.171495 systemd[1]: Stopped target basic.target - Basic System. Dec 13 02:40:16.174206 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Dec 13 02:40:16.176710 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Dec 13 02:40:16.179308 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Dec 13 02:40:16.182066 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Dec 13 02:40:16.184864 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Dec 13 02:40:16.187793 systemd[1]: Stopped target sysinit.target - System Initialization. Dec 13 02:40:16.199529 systemd[1]: Stopped target local-fs.target - Local File Systems. Dec 13 02:40:16.202395 systemd[1]: Stopped target swap.target - Swaps. Dec 13 02:40:16.204866 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Dec 13 02:40:16.205201 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Dec 13 02:40:16.208197 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Dec 13 02:40:16.209448 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Dec 13 02:40:16.211203 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Dec 13 02:40:16.213639 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Dec 13 02:40:16.214925 systemd[1]: dracut-initqueue.service: Deactivated successfully. Dec 13 02:40:16.215102 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Dec 13 02:40:16.217840 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Dec 13 02:40:16.218052 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Dec 13 02:40:16.218996 systemd[1]: ignition-files.service: Deactivated successfully. Dec 13 02:40:16.219111 systemd[1]: Stopped ignition-files.service - Ignition (files). Dec 13 02:40:16.230332 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Dec 13 02:40:16.232551 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Dec 13 02:40:16.233064 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Dec 13 02:40:16.233195 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Dec 13 02:40:16.235075 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Dec 13 02:40:16.235212 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Dec 13 02:40:16.244511 systemd[1]: initrd-cleanup.service: Deactivated successfully. Dec 13 02:40:16.245208 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Dec 13 02:40:16.255754 ignition[950]: INFO : Ignition 2.19.0 Dec 13 02:40:16.257971 ignition[950]: INFO : Stage: umount Dec 13 02:40:16.257971 ignition[950]: INFO : no configs at "/usr/lib/ignition/base.d" Dec 13 02:40:16.257971 ignition[950]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/openstack" Dec 13 02:40:16.257971 ignition[950]: INFO : umount: umount passed Dec 13 02:40:16.257971 ignition[950]: INFO : Ignition finished successfully Dec 13 02:40:16.261135 systemd[1]: ignition-mount.service: Deactivated successfully. Dec 13 02:40:16.261817 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Dec 13 02:40:16.262741 systemd[1]: ignition-disks.service: Deactivated successfully. Dec 13 02:40:16.262810 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Dec 13 02:40:16.263378 systemd[1]: ignition-kargs.service: Deactivated successfully. Dec 13 02:40:16.263419 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Dec 13 02:40:16.264419 systemd[1]: ignition-fetch.service: Deactivated successfully. Dec 13 02:40:16.264459 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Dec 13 02:40:16.265445 systemd[1]: Stopped target network.target - Network. Dec 13 02:40:16.266420 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Dec 13 02:40:16.266467 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Dec 13 02:40:16.267521 systemd[1]: Stopped target paths.target - Path Units. Dec 13 02:40:16.268459 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Dec 13 02:40:16.270038 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Dec 13 02:40:16.270641 systemd[1]: Stopped target slices.target - Slice Units. Dec 13 02:40:16.271760 systemd[1]: Stopped target sockets.target - Socket Units. Dec 13 02:40:16.272976 systemd[1]: iscsid.socket: Deactivated successfully. Dec 13 02:40:16.273014 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Dec 13 02:40:16.273909 systemd[1]: iscsiuio.socket: Deactivated successfully. Dec 13 02:40:16.273945 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Dec 13 02:40:16.275004 systemd[1]: ignition-setup.service: Deactivated successfully. Dec 13 02:40:16.275046 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Dec 13 02:40:16.276234 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Dec 13 02:40:16.276274 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Dec 13 02:40:16.277366 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Dec 13 02:40:16.278410 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Dec 13 02:40:16.281990 systemd-networkd[694]: eth0: DHCPv6 lease lost Dec 13 02:40:16.285618 systemd[1]: systemd-resolved.service: Deactivated successfully. Dec 13 02:40:16.285745 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Dec 13 02:40:16.288801 systemd[1]: systemd-networkd.service: Deactivated successfully. Dec 13 02:40:16.288947 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Dec 13 02:40:16.290920 systemd[1]: systemd-networkd.socket: Deactivated successfully. Dec 13 02:40:16.291171 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Dec 13 02:40:16.298086 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Dec 13 02:40:16.300857 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Dec 13 02:40:16.301607 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Dec 13 02:40:16.302222 systemd[1]: systemd-sysctl.service: Deactivated successfully. Dec 13 02:40:16.302265 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Dec 13 02:40:16.302744 systemd[1]: systemd-modules-load.service: Deactivated successfully. Dec 13 02:40:16.302781 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Dec 13 02:40:16.304441 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Dec 13 02:40:16.304506 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Dec 13 02:40:16.305405 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Dec 13 02:40:16.313154 systemd[1]: sysroot-boot.mount: Deactivated successfully. Dec 13 02:40:16.317088 systemd[1]: systemd-udevd.service: Deactivated successfully. Dec 13 02:40:16.317237 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Dec 13 02:40:16.318106 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Dec 13 02:40:16.318146 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Dec 13 02:40:16.319467 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Dec 13 02:40:16.319500 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Dec 13 02:40:16.320727 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Dec 13 02:40:16.320778 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Dec 13 02:40:16.322445 systemd[1]: dracut-cmdline.service: Deactivated successfully. Dec 13 02:40:16.322489 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Dec 13 02:40:16.323700 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Dec 13 02:40:16.323749 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Dec 13 02:40:16.331320 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Dec 13 02:40:16.331926 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Dec 13 02:40:16.331986 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Dec 13 02:40:16.332540 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Dec 13 02:40:16.332586 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Dec 13 02:40:16.333214 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Dec 13 02:40:16.333265 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Dec 13 02:40:16.334462 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Dec 13 02:40:16.334505 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Dec 13 02:40:16.341070 systemd[1]: network-cleanup.service: Deactivated successfully. Dec 13 02:40:16.341172 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Dec 13 02:40:16.342316 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Dec 13 02:40:16.342398 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Dec 13 02:40:16.663938 systemd[1]: sysroot-boot.service: Deactivated successfully. Dec 13 02:40:16.664227 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Dec 13 02:40:16.668646 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Dec 13 02:40:16.670145 systemd[1]: initrd-setup-root.service: Deactivated successfully. Dec 13 02:40:16.670289 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Dec 13 02:40:16.686283 systemd[1]: Starting initrd-switch-root.service - Switch Root... Dec 13 02:40:16.708745 systemd[1]: Switching root. Dec 13 02:40:16.765507 systemd-journald[185]: Journal stopped Dec 13 02:40:23.275175 systemd-journald[185]: Received SIGTERM from PID 1 (systemd). Dec 13 02:40:23.275249 kernel: SELinux: policy capability network_peer_controls=1 Dec 13 02:40:23.275269 kernel: SELinux: policy capability open_perms=1 Dec 13 02:40:23.275281 kernel: SELinux: policy capability extended_socket_class=1 Dec 13 02:40:23.275298 kernel: SELinux: policy capability always_check_network=0 Dec 13 02:40:23.275315 kernel: SELinux: policy capability cgroup_seclabel=1 Dec 13 02:40:23.275328 kernel: SELinux: policy capability nnp_nosuid_transition=1 Dec 13 02:40:23.275340 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Dec 13 02:40:23.275352 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Dec 13 02:40:23.275368 kernel: audit: type=1403 audit(1734057619.309:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Dec 13 02:40:23.275394 systemd[1]: Successfully loaded SELinux policy in 144.906ms. Dec 13 02:40:23.275429 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 26.435ms. Dec 13 02:40:23.275451 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Dec 13 02:40:23.275469 systemd[1]: Detected virtualization kvm. Dec 13 02:40:23.275483 systemd[1]: Detected architecture x86-64. Dec 13 02:40:23.275496 systemd[1]: Detected first boot. Dec 13 02:40:23.275509 systemd[1]: Hostname set to . Dec 13 02:40:23.275524 systemd[1]: Initializing machine ID from VM UUID. Dec 13 02:40:23.275542 zram_generator::config[992]: No configuration found. Dec 13 02:40:23.275556 systemd[1]: Populated /etc with preset unit settings. Dec 13 02:40:23.275568 systemd[1]: initrd-switch-root.service: Deactivated successfully. Dec 13 02:40:23.275581 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Dec 13 02:40:23.275594 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Dec 13 02:40:23.275608 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Dec 13 02:40:23.275621 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Dec 13 02:40:23.275633 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Dec 13 02:40:23.275648 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Dec 13 02:40:23.275661 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Dec 13 02:40:23.275674 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Dec 13 02:40:23.275688 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Dec 13 02:40:23.275701 systemd[1]: Created slice user.slice - User and Session Slice. Dec 13 02:40:23.275714 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Dec 13 02:40:23.275727 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Dec 13 02:40:23.275740 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Dec 13 02:40:23.275752 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Dec 13 02:40:23.275769 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Dec 13 02:40:23.275782 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Dec 13 02:40:23.275795 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Dec 13 02:40:23.275808 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Dec 13 02:40:23.275822 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Dec 13 02:40:23.276100 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Dec 13 02:40:23.276138 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Dec 13 02:40:23.276151 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Dec 13 02:40:23.276165 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Dec 13 02:40:23.276178 systemd[1]: Reached target remote-fs.target - Remote File Systems. Dec 13 02:40:23.276191 systemd[1]: Reached target slices.target - Slice Units. Dec 13 02:40:23.276203 systemd[1]: Reached target swap.target - Swaps. Dec 13 02:40:23.276217 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Dec 13 02:40:23.276230 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Dec 13 02:40:23.276243 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Dec 13 02:40:23.276259 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Dec 13 02:40:23.276272 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Dec 13 02:40:23.276285 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Dec 13 02:40:23.276298 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Dec 13 02:40:23.276311 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Dec 13 02:40:23.276325 systemd[1]: Mounting media.mount - External Media Directory... Dec 13 02:40:23.276342 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 13 02:40:23.276356 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Dec 13 02:40:23.276369 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Dec 13 02:40:23.276385 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Dec 13 02:40:23.276399 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Dec 13 02:40:23.276412 systemd[1]: Reached target machines.target - Containers. Dec 13 02:40:23.276425 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Dec 13 02:40:23.276438 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Dec 13 02:40:23.276452 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Dec 13 02:40:23.276465 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Dec 13 02:40:23.276477 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Dec 13 02:40:23.276493 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Dec 13 02:40:23.276506 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Dec 13 02:40:23.276519 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Dec 13 02:40:23.276532 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Dec 13 02:40:23.276545 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Dec 13 02:40:23.276560 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Dec 13 02:40:23.276573 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Dec 13 02:40:23.276586 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Dec 13 02:40:23.276599 systemd[1]: Stopped systemd-fsck-usr.service. Dec 13 02:40:23.276614 kernel: fuse: init (API version 7.39) Dec 13 02:40:23.276627 systemd[1]: Starting systemd-journald.service - Journal Service... Dec 13 02:40:23.276640 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Dec 13 02:40:23.276652 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Dec 13 02:40:23.276665 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Dec 13 02:40:23.276678 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Dec 13 02:40:23.276692 systemd[1]: verity-setup.service: Deactivated successfully. Dec 13 02:40:23.276705 systemd[1]: Stopped verity-setup.service. Dec 13 02:40:23.276718 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 13 02:40:23.276735 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Dec 13 02:40:23.276747 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Dec 13 02:40:23.276761 systemd[1]: Mounted media.mount - External Media Directory. Dec 13 02:40:23.276773 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Dec 13 02:40:23.276786 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Dec 13 02:40:23.276802 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Dec 13 02:40:23.276815 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Dec 13 02:40:23.276828 systemd[1]: modprobe@configfs.service: Deactivated successfully. Dec 13 02:40:23.276842 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Dec 13 02:40:23.276855 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Dec 13 02:40:23.276869 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Dec 13 02:40:23.276884 kernel: ACPI: bus type drm_connector registered Dec 13 02:40:23.277178 systemd[1]: modprobe@drm.service: Deactivated successfully. Dec 13 02:40:23.277198 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Dec 13 02:40:23.277211 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Dec 13 02:40:23.277225 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Dec 13 02:40:23.277238 systemd[1]: modprobe@fuse.service: Deactivated successfully. Dec 13 02:40:23.277251 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Dec 13 02:40:23.277264 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Dec 13 02:40:23.277282 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Dec 13 02:40:23.277295 kernel: loop: module loaded Dec 13 02:40:23.277308 systemd[1]: modprobe@loop.service: Deactivated successfully. Dec 13 02:40:23.277321 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Dec 13 02:40:23.277334 systemd[1]: Reached target network-pre.target - Preparation for Network. Dec 13 02:40:23.277347 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Dec 13 02:40:23.277360 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Dec 13 02:40:23.277374 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Dec 13 02:40:23.277389 systemd[1]: Reached target local-fs.target - Local File Systems. Dec 13 02:40:23.277402 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Dec 13 02:40:23.277416 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Dec 13 02:40:23.277429 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Dec 13 02:40:23.277443 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Dec 13 02:40:23.277503 systemd-journald[1078]: Collecting audit messages is disabled. Dec 13 02:40:23.277533 systemd-journald[1078]: Journal started Dec 13 02:40:23.277562 systemd-journald[1078]: Runtime Journal (/run/log/journal/2abc6322425547259962dc510fbd41be) is 4.9M, max 39.3M, 34.4M free. Dec 13 02:40:22.760617 systemd[1]: Queued start job for default target multi-user.target. Dec 13 02:40:22.805083 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Dec 13 02:40:22.805500 systemd[1]: systemd-journald.service: Deactivated successfully. Dec 13 02:40:23.285068 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Dec 13 02:40:23.285147 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Dec 13 02:40:23.297942 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Dec 13 02:40:23.298033 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Dec 13 02:40:23.303936 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Dec 13 02:40:23.308018 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Dec 13 02:40:23.313013 systemd[1]: Started systemd-journald.service - Journal Service. Dec 13 02:40:23.315983 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Dec 13 02:40:23.316760 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Dec 13 02:40:23.317423 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Dec 13 02:40:23.318432 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Dec 13 02:40:23.342724 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Dec 13 02:40:23.353253 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Dec 13 02:40:23.360041 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Dec 13 02:40:23.367228 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Dec 13 02:40:23.368334 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Dec 13 02:40:23.369661 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Dec 13 02:40:23.379880 kernel: loop0: detected capacity change from 0 to 211296 Dec 13 02:40:23.379347 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Dec 13 02:40:23.390137 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Dec 13 02:40:23.413830 systemd-journald[1078]: Time spent on flushing to /var/log/journal/2abc6322425547259962dc510fbd41be is 26.743ms for 928 entries. Dec 13 02:40:23.413830 systemd-journald[1078]: System Journal (/var/log/journal/2abc6322425547259962dc510fbd41be) is 8.0M, max 584.8M, 576.8M free. Dec 13 02:40:23.463966 systemd-journald[1078]: Received client request to flush runtime journal. Dec 13 02:40:23.428084 udevadm[1130]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in. Dec 13 02:40:23.434492 systemd-tmpfiles[1104]: ACLs are not supported, ignoring. Dec 13 02:40:23.434510 systemd-tmpfiles[1104]: ACLs are not supported, ignoring. Dec 13 02:40:23.449683 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Dec 13 02:40:23.464428 systemd[1]: Starting systemd-sysusers.service - Create System Users... Dec 13 02:40:23.470620 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Dec 13 02:40:23.478936 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Dec 13 02:40:23.587990 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Dec 13 02:40:23.685566 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Dec 13 02:40:23.687272 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Dec 13 02:40:23.787963 kernel: loop1: detected capacity change from 0 to 142488 Dec 13 02:40:24.031388 systemd[1]: Finished systemd-sysusers.service - Create System Users. Dec 13 02:40:24.043374 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Dec 13 02:40:24.077742 systemd-tmpfiles[1148]: ACLs are not supported, ignoring. Dec 13 02:40:24.078327 systemd-tmpfiles[1148]: ACLs are not supported, ignoring. Dec 13 02:40:24.088966 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Dec 13 02:40:24.226963 kernel: loop2: detected capacity change from 0 to 140768 Dec 13 02:40:24.351547 kernel: loop3: detected capacity change from 0 to 8 Dec 13 02:40:24.377005 kernel: loop4: detected capacity change from 0 to 211296 Dec 13 02:40:24.514958 kernel: loop5: detected capacity change from 0 to 142488 Dec 13 02:40:24.650919 kernel: loop6: detected capacity change from 0 to 140768 Dec 13 02:40:24.874981 kernel: loop7: detected capacity change from 0 to 8 Dec 13 02:40:24.878667 (sd-merge)[1155]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-openstack'. Dec 13 02:40:24.879790 (sd-merge)[1155]: Merged extensions into '/usr'. Dec 13 02:40:24.889246 systemd[1]: Reloading requested from client PID 1103 ('systemd-sysext') (unit systemd-sysext.service)... Dec 13 02:40:24.889285 systemd[1]: Reloading... Dec 13 02:40:25.011266 zram_generator::config[1181]: No configuration found. Dec 13 02:40:25.183122 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Dec 13 02:40:25.242704 systemd[1]: Reloading finished in 352 ms. Dec 13 02:40:25.287295 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Dec 13 02:40:25.289423 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Dec 13 02:40:25.300100 systemd[1]: Starting ensure-sysext.service... Dec 13 02:40:25.304107 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Dec 13 02:40:25.309124 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Dec 13 02:40:25.324843 systemd[1]: Reloading requested from client PID 1237 ('systemctl') (unit ensure-sysext.service)... Dec 13 02:40:25.324865 systemd[1]: Reloading... Dec 13 02:40:25.325094 ldconfig[1096]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Dec 13 02:40:25.344417 systemd-udevd[1239]: Using default interface naming scheme 'v255'. Dec 13 02:40:25.359313 systemd-tmpfiles[1238]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Dec 13 02:40:25.359697 systemd-tmpfiles[1238]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Dec 13 02:40:25.361292 systemd-tmpfiles[1238]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Dec 13 02:40:25.364600 systemd-tmpfiles[1238]: ACLs are not supported, ignoring. Dec 13 02:40:25.364740 systemd-tmpfiles[1238]: ACLs are not supported, ignoring. Dec 13 02:40:25.371519 systemd-tmpfiles[1238]: Detected autofs mount point /boot during canonicalization of boot. Dec 13 02:40:25.371658 systemd-tmpfiles[1238]: Skipping /boot Dec 13 02:40:25.386468 systemd-tmpfiles[1238]: Detected autofs mount point /boot during canonicalization of boot. Dec 13 02:40:25.386892 systemd-tmpfiles[1238]: Skipping /boot Dec 13 02:40:25.423951 zram_generator::config[1272]: No configuration found. Dec 13 02:40:25.492937 kernel: BTRFS info: devid 1 device path /dev/mapper/usr changed to /dev/dm-0 scanned by (udev-worker) (1295) Dec 13 02:40:25.498075 kernel: BTRFS info: devid 1 device path /dev/dm-0 changed to /dev/mapper/usr scanned by (udev-worker) (1295) Dec 13 02:40:25.502937 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 41 scanned by (udev-worker) (1278) Dec 13 02:40:25.629961 kernel: piix4_smbus 0000:00:01.3: SMBus Host Controller at 0x700, revision 0 Dec 13 02:40:25.655173 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input2 Dec 13 02:40:25.663629 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Dec 13 02:40:25.666400 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input3 Dec 13 02:40:25.666436 kernel: ACPI: button: Power Button [PWRF] Dec 13 02:40:25.710922 kernel: mousedev: PS/2 mouse device common for all mice Dec 13 02:40:25.737611 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Dec 13 02:40:25.738087 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Dec 13 02:40:25.738909 systemd[1]: Reloading finished in 413 ms. Dec 13 02:40:25.755820 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Dec 13 02:40:25.756868 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Dec 13 02:40:25.763528 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Dec 13 02:40:25.794528 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 13 02:40:25.807293 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Dec 13 02:40:25.813008 kernel: [drm] pci: virtio-vga detected at 0000:00:02.0 Dec 13 02:40:25.813083 kernel: virtio-pci 0000:00:02.0: vgaarb: deactivate vga console Dec 13 02:40:25.817314 kernel: Console: switching to colour dummy device 80x25 Dec 13 02:40:25.818015 kernel: [drm] features: -virgl +edid -resource_blob -host_visible Dec 13 02:40:25.818081 kernel: [drm] features: -context_init Dec 13 02:40:25.820973 kernel: [drm] number of scanouts: 1 Dec 13 02:40:25.821083 kernel: [drm] number of cap sets: 0 Dec 13 02:40:25.823975 kernel: [drm] Initialized virtio_gpu 0.1.0 0 for 0000:00:02.0 on minor 0 Dec 13 02:40:25.835724 kernel: fbcon: virtio_gpudrmfb (fb0) is primary device Dec 13 02:40:25.835802 kernel: Console: switching to colour frame buffer device 128x48 Dec 13 02:40:25.839282 kernel: virtio-pci 0000:00:02.0: [drm] fb0: virtio_gpudrmfb frame buffer device Dec 13 02:40:25.866373 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Dec 13 02:40:25.866728 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Dec 13 02:40:25.870220 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Dec 13 02:40:25.882565 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Dec 13 02:40:25.887253 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Dec 13 02:40:25.887532 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Dec 13 02:40:25.891462 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Dec 13 02:40:25.897222 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Dec 13 02:40:25.902305 systemd[1]: Starting systemd-networkd.service - Network Configuration... Dec 13 02:40:25.911432 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Dec 13 02:40:25.921371 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Dec 13 02:40:25.934381 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Dec 13 02:40:25.935530 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 13 02:40:25.939921 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Dec 13 02:40:25.940117 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Dec 13 02:40:25.941559 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Dec 13 02:40:25.941836 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Dec 13 02:40:25.951392 systemd[1]: modprobe@loop.service: Deactivated successfully. Dec 13 02:40:25.951586 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Dec 13 02:40:25.952366 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Dec 13 02:40:25.961091 augenrules[1378]: No rules Dec 13 02:40:25.962961 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Dec 13 02:40:25.983180 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 13 02:40:25.983531 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Dec 13 02:40:25.993240 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Dec 13 02:40:25.996832 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Dec 13 02:40:26.000255 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Dec 13 02:40:26.001533 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Dec 13 02:40:26.007393 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Dec 13 02:40:26.008078 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 13 02:40:26.012393 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Dec 13 02:40:26.017887 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Dec 13 02:40:26.018537 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Dec 13 02:40:26.035985 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Dec 13 02:40:26.036176 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Dec 13 02:40:26.042487 systemd[1]: modprobe@loop.service: Deactivated successfully. Dec 13 02:40:26.042889 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Dec 13 02:40:26.049260 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Dec 13 02:40:26.064313 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 13 02:40:26.064628 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Dec 13 02:40:26.072206 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Dec 13 02:40:26.084982 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Dec 13 02:40:26.085853 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Dec 13 02:40:26.085965 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Dec 13 02:40:26.089365 systemd[1]: Starting systemd-update-done.service - Update is Completed... Dec 13 02:40:26.091220 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Dec 13 02:40:26.091264 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 13 02:40:26.095523 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Dec 13 02:40:26.100426 systemd[1]: Finished ensure-sysext.service. Dec 13 02:40:26.104792 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Dec 13 02:40:26.105735 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Dec 13 02:40:26.105959 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Dec 13 02:40:26.106763 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Dec 13 02:40:26.108625 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Dec 13 02:40:26.111659 systemd[1]: modprobe@drm.service: Deactivated successfully. Dec 13 02:40:26.111967 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Dec 13 02:40:26.116834 systemd[1]: Started systemd-userdbd.service - User Database Manager. Dec 13 02:40:26.143181 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Dec 13 02:40:26.144146 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Dec 13 02:40:26.150856 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Dec 13 02:40:26.159415 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Dec 13 02:40:26.181448 systemd[1]: Finished systemd-update-done.service - Update is Completed. Dec 13 02:40:26.185953 lvm[1416]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Dec 13 02:40:26.236315 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Dec 13 02:40:26.241989 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Dec 13 02:40:26.257161 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Dec 13 02:40:26.260802 systemd-networkd[1371]: lo: Link UP Dec 13 02:40:26.261939 systemd-networkd[1371]: lo: Gained carrier Dec 13 02:40:26.263496 systemd-networkd[1371]: Enumeration completed Dec 13 02:40:26.265220 systemd[1]: Started systemd-networkd.service - Network Configuration. Dec 13 02:40:26.266496 systemd-networkd[1371]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 13 02:40:26.266575 systemd-networkd[1371]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Dec 13 02:40:26.268453 systemd-networkd[1371]: eth0: Link UP Dec 13 02:40:26.268546 systemd-networkd[1371]: eth0: Gained carrier Dec 13 02:40:26.268640 systemd-networkd[1371]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 13 02:40:26.279176 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Dec 13 02:40:26.288001 systemd-networkd[1371]: eth0: DHCPv4 address 172.24.4.241/24, gateway 172.24.4.1 acquired from 172.24.4.1 Dec 13 02:40:26.289063 lvm[1425]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Dec 13 02:40:26.300396 systemd-resolved[1372]: Positive Trust Anchors: Dec 13 02:40:26.300418 systemd-resolved[1372]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Dec 13 02:40:26.300467 systemd-resolved[1372]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Dec 13 02:40:26.308953 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Dec 13 02:40:26.313267 systemd-resolved[1372]: Using system hostname 'ci-4081-2-1-b-3218de15a5.novalocal'. Dec 13 02:40:26.318458 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Dec 13 02:40:26.319301 systemd[1]: Reached target network.target - Network. Dec 13 02:40:26.319807 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Dec 13 02:40:26.335368 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Dec 13 02:40:26.342438 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Dec 13 02:40:26.343482 systemd[1]: Reached target sysinit.target - System Initialization. Dec 13 02:40:26.346638 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Dec 13 02:40:26.347322 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Dec 13 02:40:26.347874 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Dec 13 02:40:26.351371 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Dec 13 02:40:26.351417 systemd[1]: Reached target paths.target - Path Units. Dec 13 02:40:26.351945 systemd[1]: Reached target time-set.target - System Time Set. Dec 13 02:40:26.352677 systemd[1]: Started logrotate.timer - Daily rotation of log files. Dec 13 02:40:26.356618 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Dec 13 02:40:26.357496 systemd[1]: Reached target timers.target - Timer Units. Dec 13 02:40:26.360285 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Dec 13 02:40:26.364304 systemd[1]: Starting docker.socket - Docker Socket for the API... Dec 13 02:40:26.372945 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Dec 13 02:40:26.375678 systemd[1]: Listening on docker.socket - Docker Socket for the API. Dec 13 02:40:26.377693 systemd[1]: Reached target sockets.target - Socket Units. Dec 13 02:40:26.378202 systemd[1]: Reached target basic.target - Basic System. Dec 13 02:40:26.378770 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Dec 13 02:40:26.378809 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Dec 13 02:40:26.381157 systemd-timesyncd[1417]: Contacted time server 5.39.80.51:123 (0.flatcar.pool.ntp.org). Dec 13 02:40:26.381227 systemd-timesyncd[1417]: Initial clock synchronization to Fri 2024-12-13 02:40:26.274758 UTC. Dec 13 02:40:26.387050 systemd[1]: Starting containerd.service - containerd container runtime... Dec 13 02:40:26.392197 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Dec 13 02:40:26.399045 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Dec 13 02:40:26.405949 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Dec 13 02:40:26.416242 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Dec 13 02:40:26.416989 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Dec 13 02:40:26.421074 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Dec 13 02:40:26.426219 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Dec 13 02:40:26.427273 jq[1436]: false Dec 13 02:40:26.434297 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Dec 13 02:40:26.448948 systemd[1]: Starting systemd-logind.service - User Login Management... Dec 13 02:40:26.450460 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Dec 13 02:40:26.454697 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Dec 13 02:40:26.462911 systemd[1]: Starting update-engine.service - Update Engine... Dec 13 02:40:26.473830 extend-filesystems[1437]: Found loop4 Dec 13 02:40:26.477351 extend-filesystems[1437]: Found loop5 Dec 13 02:40:26.477351 extend-filesystems[1437]: Found loop6 Dec 13 02:40:26.477351 extend-filesystems[1437]: Found loop7 Dec 13 02:40:26.477351 extend-filesystems[1437]: Found vda Dec 13 02:40:26.477351 extend-filesystems[1437]: Found vda1 Dec 13 02:40:26.477351 extend-filesystems[1437]: Found vda2 Dec 13 02:40:26.477351 extend-filesystems[1437]: Found vda3 Dec 13 02:40:26.477351 extend-filesystems[1437]: Found usr Dec 13 02:40:26.477351 extend-filesystems[1437]: Found vda4 Dec 13 02:40:26.477351 extend-filesystems[1437]: Found vda6 Dec 13 02:40:26.477351 extend-filesystems[1437]: Found vda7 Dec 13 02:40:26.477351 extend-filesystems[1437]: Found vda9 Dec 13 02:40:26.477351 extend-filesystems[1437]: Checking size of /dev/vda9 Dec 13 02:40:26.474046 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Dec 13 02:40:26.487481 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Dec 13 02:40:26.487687 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Dec 13 02:40:26.519548 extend-filesystems[1437]: Resized partition /dev/vda9 Dec 13 02:40:26.488036 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Dec 13 02:40:26.488181 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Dec 13 02:40:26.528272 extend-filesystems[1458]: resize2fs 1.47.1 (20-May-2024) Dec 13 02:40:26.542526 kernel: EXT4-fs (vda9): resizing filesystem from 1617920 to 4635643 blocks Dec 13 02:40:26.548865 jq[1447]: true Dec 13 02:40:26.556092 systemd[1]: Started dbus.service - D-Bus System Message Bus. Dec 13 02:40:26.555114 dbus-daemon[1435]: [system] SELinux support is enabled Dec 13 02:40:26.572713 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Dec 13 02:40:26.572756 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Dec 13 02:40:26.581310 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Dec 13 02:40:26.581343 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Dec 13 02:40:26.592385 update_engine[1445]: I20241213 02:40:26.589699 1445 main.cc:92] Flatcar Update Engine starting Dec 13 02:40:26.589423 systemd[1]: motdgen.service: Deactivated successfully. Dec 13 02:40:26.589692 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Dec 13 02:40:26.597768 (ntainerd)[1466]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Dec 13 02:40:26.604771 jq[1465]: true Dec 13 02:40:26.607627 systemd[1]: Started update-engine.service - Update Engine. Dec 13 02:40:26.609196 update_engine[1445]: I20241213 02:40:26.609120 1445 update_check_scheduler.cc:74] Next update check in 7m10s Dec 13 02:40:26.621798 systemd[1]: Started locksmithd.service - Cluster reboot manager. Dec 13 02:40:26.651959 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 41 scanned by (udev-worker) (1294) Dec 13 02:40:26.654980 systemd-logind[1441]: New seat seat0. Dec 13 02:40:26.659877 systemd-logind[1441]: Watching system buttons on /dev/input/event1 (Power Button) Dec 13 02:40:26.659929 systemd-logind[1441]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Dec 13 02:40:26.660299 systemd[1]: Started systemd-logind.service - User Login Management. Dec 13 02:40:26.666768 kernel: EXT4-fs (vda9): resized filesystem to 4635643 Dec 13 02:40:26.757943 extend-filesystems[1458]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Dec 13 02:40:26.757943 extend-filesystems[1458]: old_desc_blocks = 1, new_desc_blocks = 3 Dec 13 02:40:26.757943 extend-filesystems[1458]: The filesystem on /dev/vda9 is now 4635643 (4k) blocks long. Dec 13 02:40:26.777961 extend-filesystems[1437]: Resized filesystem in /dev/vda9 Dec 13 02:40:26.760853 systemd[1]: extend-filesystems.service: Deactivated successfully. Dec 13 02:40:26.761109 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Dec 13 02:40:26.829876 bash[1487]: Updated "/home/core/.ssh/authorized_keys" Dec 13 02:40:26.836283 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Dec 13 02:40:26.851389 systemd[1]: Starting sshkeys.service... Dec 13 02:40:26.877349 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Dec 13 02:40:26.893252 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Dec 13 02:40:26.906031 locksmithd[1472]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Dec 13 02:40:27.421646 sshd_keygen[1459]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Dec 13 02:40:27.456512 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Dec 13 02:40:27.478152 systemd[1]: Starting issuegen.service - Generate /run/issue... Dec 13 02:40:27.499790 systemd[1]: issuegen.service: Deactivated successfully. Dec 13 02:40:27.501399 systemd[1]: Finished issuegen.service - Generate /run/issue. Dec 13 02:40:27.514842 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Dec 13 02:40:27.534815 containerd[1466]: time="2024-12-13T02:40:27.534700504Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Dec 13 02:40:27.545171 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Dec 13 02:40:27.557555 systemd[1]: Started getty@tty1.service - Getty on tty1. Dec 13 02:40:27.568541 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Dec 13 02:40:27.570314 containerd[1466]: time="2024-12-13T02:40:27.570231029Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Dec 13 02:40:27.572056 systemd[1]: Reached target getty.target - Login Prompts. Dec 13 02:40:27.573792 containerd[1466]: time="2024-12-13T02:40:27.573741405Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.65-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Dec 13 02:40:27.574482 containerd[1466]: time="2024-12-13T02:40:27.573930598Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Dec 13 02:40:27.574482 containerd[1466]: time="2024-12-13T02:40:27.573960176Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Dec 13 02:40:27.574482 containerd[1466]: time="2024-12-13T02:40:27.574242298Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Dec 13 02:40:27.574482 containerd[1466]: time="2024-12-13T02:40:27.574263611Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Dec 13 02:40:27.574482 containerd[1466]: time="2024-12-13T02:40:27.574373610Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Dec 13 02:40:27.574482 containerd[1466]: time="2024-12-13T02:40:27.574390505Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Dec 13 02:40:27.577518 containerd[1466]: time="2024-12-13T02:40:27.577477076Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Dec 13 02:40:27.577615 containerd[1466]: time="2024-12-13T02:40:27.577599986Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Dec 13 02:40:27.577689 containerd[1466]: time="2024-12-13T02:40:27.577672637Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Dec 13 02:40:27.577771 containerd[1466]: time="2024-12-13T02:40:27.577751644Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Dec 13 02:40:27.578007 containerd[1466]: time="2024-12-13T02:40:27.577982981Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Dec 13 02:40:27.578557 containerd[1466]: time="2024-12-13T02:40:27.578531742Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Dec 13 02:40:27.578914 containerd[1466]: time="2024-12-13T02:40:27.578802938Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Dec 13 02:40:27.578914 containerd[1466]: time="2024-12-13T02:40:27.578828839Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Dec 13 02:40:27.579002 containerd[1466]: time="2024-12-13T02:40:27.578980922Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Dec 13 02:40:27.579063 containerd[1466]: time="2024-12-13T02:40:27.579041176Z" level=info msg="metadata content store policy set" policy=shared Dec 13 02:40:27.585332 containerd[1466]: time="2024-12-13T02:40:27.585303834Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Dec 13 02:40:27.585385 containerd[1466]: time="2024-12-13T02:40:27.585354102Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Dec 13 02:40:27.585385 containerd[1466]: time="2024-12-13T02:40:27.585373183Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Dec 13 02:40:27.585436 containerd[1466]: time="2024-12-13T02:40:27.585389553Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Dec 13 02:40:27.585436 containerd[1466]: time="2024-12-13T02:40:27.585405519Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Dec 13 02:40:27.585568 containerd[1466]: time="2024-12-13T02:40:27.585542575Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Dec 13 02:40:27.585825 containerd[1466]: time="2024-12-13T02:40:27.585807564Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Dec 13 02:40:27.585973 containerd[1466]: time="2024-12-13T02:40:27.585949632Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Dec 13 02:40:27.586002 containerd[1466]: time="2024-12-13T02:40:27.585976137Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Dec 13 02:40:27.586002 containerd[1466]: time="2024-12-13T02:40:27.585992250Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Dec 13 02:40:27.586053 containerd[1466]: time="2024-12-13T02:40:27.586013584Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Dec 13 02:40:27.586053 containerd[1466]: time="2024-12-13T02:40:27.586028739Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Dec 13 02:40:27.586053 containerd[1466]: time="2024-12-13T02:40:27.586046751Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Dec 13 02:40:27.586115 containerd[1466]: time="2024-12-13T02:40:27.586063606Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Dec 13 02:40:27.586115 containerd[1466]: time="2024-12-13T02:40:27.586081074Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Dec 13 02:40:27.586115 containerd[1466]: time="2024-12-13T02:40:27.586096416Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Dec 13 02:40:27.586174 containerd[1466]: time="2024-12-13T02:40:27.586113400Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Dec 13 02:40:27.586174 containerd[1466]: time="2024-12-13T02:40:27.586126658Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Dec 13 02:40:27.586174 containerd[1466]: time="2024-12-13T02:40:27.586148959Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Dec 13 02:40:27.586174 containerd[1466]: time="2024-12-13T02:40:27.586164629Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Dec 13 02:40:27.586254 containerd[1466]: time="2024-12-13T02:40:27.586179091Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Dec 13 02:40:27.586254 containerd[1466]: time="2024-12-13T02:40:27.586193604Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Dec 13 02:40:27.586254 containerd[1466]: time="2024-12-13T02:40:27.586207315Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Dec 13 02:40:27.586254 containerd[1466]: time="2024-12-13T02:40:27.586222243Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Dec 13 02:40:27.586254 containerd[1466]: time="2024-12-13T02:40:27.586236014Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Dec 13 02:40:27.586357 containerd[1466]: time="2024-12-13T02:40:27.586252068Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Dec 13 02:40:27.586357 containerd[1466]: time="2024-12-13T02:40:27.586267737Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Dec 13 02:40:27.586357 containerd[1466]: time="2024-12-13T02:40:27.586284474Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Dec 13 02:40:27.586357 containerd[1466]: time="2024-12-13T02:40:27.586298789Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Dec 13 02:40:27.586357 containerd[1466]: time="2024-12-13T02:40:27.586313330Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Dec 13 02:40:27.586357 containerd[1466]: time="2024-12-13T02:40:27.586327200Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Dec 13 02:40:27.586357 containerd[1466]: time="2024-12-13T02:40:27.586344787Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Dec 13 02:40:27.586505 containerd[1466]: time="2024-12-13T02:40:27.586367267Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Dec 13 02:40:27.586505 containerd[1466]: time="2024-12-13T02:40:27.586380831Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Dec 13 02:40:27.586505 containerd[1466]: time="2024-12-13T02:40:27.586393613Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Dec 13 02:40:27.586505 containerd[1466]: time="2024-12-13T02:40:27.586443319Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Dec 13 02:40:27.586505 containerd[1466]: time="2024-12-13T02:40:27.586463951Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Dec 13 02:40:27.586505 containerd[1466]: time="2024-12-13T02:40:27.586476041Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Dec 13 02:40:27.586505 containerd[1466]: time="2024-12-13T02:40:27.586490068Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Dec 13 02:40:27.586505 containerd[1466]: time="2024-12-13T02:40:27.586501654Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Dec 13 02:40:27.586664 containerd[1466]: time="2024-12-13T02:40:27.586515169Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Dec 13 02:40:27.586664 containerd[1466]: time="2024-12-13T02:40:27.586530897Z" level=info msg="NRI interface is disabled by configuration." Dec 13 02:40:27.586664 containerd[1466]: time="2024-12-13T02:40:27.586542384Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Dec 13 02:40:27.587002 containerd[1466]: time="2024-12-13T02:40:27.586922038Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Dec 13 02:40:27.587142 containerd[1466]: time="2024-12-13T02:40:27.587006473Z" level=info msg="Connect containerd service" Dec 13 02:40:27.587142 containerd[1466]: time="2024-12-13T02:40:27.587041349Z" level=info msg="using legacy CRI server" Dec 13 02:40:27.587142 containerd[1466]: time="2024-12-13T02:40:27.587049031Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Dec 13 02:40:27.587231 containerd[1466]: time="2024-12-13T02:40:27.587168530Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Dec 13 02:40:27.587859 containerd[1466]: time="2024-12-13T02:40:27.587819806Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Dec 13 02:40:27.588156 containerd[1466]: time="2024-12-13T02:40:27.588024066Z" level=info msg="Start subscribing containerd event" Dec 13 02:40:27.588156 containerd[1466]: time="2024-12-13T02:40:27.588091270Z" level=info msg="Start recovering state" Dec 13 02:40:27.588225 containerd[1466]: time="2024-12-13T02:40:27.588207438Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Dec 13 02:40:27.588285 containerd[1466]: time="2024-12-13T02:40:27.588269649Z" level=info msg="Start event monitor" Dec 13 02:40:27.589913 containerd[1466]: time="2024-12-13T02:40:27.588330417Z" level=info msg="Start snapshots syncer" Dec 13 02:40:27.589913 containerd[1466]: time="2024-12-13T02:40:27.588346165Z" level=info msg="Start cni network conf syncer for default" Dec 13 02:40:27.589913 containerd[1466]: time="2024-12-13T02:40:27.588295046Z" level=info msg=serving... address=/run/containerd/containerd.sock Dec 13 02:40:27.589913 containerd[1466]: time="2024-12-13T02:40:27.588355962Z" level=info msg="Start streaming server" Dec 13 02:40:27.589913 containerd[1466]: time="2024-12-13T02:40:27.588460574Z" level=info msg="containerd successfully booted in 0.057889s" Dec 13 02:40:27.588663 systemd[1]: Started containerd.service - containerd container runtime. Dec 13 02:40:28.060354 systemd-networkd[1371]: eth0: Gained IPv6LL Dec 13 02:40:28.065326 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Dec 13 02:40:28.073515 systemd[1]: Reached target network-online.target - Network is Online. Dec 13 02:40:28.085412 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 13 02:40:28.102114 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Dec 13 02:40:28.144553 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Dec 13 02:40:31.603253 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 13 02:40:31.617643 (kubelet)[1543]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 13 02:40:32.218018 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Dec 13 02:40:32.232686 systemd[1]: Started sshd@0-172.24.4.241:22-172.24.4.1:50132.service - OpenSSH per-connection server daemon (172.24.4.1:50132). Dec 13 02:40:33.589762 coreos-metadata[1434]: Dec 13 02:40:33.589 WARN failed to locate config-drive, using the metadata service API instead Dec 13 02:40:33.715994 coreos-metadata[1434]: Dec 13 02:40:33.715 INFO Fetching http://169.254.169.254/openstack/2012-08-10/meta_data.json: Attempt #1 Dec 13 02:40:33.730738 sshd[1545]: Accepted publickey for core from 172.24.4.1 port 50132 ssh2: RSA SHA256:s+jMJkc8yzesvkj+g1MqwY5XQAL52YjwOYy7JiKKino Dec 13 02:40:33.791603 login[1520]: pam_unix(login:session): session opened for user core(uid=500) by LOGIN(uid=0) Dec 13 02:40:33.792181 sshd[1545]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 02:40:33.800027 login[1521]: pam_unix(login:session): session opened for user core(uid=500) by LOGIN(uid=0) Dec 13 02:40:33.832645 systemd-logind[1441]: New session 3 of user core. Dec 13 02:40:33.838241 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Dec 13 02:40:33.848564 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Dec 13 02:40:33.853009 systemd-logind[1441]: New session 1 of user core. Dec 13 02:40:33.867614 systemd-logind[1441]: New session 2 of user core. Dec 13 02:40:33.881389 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Dec 13 02:40:33.893430 systemd[1]: Starting user@500.service - User Manager for UID 500... Dec 13 02:40:33.923036 (systemd)[1560]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Dec 13 02:40:33.950828 coreos-metadata[1434]: Dec 13 02:40:33.950 INFO Fetch successful Dec 13 02:40:33.950828 coreos-metadata[1434]: Dec 13 02:40:33.950 INFO Fetching http://169.254.169.254/latest/meta-data/hostname: Attempt #1 Dec 13 02:40:33.966099 coreos-metadata[1434]: Dec 13 02:40:33.966 INFO Fetch successful Dec 13 02:40:33.966251 coreos-metadata[1434]: Dec 13 02:40:33.966 INFO Fetching http://169.254.169.254/latest/meta-data/instance-id: Attempt #1 Dec 13 02:40:33.973482 coreos-metadata[1496]: Dec 13 02:40:33.973 WARN failed to locate config-drive, using the metadata service API instead Dec 13 02:40:33.984013 coreos-metadata[1434]: Dec 13 02:40:33.983 INFO Fetch successful Dec 13 02:40:33.984013 coreos-metadata[1434]: Dec 13 02:40:33.984 INFO Fetching http://169.254.169.254/latest/meta-data/instance-type: Attempt #1 Dec 13 02:40:33.992023 coreos-metadata[1496]: Dec 13 02:40:33.991 INFO Fetching http://169.254.169.254/latest/meta-data/public-keys: Attempt #1 Dec 13 02:40:33.998439 coreos-metadata[1434]: Dec 13 02:40:33.998 INFO Fetch successful Dec 13 02:40:33.998651 coreos-metadata[1434]: Dec 13 02:40:33.998 INFO Fetching http://169.254.169.254/latest/meta-data/local-ipv4: Attempt #1 Dec 13 02:40:34.006262 coreos-metadata[1496]: Dec 13 02:40:34.006 INFO Fetch successful Dec 13 02:40:34.006394 coreos-metadata[1496]: Dec 13 02:40:34.006 INFO Fetching http://169.254.169.254/latest/meta-data/public-keys/0/openssh-key: Attempt #1 Dec 13 02:40:34.011066 coreos-metadata[1434]: Dec 13 02:40:34.011 INFO Fetch successful Dec 13 02:40:34.011204 coreos-metadata[1434]: Dec 13 02:40:34.011 INFO Fetching http://169.254.169.254/latest/meta-data/public-ipv4: Attempt #1 Dec 13 02:40:34.019058 coreos-metadata[1434]: Dec 13 02:40:34.019 INFO Fetch successful Dec 13 02:40:34.023910 coreos-metadata[1496]: Dec 13 02:40:34.022 INFO Fetch successful Dec 13 02:40:34.041522 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Dec 13 02:40:34.043571 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Dec 13 02:40:34.354230 unknown[1496]: wrote ssh authorized keys file for user: core Dec 13 02:40:34.557363 systemd[1560]: Queued start job for default target default.target. Dec 13 02:40:34.567006 systemd[1560]: Created slice app.slice - User Application Slice. Dec 13 02:40:34.567037 systemd[1560]: Reached target paths.target - Paths. Dec 13 02:40:34.567052 systemd[1560]: Reached target timers.target - Timers. Dec 13 02:40:34.570994 systemd[1560]: Starting dbus.socket - D-Bus User Message Bus Socket... Dec 13 02:40:34.592407 systemd[1560]: Listening on dbus.socket - D-Bus User Message Bus Socket. Dec 13 02:40:34.592547 systemd[1560]: Reached target sockets.target - Sockets. Dec 13 02:40:34.592566 systemd[1560]: Reached target basic.target - Basic System. Dec 13 02:40:34.592614 systemd[1560]: Reached target default.target - Main User Target. Dec 13 02:40:34.592645 systemd[1560]: Startup finished in 662ms. Dec 13 02:40:34.592936 systemd[1]: Started user@500.service - User Manager for UID 500. Dec 13 02:40:34.600356 systemd[1]: Started session-1.scope - Session 1 of User core. Dec 13 02:40:34.602332 systemd[1]: Started session-2.scope - Session 2 of User core. Dec 13 02:40:34.605169 systemd[1]: Started session-3.scope - Session 3 of User core. Dec 13 02:40:34.806448 update-ssh-keys[1573]: Updated "/home/core/.ssh/authorized_keys" Dec 13 02:40:34.805604 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Dec 13 02:40:34.811696 systemd[1]: Finished sshkeys.service. Dec 13 02:40:34.819870 systemd[1]: Reached target multi-user.target - Multi-User System. Dec 13 02:40:34.820503 systemd[1]: Startup finished in 1.171s (kernel) + 18.416s (initrd) + 15.654s (userspace) = 35.242s. Dec 13 02:40:34.990613 systemd[1]: Started sshd@1-172.24.4.241:22-172.24.4.1:55092.service - OpenSSH per-connection server daemon (172.24.4.1:55092). Dec 13 02:40:36.980300 sshd[1605]: Accepted publickey for core from 172.24.4.1 port 55092 ssh2: RSA SHA256:s+jMJkc8yzesvkj+g1MqwY5XQAL52YjwOYy7JiKKino Dec 13 02:40:36.986158 sshd[1605]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 02:40:36.999744 systemd-logind[1441]: New session 4 of user core. Dec 13 02:40:37.005242 systemd[1]: Started session-4.scope - Session 4 of User core. Dec 13 02:40:37.492209 kubelet[1543]: E1213 02:40:37.491971 1543 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 13 02:40:37.497500 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 13 02:40:37.497883 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 13 02:40:37.498794 systemd[1]: kubelet.service: Consumed 2.646s CPU time. Dec 13 02:40:37.824988 sshd[1605]: pam_unix(sshd:session): session closed for user core Dec 13 02:40:37.834368 systemd[1]: sshd@1-172.24.4.241:22-172.24.4.1:55092.service: Deactivated successfully. Dec 13 02:40:37.837957 systemd[1]: session-4.scope: Deactivated successfully. Dec 13 02:40:37.839638 systemd-logind[1441]: Session 4 logged out. Waiting for processes to exit. Dec 13 02:40:37.847533 systemd[1]: Started sshd@2-172.24.4.241:22-172.24.4.1:55096.service - OpenSSH per-connection server daemon (172.24.4.1:55096). Dec 13 02:40:37.850320 systemd-logind[1441]: Removed session 4. Dec 13 02:40:39.161287 sshd[1614]: Accepted publickey for core from 172.24.4.1 port 55096 ssh2: RSA SHA256:s+jMJkc8yzesvkj+g1MqwY5XQAL52YjwOYy7JiKKino Dec 13 02:40:39.163866 sshd[1614]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 02:40:39.172948 systemd-logind[1441]: New session 5 of user core. Dec 13 02:40:39.178113 systemd[1]: Started session-5.scope - Session 5 of User core. Dec 13 02:40:39.804216 sshd[1614]: pam_unix(sshd:session): session closed for user core Dec 13 02:40:39.811638 systemd[1]: sshd@2-172.24.4.241:22-172.24.4.1:55096.service: Deactivated successfully. Dec 13 02:40:39.814481 systemd[1]: session-5.scope: Deactivated successfully. Dec 13 02:40:39.816010 systemd-logind[1441]: Session 5 logged out. Waiting for processes to exit. Dec 13 02:40:39.823437 systemd[1]: Started sshd@3-172.24.4.241:22-172.24.4.1:55100.service - OpenSSH per-connection server daemon (172.24.4.1:55100). Dec 13 02:40:39.826198 systemd-logind[1441]: Removed session 5. Dec 13 02:40:40.952364 sshd[1621]: Accepted publickey for core from 172.24.4.1 port 55100 ssh2: RSA SHA256:s+jMJkc8yzesvkj+g1MqwY5XQAL52YjwOYy7JiKKino Dec 13 02:40:40.954957 sshd[1621]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 02:40:40.964942 systemd-logind[1441]: New session 6 of user core. Dec 13 02:40:40.975271 systemd[1]: Started session-6.scope - Session 6 of User core. Dec 13 02:40:41.592864 sshd[1621]: pam_unix(sshd:session): session closed for user core Dec 13 02:40:41.602886 systemd[1]: sshd@3-172.24.4.241:22-172.24.4.1:55100.service: Deactivated successfully. Dec 13 02:40:41.606590 systemd[1]: session-6.scope: Deactivated successfully. Dec 13 02:40:41.608216 systemd-logind[1441]: Session 6 logged out. Waiting for processes to exit. Dec 13 02:40:41.615541 systemd[1]: Started sshd@4-172.24.4.241:22-172.24.4.1:55116.service - OpenSSH per-connection server daemon (172.24.4.1:55116). Dec 13 02:40:41.619590 systemd-logind[1441]: Removed session 6. Dec 13 02:40:42.967693 sshd[1628]: Accepted publickey for core from 172.24.4.1 port 55116 ssh2: RSA SHA256:s+jMJkc8yzesvkj+g1MqwY5XQAL52YjwOYy7JiKKino Dec 13 02:40:42.970644 sshd[1628]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 02:40:42.980524 systemd-logind[1441]: New session 7 of user core. Dec 13 02:40:42.993217 systemd[1]: Started session-7.scope - Session 7 of User core. Dec 13 02:40:43.304611 sudo[1631]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Dec 13 02:40:43.305325 sudo[1631]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Dec 13 02:40:43.324210 sudo[1631]: pam_unix(sudo:session): session closed for user root Dec 13 02:40:43.586755 sshd[1628]: pam_unix(sshd:session): session closed for user core Dec 13 02:40:43.597469 systemd[1]: sshd@4-172.24.4.241:22-172.24.4.1:55116.service: Deactivated successfully. Dec 13 02:40:43.600528 systemd[1]: session-7.scope: Deactivated successfully. Dec 13 02:40:43.602028 systemd-logind[1441]: Session 7 logged out. Waiting for processes to exit. Dec 13 02:40:43.609512 systemd[1]: Started sshd@5-172.24.4.241:22-172.24.4.1:55120.service - OpenSSH per-connection server daemon (172.24.4.1:55120). Dec 13 02:40:43.611986 systemd-logind[1441]: Removed session 7. Dec 13 02:40:45.102732 sshd[1636]: Accepted publickey for core from 172.24.4.1 port 55120 ssh2: RSA SHA256:s+jMJkc8yzesvkj+g1MqwY5XQAL52YjwOYy7JiKKino Dec 13 02:40:45.134381 sshd[1636]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 02:40:45.146496 systemd-logind[1441]: New session 8 of user core. Dec 13 02:40:45.158422 systemd[1]: Started session-8.scope - Session 8 of User core. Dec 13 02:40:45.725071 sudo[1640]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Dec 13 02:40:45.725892 sudo[1640]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Dec 13 02:40:45.736242 sudo[1640]: pam_unix(sudo:session): session closed for user root Dec 13 02:40:45.749560 sudo[1639]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Dec 13 02:40:45.750265 sudo[1639]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Dec 13 02:40:45.780622 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Dec 13 02:40:45.789686 auditctl[1643]: No rules Dec 13 02:40:45.791085 systemd[1]: audit-rules.service: Deactivated successfully. Dec 13 02:40:45.791627 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Dec 13 02:40:45.800738 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Dec 13 02:40:45.869844 augenrules[1661]: No rules Dec 13 02:40:45.872423 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Dec 13 02:40:45.874790 sudo[1639]: pam_unix(sudo:session): session closed for user root Dec 13 02:40:46.020650 sshd[1636]: pam_unix(sshd:session): session closed for user core Dec 13 02:40:46.034205 systemd[1]: sshd@5-172.24.4.241:22-172.24.4.1:55120.service: Deactivated successfully. Dec 13 02:40:46.037227 systemd[1]: session-8.scope: Deactivated successfully. Dec 13 02:40:46.038803 systemd-logind[1441]: Session 8 logged out. Waiting for processes to exit. Dec 13 02:40:46.047613 systemd[1]: Started sshd@6-172.24.4.241:22-172.24.4.1:48258.service - OpenSSH per-connection server daemon (172.24.4.1:48258). Dec 13 02:40:46.050748 systemd-logind[1441]: Removed session 8. Dec 13 02:40:47.271184 sshd[1669]: Accepted publickey for core from 172.24.4.1 port 48258 ssh2: RSA SHA256:s+jMJkc8yzesvkj+g1MqwY5XQAL52YjwOYy7JiKKino Dec 13 02:40:47.272776 sshd[1669]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 02:40:47.280960 systemd-logind[1441]: New session 9 of user core. Dec 13 02:40:47.284302 systemd[1]: Started session-9.scope - Session 9 of User core. Dec 13 02:40:47.653236 sudo[1672]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Dec 13 02:40:47.653631 sudo[1672]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Dec 13 02:40:47.655186 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Dec 13 02:40:47.663265 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 13 02:40:48.018742 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 13 02:40:48.023546 (kubelet)[1691]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 13 02:40:48.482721 kubelet[1691]: E1213 02:40:48.482184 1691 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 13 02:40:48.492667 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 13 02:40:48.493761 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 13 02:40:50.014689 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Dec 13 02:40:50.031511 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 13 02:40:50.085461 systemd[1]: Reloading requested from client PID 1727 ('systemctl') (unit session-9.scope)... Dec 13 02:40:50.085500 systemd[1]: Reloading... Dec 13 02:40:50.193928 zram_generator::config[1768]: No configuration found. Dec 13 02:40:50.364498 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Dec 13 02:40:50.456403 systemd[1]: Reloading finished in 370 ms. Dec 13 02:40:50.511012 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Dec 13 02:40:50.511095 systemd[1]: kubelet.service: Failed with result 'signal'. Dec 13 02:40:50.511437 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Dec 13 02:40:50.513678 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 13 02:40:51.166851 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 13 02:40:51.172596 (kubelet)[1830]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Dec 13 02:40:51.636713 kubelet[1830]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 13 02:40:51.636713 kubelet[1830]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 13 02:40:51.636713 kubelet[1830]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 13 02:40:51.637522 kubelet[1830]: I1213 02:40:51.637186 1830 server.go:204] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 13 02:40:52.208510 kubelet[1830]: I1213 02:40:52.207792 1830 server.go:487] "Kubelet version" kubeletVersion="v1.29.2" Dec 13 02:40:52.208510 kubelet[1830]: I1213 02:40:52.207858 1830 server.go:489] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 13 02:40:52.208510 kubelet[1830]: I1213 02:40:52.208383 1830 server.go:919] "Client rotation is on, will bootstrap in background" Dec 13 02:40:52.245562 kubelet[1830]: I1213 02:40:52.245489 1830 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Dec 13 02:40:52.282111 kubelet[1830]: I1213 02:40:52.282051 1830 server.go:745] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Dec 13 02:40:52.283003 kubelet[1830]: I1213 02:40:52.282970 1830 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 13 02:40:52.284013 kubelet[1830]: I1213 02:40:52.283607 1830 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} Dec 13 02:40:52.284013 kubelet[1830]: I1213 02:40:52.283805 1830 topology_manager.go:138] "Creating topology manager with none policy" Dec 13 02:40:52.284013 kubelet[1830]: I1213 02:40:52.283838 1830 container_manager_linux.go:301] "Creating device plugin manager" Dec 13 02:40:52.288636 kubelet[1830]: I1213 02:40:52.288533 1830 state_mem.go:36] "Initialized new in-memory state store" Dec 13 02:40:52.288852 kubelet[1830]: I1213 02:40:52.288803 1830 kubelet.go:396] "Attempting to sync node with API server" Dec 13 02:40:52.288852 kubelet[1830]: I1213 02:40:52.288853 1830 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 13 02:40:52.289731 kubelet[1830]: I1213 02:40:52.288972 1830 kubelet.go:312] "Adding apiserver pod source" Dec 13 02:40:52.289731 kubelet[1830]: I1213 02:40:52.289015 1830 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 13 02:40:52.292781 kubelet[1830]: E1213 02:40:52.292456 1830 file.go:98] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:40:52.292781 kubelet[1830]: E1213 02:40:52.292578 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:40:52.296431 kubelet[1830]: I1213 02:40:52.296343 1830 kuberuntime_manager.go:258] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Dec 13 02:40:52.302745 kubelet[1830]: W1213 02:40:52.302699 1830 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.Node: nodes "172.24.4.241" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Dec 13 02:40:52.304938 kubelet[1830]: E1213 02:40:52.303063 1830 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.Node: failed to list *v1.Node: nodes "172.24.4.241" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Dec 13 02:40:52.305500 kubelet[1830]: I1213 02:40:52.305440 1830 kubelet.go:809] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 13 02:40:52.305679 kubelet[1830]: W1213 02:40:52.305636 1830 probe.go:268] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Dec 13 02:40:52.307374 kubelet[1830]: I1213 02:40:52.307322 1830 server.go:1256] "Started kubelet" Dec 13 02:40:52.310210 kubelet[1830]: W1213 02:40:52.310167 1830 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Dec 13 02:40:52.310400 kubelet[1830]: E1213 02:40:52.310376 1830 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Dec 13 02:40:52.310635 kubelet[1830]: I1213 02:40:52.310605 1830 server.go:162] "Starting to listen" address="0.0.0.0" port=10250 Dec 13 02:40:52.312709 kubelet[1830]: I1213 02:40:52.312672 1830 server.go:461] "Adding debug handlers to kubelet server" Dec 13 02:40:52.317306 kubelet[1830]: I1213 02:40:52.317235 1830 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 13 02:40:52.317848 kubelet[1830]: I1213 02:40:52.317789 1830 server.go:233] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 13 02:40:52.320327 kubelet[1830]: I1213 02:40:52.320247 1830 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 13 02:40:52.335793 kubelet[1830]: I1213 02:40:52.335715 1830 volume_manager.go:291] "Starting Kubelet Volume Manager" Dec 13 02:40:52.338377 kubelet[1830]: I1213 02:40:52.336063 1830 desired_state_of_world_populator.go:151] "Desired state populator starts to run" Dec 13 02:40:52.338641 kubelet[1830]: I1213 02:40:52.338539 1830 reconciler_new.go:29] "Reconciler: start to sync state" Dec 13 02:40:52.341137 kubelet[1830]: E1213 02:40:52.341021 1830 event.go:346] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{172.24.4.241.18109c3f51d28ede default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:172.24.4.241,UID:172.24.4.241,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:172.24.4.241,},FirstTimestamp:2024-12-13 02:40:52.307267294 +0000 UTC m=+1.124303185,LastTimestamp:2024-12-13 02:40:52.307267294 +0000 UTC m=+1.124303185,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:172.24.4.241,}" Dec 13 02:40:52.342828 kubelet[1830]: E1213 02:40:52.341665 1830 kubelet.go:1462] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Dec 13 02:40:52.344140 kubelet[1830]: I1213 02:40:52.344104 1830 factory.go:221] Registration of the systemd container factory successfully Dec 13 02:40:52.347149 kubelet[1830]: I1213 02:40:52.347099 1830 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Dec 13 02:40:52.352818 kubelet[1830]: E1213 02:40:52.351728 1830 event.go:346] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{172.24.4.241.18109c3f53decd45 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:172.24.4.241,UID:172.24.4.241,APIVersion:,ResourceVersion:,FieldPath:,},Reason:InvalidDiskCapacity,Message:invalid capacity 0 on image filesystem,Source:EventSource{Component:kubelet,Host:172.24.4.241,},FirstTimestamp:2024-12-13 02:40:52.341624133 +0000 UTC m=+1.158659964,LastTimestamp:2024-12-13 02:40:52.341624133 +0000 UTC m=+1.158659964,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:172.24.4.241,}" Dec 13 02:40:52.353666 kubelet[1830]: I1213 02:40:52.353457 1830 factory.go:221] Registration of the containerd container factory successfully Dec 13 02:40:52.373997 kubelet[1830]: W1213 02:40:52.373959 1830 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Dec 13 02:40:52.374206 kubelet[1830]: E1213 02:40:52.374193 1830 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Dec 13 02:40:52.374394 kubelet[1830]: E1213 02:40:52.374380 1830 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"172.24.4.241\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="200ms" Dec 13 02:40:52.388718 kubelet[1830]: E1213 02:40:52.388674 1830 event.go:346] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{172.24.4.241.18109c3f5677c412 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:172.24.4.241,UID:172.24.4.241,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node 172.24.4.241 status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:172.24.4.241,},FirstTimestamp:2024-12-13 02:40:52.385203218 +0000 UTC m=+1.202238999,LastTimestamp:2024-12-13 02:40:52.385203218 +0000 UTC m=+1.202238999,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:172.24.4.241,}" Dec 13 02:40:52.390183 kubelet[1830]: I1213 02:40:52.390167 1830 cpu_manager.go:214] "Starting CPU manager" policy="none" Dec 13 02:40:52.390276 kubelet[1830]: I1213 02:40:52.390266 1830 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Dec 13 02:40:52.390351 kubelet[1830]: I1213 02:40:52.390342 1830 state_mem.go:36] "Initialized new in-memory state store" Dec 13 02:40:52.398051 kubelet[1830]: I1213 02:40:52.397984 1830 policy_none.go:49] "None policy: Start" Dec 13 02:40:52.399568 kubelet[1830]: E1213 02:40:52.399476 1830 event.go:346] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{172.24.4.241.18109c3f5677dfd8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:172.24.4.241,UID:172.24.4.241,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node 172.24.4.241 status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:172.24.4.241,},FirstTimestamp:2024-12-13 02:40:52.385210328 +0000 UTC m=+1.202246109,LastTimestamp:2024-12-13 02:40:52.385210328 +0000 UTC m=+1.202246109,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:172.24.4.241,}" Dec 13 02:40:52.400127 kubelet[1830]: I1213 02:40:52.400109 1830 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 13 02:40:52.400215 kubelet[1830]: I1213 02:40:52.400204 1830 state_mem.go:35] "Initializing new in-memory state store" Dec 13 02:40:52.437726 kubelet[1830]: I1213 02:40:52.437697 1830 kubelet_node_status.go:73] "Attempting to register node" node="172.24.4.241" Dec 13 02:40:52.455690 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Dec 13 02:40:52.481762 kubelet[1830]: I1213 02:40:52.480594 1830 kubelet_node_status.go:76] "Successfully registered node" node="172.24.4.241" Dec 13 02:40:52.490810 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Dec 13 02:40:52.510984 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Dec 13 02:40:52.520876 kubelet[1830]: I1213 02:40:52.520101 1830 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 13 02:40:52.520876 kubelet[1830]: I1213 02:40:52.520462 1830 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 13 02:40:52.526147 kubelet[1830]: E1213 02:40:52.526099 1830 eviction_manager.go:282] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"172.24.4.241\" not found" Dec 13 02:40:52.597386 kubelet[1830]: E1213 02:40:52.596887 1830 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"172.24.4.241\" not found" Dec 13 02:40:52.609112 kubelet[1830]: I1213 02:40:52.608846 1830 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 13 02:40:52.611551 kubelet[1830]: I1213 02:40:52.611396 1830 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 13 02:40:52.612075 kubelet[1830]: I1213 02:40:52.611720 1830 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 13 02:40:52.612075 kubelet[1830]: I1213 02:40:52.612028 1830 kubelet.go:2329] "Starting kubelet main sync loop" Dec 13 02:40:52.612241 kubelet[1830]: E1213 02:40:52.612206 1830 kubelet.go:2353] "Skipping pod synchronization" err="PLEG is not healthy: pleg has yet to be successful" Dec 13 02:40:52.698303 kubelet[1830]: E1213 02:40:52.698209 1830 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"172.24.4.241\" not found" Dec 13 02:40:52.800026 kubelet[1830]: E1213 02:40:52.799605 1830 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"172.24.4.241\" not found" Dec 13 02:40:52.900008 kubelet[1830]: E1213 02:40:52.899824 1830 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"172.24.4.241\" not found" Dec 13 02:40:53.001110 kubelet[1830]: E1213 02:40:53.000986 1830 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"172.24.4.241\" not found" Dec 13 02:40:53.101546 kubelet[1830]: E1213 02:40:53.101273 1830 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"172.24.4.241\" not found" Dec 13 02:40:53.202400 kubelet[1830]: E1213 02:40:53.202124 1830 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"172.24.4.241\" not found" Dec 13 02:40:53.213722 kubelet[1830]: I1213 02:40:53.213537 1830 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Dec 13 02:40:53.214022 kubelet[1830]: W1213 02:40:53.213836 1830 reflector.go:462] vendor/k8s.io/client-go/informers/factory.go:159: watch of *v1.RuntimeClass ended with: very short watch: vendor/k8s.io/client-go/informers/factory.go:159: Unexpected watch close - watch lasted less than a second and no items received Dec 13 02:40:53.292972 kubelet[1830]: E1213 02:40:53.292764 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:40:53.303355 kubelet[1830]: E1213 02:40:53.303257 1830 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"172.24.4.241\" not found" Dec 13 02:40:53.405317 kubelet[1830]: I1213 02:40:53.404990 1830 kuberuntime_manager.go:1529] "Updating runtime config through cri with podcidr" CIDR="192.168.1.0/24" Dec 13 02:40:53.406072 containerd[1466]: time="2024-12-13T02:40:53.405631201Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Dec 13 02:40:53.407702 kubelet[1830]: I1213 02:40:53.406520 1830 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.1.0/24" Dec 13 02:40:53.728568 sudo[1672]: pam_unix(sudo:session): session closed for user root Dec 13 02:40:54.003584 sshd[1669]: pam_unix(sshd:session): session closed for user core Dec 13 02:40:54.012044 systemd[1]: sshd@6-172.24.4.241:22-172.24.4.1:48258.service: Deactivated successfully. Dec 13 02:40:54.016430 systemd[1]: session-9.scope: Deactivated successfully. Dec 13 02:40:54.017104 systemd[1]: session-9.scope: Consumed 1.084s CPU time, 109.5M memory peak, 0B memory swap peak. Dec 13 02:40:54.019865 systemd-logind[1441]: Session 9 logged out. Waiting for processes to exit. Dec 13 02:40:54.023408 systemd-logind[1441]: Removed session 9. Dec 13 02:40:54.293417 kubelet[1830]: E1213 02:40:54.293105 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:40:54.293417 kubelet[1830]: I1213 02:40:54.293202 1830 apiserver.go:52] "Watching apiserver" Dec 13 02:40:54.306773 kubelet[1830]: I1213 02:40:54.306688 1830 topology_manager.go:215] "Topology Admit Handler" podUID="6bb2af7a-92b5-42f4-ba83-0270d49ce026" podNamespace="calico-system" podName="calico-node-n8hkc" Dec 13 02:40:54.307068 kubelet[1830]: I1213 02:40:54.306842 1830 topology_manager.go:215] "Topology Admit Handler" podUID="ed687165-b432-4311-a017-bee89dd86d6d" podNamespace="calico-system" podName="csi-node-driver-mn4q7" Dec 13 02:40:54.307068 kubelet[1830]: I1213 02:40:54.306912 1830 topology_manager.go:215] "Topology Admit Handler" podUID="99595cfe-3433-47c0-8c88-589f72f1e421" podNamespace="kube-system" podName="kube-proxy-j7n7v" Dec 13 02:40:54.308347 kubelet[1830]: E1213 02:40:54.307488 1830 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-mn4q7" podUID="ed687165-b432-4311-a017-bee89dd86d6d" Dec 13 02:40:54.321738 systemd[1]: Created slice kubepods-besteffort-pod99595cfe_3433_47c0_8c88_589f72f1e421.slice - libcontainer container kubepods-besteffort-pod99595cfe_3433_47c0_8c88_589f72f1e421.slice. Dec 13 02:40:54.334368 systemd[1]: Created slice kubepods-besteffort-pod6bb2af7a_92b5_42f4_ba83_0270d49ce026.slice - libcontainer container kubepods-besteffort-pod6bb2af7a_92b5_42f4_ba83_0270d49ce026.slice. Dec 13 02:40:54.344679 kubelet[1830]: I1213 02:40:54.342836 1830 desired_state_of_world_populator.go:159] "Finished populating initial desired state of world" Dec 13 02:40:54.351979 kubelet[1830]: I1213 02:40:54.351944 1830 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/6bb2af7a-92b5-42f4-ba83-0270d49ce026-flexvol-driver-host\") pod \"calico-node-n8hkc\" (UID: \"6bb2af7a-92b5-42f4-ba83-0270d49ce026\") " pod="calico-system/calico-node-n8hkc" Dec 13 02:40:54.352273 kubelet[1830]: I1213 02:40:54.352260 1830 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ed687165-b432-4311-a017-bee89dd86d6d-kubelet-dir\") pod \"csi-node-driver-mn4q7\" (UID: \"ed687165-b432-4311-a017-bee89dd86d6d\") " pod="calico-system/csi-node-driver-mn4q7" Dec 13 02:40:54.352395 kubelet[1830]: I1213 02:40:54.352384 1830 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99kth\" (UniqueName: \"kubernetes.io/projected/ed687165-b432-4311-a017-bee89dd86d6d-kube-api-access-99kth\") pod \"csi-node-driver-mn4q7\" (UID: \"ed687165-b432-4311-a017-bee89dd86d6d\") " pod="calico-system/csi-node-driver-mn4q7" Dec 13 02:40:54.352492 kubelet[1830]: I1213 02:40:54.352481 1830 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/99595cfe-3433-47c0-8c88-589f72f1e421-kube-proxy\") pod \"kube-proxy-j7n7v\" (UID: \"99595cfe-3433-47c0-8c88-589f72f1e421\") " pod="kube-system/kube-proxy-j7n7v" Dec 13 02:40:54.352653 kubelet[1830]: I1213 02:40:54.352641 1830 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/6bb2af7a-92b5-42f4-ba83-0270d49ce026-lib-modules\") pod \"calico-node-n8hkc\" (UID: \"6bb2af7a-92b5-42f4-ba83-0270d49ce026\") " pod="calico-system/calico-node-n8hkc" Dec 13 02:40:54.352757 kubelet[1830]: I1213 02:40:54.352745 1830 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/6bb2af7a-92b5-42f4-ba83-0270d49ce026-xtables-lock\") pod \"calico-node-n8hkc\" (UID: \"6bb2af7a-92b5-42f4-ba83-0270d49ce026\") " pod="calico-system/calico-node-n8hkc" Dec 13 02:40:54.352883 kubelet[1830]: I1213 02:40:54.352872 1830 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/6bb2af7a-92b5-42f4-ba83-0270d49ce026-node-certs\") pod \"calico-node-n8hkc\" (UID: \"6bb2af7a-92b5-42f4-ba83-0270d49ce026\") " pod="calico-system/calico-node-n8hkc" Dec 13 02:40:54.353045 kubelet[1830]: I1213 02:40:54.353029 1830 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/ed687165-b432-4311-a017-bee89dd86d6d-varrun\") pod \"csi-node-driver-mn4q7\" (UID: \"ed687165-b432-4311-a017-bee89dd86d6d\") " pod="calico-system/csi-node-driver-mn4q7" Dec 13 02:40:54.353172 kubelet[1830]: I1213 02:40:54.353161 1830 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcbgw\" (UniqueName: \"kubernetes.io/projected/6bb2af7a-92b5-42f4-ba83-0270d49ce026-kube-api-access-rcbgw\") pod \"calico-node-n8hkc\" (UID: \"6bb2af7a-92b5-42f4-ba83-0270d49ce026\") " pod="calico-system/calico-node-n8hkc" Dec 13 02:40:54.353304 kubelet[1830]: I1213 02:40:54.353292 1830 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/99595cfe-3433-47c0-8c88-589f72f1e421-xtables-lock\") pod \"kube-proxy-j7n7v\" (UID: \"99595cfe-3433-47c0-8c88-589f72f1e421\") " pod="kube-system/kube-proxy-j7n7v" Dec 13 02:40:54.353429 kubelet[1830]: I1213 02:40:54.353418 1830 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/99595cfe-3433-47c0-8c88-589f72f1e421-lib-modules\") pod \"kube-proxy-j7n7v\" (UID: \"99595cfe-3433-47c0-8c88-589f72f1e421\") " pod="kube-system/kube-proxy-j7n7v" Dec 13 02:40:54.353547 kubelet[1830]: I1213 02:40:54.353536 1830 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6bb2af7a-92b5-42f4-ba83-0270d49ce026-tigera-ca-bundle\") pod \"calico-node-n8hkc\" (UID: \"6bb2af7a-92b5-42f4-ba83-0270d49ce026\") " pod="calico-system/calico-node-n8hkc" Dec 13 02:40:54.353655 kubelet[1830]: I1213 02:40:54.353645 1830 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/6bb2af7a-92b5-42f4-ba83-0270d49ce026-var-run-calico\") pod \"calico-node-n8hkc\" (UID: \"6bb2af7a-92b5-42f4-ba83-0270d49ce026\") " pod="calico-system/calico-node-n8hkc" Dec 13 02:40:54.353766 kubelet[1830]: I1213 02:40:54.353755 1830 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/6bb2af7a-92b5-42f4-ba83-0270d49ce026-var-lib-calico\") pod \"calico-node-n8hkc\" (UID: \"6bb2af7a-92b5-42f4-ba83-0270d49ce026\") " pod="calico-system/calico-node-n8hkc" Dec 13 02:40:54.353849 kubelet[1830]: I1213 02:40:54.353839 1830 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/6bb2af7a-92b5-42f4-ba83-0270d49ce026-cni-bin-dir\") pod \"calico-node-n8hkc\" (UID: \"6bb2af7a-92b5-42f4-ba83-0270d49ce026\") " pod="calico-system/calico-node-n8hkc" Dec 13 02:40:54.354008 kubelet[1830]: I1213 02:40:54.353997 1830 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/6bb2af7a-92b5-42f4-ba83-0270d49ce026-cni-log-dir\") pod \"calico-node-n8hkc\" (UID: \"6bb2af7a-92b5-42f4-ba83-0270d49ce026\") " pod="calico-system/calico-node-n8hkc" Dec 13 02:40:54.354099 kubelet[1830]: I1213 02:40:54.354089 1830 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/6bb2af7a-92b5-42f4-ba83-0270d49ce026-policysync\") pod \"calico-node-n8hkc\" (UID: \"6bb2af7a-92b5-42f4-ba83-0270d49ce026\") " pod="calico-system/calico-node-n8hkc" Dec 13 02:40:54.354366 kubelet[1830]: I1213 02:40:54.354209 1830 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/6bb2af7a-92b5-42f4-ba83-0270d49ce026-cni-net-dir\") pod \"calico-node-n8hkc\" (UID: \"6bb2af7a-92b5-42f4-ba83-0270d49ce026\") " pod="calico-system/calico-node-n8hkc" Dec 13 02:40:54.354366 kubelet[1830]: I1213 02:40:54.354238 1830 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ed687165-b432-4311-a017-bee89dd86d6d-socket-dir\") pod \"csi-node-driver-mn4q7\" (UID: \"ed687165-b432-4311-a017-bee89dd86d6d\") " pod="calico-system/csi-node-driver-mn4q7" Dec 13 02:40:54.354366 kubelet[1830]: I1213 02:40:54.354264 1830 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ed687165-b432-4311-a017-bee89dd86d6d-registration-dir\") pod \"csi-node-driver-mn4q7\" (UID: \"ed687165-b432-4311-a017-bee89dd86d6d\") " pod="calico-system/csi-node-driver-mn4q7" Dec 13 02:40:54.354366 kubelet[1830]: I1213 02:40:54.354290 1830 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6glk9\" (UniqueName: \"kubernetes.io/projected/99595cfe-3433-47c0-8c88-589f72f1e421-kube-api-access-6glk9\") pod \"kube-proxy-j7n7v\" (UID: \"99595cfe-3433-47c0-8c88-589f72f1e421\") " pod="kube-system/kube-proxy-j7n7v" Dec 13 02:40:54.463381 kubelet[1830]: E1213 02:40:54.463310 1830 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 13 02:40:54.466069 kubelet[1830]: W1213 02:40:54.463608 1830 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 13 02:40:54.466957 kubelet[1830]: E1213 02:40:54.466843 1830 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 13 02:40:54.467184 kubelet[1830]: E1213 02:40:54.466015 1830 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 13 02:40:54.467499 kubelet[1830]: W1213 02:40:54.467315 1830 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 13 02:40:54.467499 kubelet[1830]: E1213 02:40:54.467410 1830 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 13 02:40:54.469441 kubelet[1830]: E1213 02:40:54.469396 1830 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 13 02:40:54.469441 kubelet[1830]: W1213 02:40:54.469435 1830 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 13 02:40:54.469789 kubelet[1830]: E1213 02:40:54.469538 1830 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 13 02:40:54.470003 kubelet[1830]: E1213 02:40:54.469800 1830 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 13 02:40:54.470003 kubelet[1830]: W1213 02:40:54.469820 1830 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 13 02:40:54.470287 kubelet[1830]: E1213 02:40:54.470168 1830 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 13 02:40:54.470287 kubelet[1830]: E1213 02:40:54.470183 1830 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 13 02:40:54.470287 kubelet[1830]: W1213 02:40:54.470203 1830 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 13 02:40:54.472034 kubelet[1830]: E1213 02:40:54.470346 1830 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 13 02:40:54.472034 kubelet[1830]: E1213 02:40:54.470653 1830 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 13 02:40:54.472034 kubelet[1830]: W1213 02:40:54.470672 1830 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 13 02:40:54.472034 kubelet[1830]: E1213 02:40:54.470884 1830 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 13 02:40:54.472034 kubelet[1830]: E1213 02:40:54.471124 1830 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 13 02:40:54.472034 kubelet[1830]: W1213 02:40:54.471142 1830 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 13 02:40:54.472034 kubelet[1830]: E1213 02:40:54.471309 1830 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 13 02:40:54.472034 kubelet[1830]: E1213 02:40:54.471498 1830 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 13 02:40:54.472034 kubelet[1830]: W1213 02:40:54.471516 1830 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 13 02:40:54.472034 kubelet[1830]: E1213 02:40:54.471671 1830 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 13 02:40:54.477346 kubelet[1830]: E1213 02:40:54.471960 1830 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 13 02:40:54.477346 kubelet[1830]: W1213 02:40:54.471979 1830 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 13 02:40:54.477346 kubelet[1830]: E1213 02:40:54.472135 1830 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 13 02:40:54.477346 kubelet[1830]: E1213 02:40:54.472330 1830 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 13 02:40:54.477346 kubelet[1830]: W1213 02:40:54.472351 1830 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 13 02:40:54.477346 kubelet[1830]: E1213 02:40:54.472493 1830 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 13 02:40:54.477346 kubelet[1830]: E1213 02:40:54.472845 1830 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 13 02:40:54.477346 kubelet[1830]: W1213 02:40:54.472863 1830 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 13 02:40:54.477346 kubelet[1830]: E1213 02:40:54.473079 1830 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 13 02:40:54.477346 kubelet[1830]: E1213 02:40:54.473272 1830 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 13 02:40:54.477888 kubelet[1830]: W1213 02:40:54.473290 1830 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 13 02:40:54.477888 kubelet[1830]: E1213 02:40:54.473576 1830 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 13 02:40:54.477888 kubelet[1830]: W1213 02:40:54.473595 1830 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 13 02:40:54.477888 kubelet[1830]: E1213 02:40:54.473858 1830 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 13 02:40:54.477888 kubelet[1830]: W1213 02:40:54.473877 1830 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 13 02:40:54.477888 kubelet[1830]: E1213 02:40:54.474314 1830 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 13 02:40:54.477888 kubelet[1830]: W1213 02:40:54.474334 1830 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 13 02:40:54.477888 kubelet[1830]: E1213 02:40:54.474667 1830 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 13 02:40:54.477888 kubelet[1830]: W1213 02:40:54.474685 1830 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 13 02:40:54.477888 kubelet[1830]: E1213 02:40:54.475050 1830 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 13 02:40:54.478430 kubelet[1830]: W1213 02:40:54.475068 1830 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 13 02:40:54.478430 kubelet[1830]: E1213 02:40:54.475096 1830 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 13 02:40:54.478430 kubelet[1830]: E1213 02:40:54.476043 1830 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 13 02:40:54.478430 kubelet[1830]: W1213 02:40:54.476064 1830 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 13 02:40:54.478430 kubelet[1830]: E1213 02:40:54.476093 1830 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 13 02:40:54.478430 kubelet[1830]: E1213 02:40:54.476132 1830 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 13 02:40:54.478430 kubelet[1830]: E1213 02:40:54.476433 1830 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 13 02:40:54.478430 kubelet[1830]: W1213 02:40:54.476451 1830 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 13 02:40:54.478430 kubelet[1830]: E1213 02:40:54.476479 1830 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 13 02:40:54.478430 kubelet[1830]: E1213 02:40:54.476781 1830 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 13 02:40:54.480867 kubelet[1830]: W1213 02:40:54.476800 1830 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 13 02:40:54.480867 kubelet[1830]: E1213 02:40:54.476841 1830 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 13 02:40:54.480867 kubelet[1830]: E1213 02:40:54.476879 1830 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 13 02:40:54.480867 kubelet[1830]: E1213 02:40:54.477034 1830 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 13 02:40:54.480867 kubelet[1830]: E1213 02:40:54.477192 1830 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 13 02:40:54.480867 kubelet[1830]: E1213 02:40:54.477211 1830 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 13 02:40:54.480867 kubelet[1830]: W1213 02:40:54.477212 1830 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 13 02:40:54.480867 kubelet[1830]: E1213 02:40:54.477237 1830 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 13 02:40:54.480867 kubelet[1830]: E1213 02:40:54.477256 1830 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 13 02:40:54.480867 kubelet[1830]: E1213 02:40:54.477563 1830 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 13 02:40:54.482818 kubelet[1830]: W1213 02:40:54.477582 1830 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 13 02:40:54.482818 kubelet[1830]: E1213 02:40:54.477645 1830 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 13 02:40:54.482818 kubelet[1830]: E1213 02:40:54.478034 1830 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 13 02:40:54.482818 kubelet[1830]: W1213 02:40:54.478053 1830 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 13 02:40:54.482818 kubelet[1830]: E1213 02:40:54.478112 1830 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 13 02:40:54.482818 kubelet[1830]: E1213 02:40:54.478546 1830 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 13 02:40:54.482818 kubelet[1830]: W1213 02:40:54.478568 1830 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 13 02:40:54.482818 kubelet[1830]: E1213 02:40:54.478634 1830 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 13 02:40:54.482818 kubelet[1830]: E1213 02:40:54.479023 1830 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 13 02:40:54.482818 kubelet[1830]: W1213 02:40:54.479042 1830 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 13 02:40:54.483415 kubelet[1830]: E1213 02:40:54.479068 1830 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 13 02:40:54.483415 kubelet[1830]: E1213 02:40:54.481979 1830 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 13 02:40:54.483415 kubelet[1830]: W1213 02:40:54.482003 1830 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 13 02:40:54.483415 kubelet[1830]: E1213 02:40:54.482033 1830 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 13 02:40:54.557258 kubelet[1830]: E1213 02:40:54.557111 1830 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 13 02:40:54.557710 kubelet[1830]: W1213 02:40:54.557472 1830 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 13 02:40:54.557710 kubelet[1830]: E1213 02:40:54.557521 1830 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 13 02:40:54.558321 kubelet[1830]: E1213 02:40:54.558163 1830 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 13 02:40:54.558321 kubelet[1830]: W1213 02:40:54.558191 1830 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 13 02:40:54.558321 kubelet[1830]: E1213 02:40:54.558220 1830 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 13 02:40:54.558989 kubelet[1830]: E1213 02:40:54.558852 1830 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 13 02:40:54.558989 kubelet[1830]: W1213 02:40:54.558876 1830 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 13 02:40:54.558989 kubelet[1830]: E1213 02:40:54.558943 1830 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 13 02:40:54.650523 kubelet[1830]: E1213 02:40:54.648114 1830 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 13 02:40:54.650523 kubelet[1830]: W1213 02:40:54.648136 1830 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 13 02:40:54.650523 kubelet[1830]: E1213 02:40:54.648164 1830 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 13 02:40:54.653072 kubelet[1830]: E1213 02:40:54.653048 1830 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 13 02:40:54.653185 kubelet[1830]: W1213 02:40:54.653167 1830 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 13 02:40:54.653269 kubelet[1830]: E1213 02:40:54.653257 1830 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 13 02:40:54.658841 kubelet[1830]: E1213 02:40:54.658638 1830 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 13 02:40:54.658841 kubelet[1830]: W1213 02:40:54.658712 1830 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 13 02:40:54.658841 kubelet[1830]: E1213 02:40:54.658771 1830 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 13 02:40:54.932857 containerd[1466]: time="2024-12-13T02:40:54.932597054Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-j7n7v,Uid:99595cfe-3433-47c0-8c88-589f72f1e421,Namespace:kube-system,Attempt:0,}" Dec 13 02:40:54.948019 containerd[1466]: time="2024-12-13T02:40:54.947821413Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-n8hkc,Uid:6bb2af7a-92b5-42f4-ba83-0270d49ce026,Namespace:calico-system,Attempt:0,}" Dec 13 02:40:55.293503 kubelet[1830]: E1213 02:40:55.293394 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:40:55.612851 kubelet[1830]: E1213 02:40:55.612591 1830 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-mn4q7" podUID="ed687165-b432-4311-a017-bee89dd86d6d" Dec 13 02:40:55.709236 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3633033840.mount: Deactivated successfully. Dec 13 02:40:55.723858 containerd[1466]: time="2024-12-13T02:40:55.723637785Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Dec 13 02:40:55.725933 containerd[1466]: time="2024-12-13T02:40:55.725781831Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Dec 13 02:40:55.728306 containerd[1466]: time="2024-12-13T02:40:55.728217080Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312064" Dec 13 02:40:55.728458 containerd[1466]: time="2024-12-13T02:40:55.728312169Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Dec 13 02:40:55.730134 containerd[1466]: time="2024-12-13T02:40:55.730022487Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Dec 13 02:40:55.735273 containerd[1466]: time="2024-12-13T02:40:55.735206697Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Dec 13 02:40:55.737934 containerd[1466]: time="2024-12-13T02:40:55.737276085Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 789.212995ms" Dec 13 02:40:55.748092 containerd[1466]: time="2024-12-13T02:40:55.747970592Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 815.126432ms" Dec 13 02:40:56.088938 containerd[1466]: time="2024-12-13T02:40:56.088677786Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 02:40:56.088938 containerd[1466]: time="2024-12-13T02:40:56.088751074Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 02:40:56.088938 containerd[1466]: time="2024-12-13T02:40:56.088772818Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 02:40:56.089372 containerd[1466]: time="2024-12-13T02:40:56.088863533Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 02:40:56.090376 containerd[1466]: time="2024-12-13T02:40:56.090054488Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 02:40:56.090376 containerd[1466]: time="2024-12-13T02:40:56.090130490Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 02:40:56.090376 containerd[1466]: time="2024-12-13T02:40:56.090151554Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 02:40:56.090376 containerd[1466]: time="2024-12-13T02:40:56.090273669Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 02:40:56.271470 systemd[1]: Started cri-containerd-24813c3ebfa164c37d23c72368b37932d58214422a622d4b386bfe0684c58adb.scope - libcontainer container 24813c3ebfa164c37d23c72368b37932d58214422a622d4b386bfe0684c58adb. Dec 13 02:40:56.277186 systemd[1]: Started cri-containerd-e46f0c1f0a29994af5782e2279cfd4f05f76ed72d0e3a686adb5696b628f8b59.scope - libcontainer container e46f0c1f0a29994af5782e2279cfd4f05f76ed72d0e3a686adb5696b628f8b59. Dec 13 02:40:56.294608 kubelet[1830]: E1213 02:40:56.294530 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:40:56.319249 containerd[1466]: time="2024-12-13T02:40:56.319136844Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-n8hkc,Uid:6bb2af7a-92b5-42f4-ba83-0270d49ce026,Namespace:calico-system,Attempt:0,} returns sandbox id \"24813c3ebfa164c37d23c72368b37932d58214422a622d4b386bfe0684c58adb\"" Dec 13 02:40:56.322743 containerd[1466]: time="2024-12-13T02:40:56.322614086Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\"" Dec 13 02:40:56.329535 containerd[1466]: time="2024-12-13T02:40:56.329372428Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-j7n7v,Uid:99595cfe-3433-47c0-8c88-589f72f1e421,Namespace:kube-system,Attempt:0,} returns sandbox id \"e46f0c1f0a29994af5782e2279cfd4f05f76ed72d0e3a686adb5696b628f8b59\"" Dec 13 02:40:57.295778 kubelet[1830]: E1213 02:40:57.295668 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:40:57.614340 kubelet[1830]: E1213 02:40:57.613439 1830 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-mn4q7" podUID="ed687165-b432-4311-a017-bee89dd86d6d" Dec 13 02:40:58.296180 kubelet[1830]: E1213 02:40:58.296070 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:40:58.745569 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3038492072.mount: Deactivated successfully. Dec 13 02:40:59.016727 containerd[1466]: time="2024-12-13T02:40:59.016551157Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 02:40:59.017878 containerd[1466]: time="2024-12-13T02:40:59.017815000Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1: active requests=0, bytes read=6855343" Dec 13 02:40:59.019078 containerd[1466]: time="2024-12-13T02:40:59.018989189Z" level=info msg="ImageCreate event name:\"sha256:2b7452b763ec8833ca0386ada5fd066e552a9b3b02b8538a5e34cc3d6d3840a6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 02:40:59.021574 containerd[1466]: time="2024-12-13T02:40:59.021517486Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:a63f8b4ff531912d12d143664eb263fdbc6cd7b3ff4aa777dfb6e318a090462c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 02:40:59.022496 containerd[1466]: time="2024-12-13T02:40:59.022240039Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\" with image id \"sha256:2b7452b763ec8833ca0386ada5fd066e552a9b3b02b8538a5e34cc3d6d3840a6\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:a63f8b4ff531912d12d143664eb263fdbc6cd7b3ff4aa777dfb6e318a090462c\", size \"6855165\" in 2.699580367s" Dec 13 02:40:59.022496 containerd[1466]: time="2024-12-13T02:40:59.022300409Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\" returns image reference \"sha256:2b7452b763ec8833ca0386ada5fd066e552a9b3b02b8538a5e34cc3d6d3840a6\"" Dec 13 02:40:59.023873 containerd[1466]: time="2024-12-13T02:40:59.023825330Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.29.12\"" Dec 13 02:40:59.025361 containerd[1466]: time="2024-12-13T02:40:59.025196400Z" level=info msg="CreateContainer within sandbox \"24813c3ebfa164c37d23c72368b37932d58214422a622d4b386bfe0684c58adb\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Dec 13 02:40:59.055085 containerd[1466]: time="2024-12-13T02:40:59.055014223Z" level=info msg="CreateContainer within sandbox \"24813c3ebfa164c37d23c72368b37932d58214422a622d4b386bfe0684c58adb\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"50991e08c818c74bbd44bd704be25d60f5e0b0b141d4ab019b02bce9b3c75ef0\"" Dec 13 02:40:59.056262 containerd[1466]: time="2024-12-13T02:40:59.056215091Z" level=info msg="StartContainer for \"50991e08c818c74bbd44bd704be25d60f5e0b0b141d4ab019b02bce9b3c75ef0\"" Dec 13 02:40:59.103108 systemd[1]: Started cri-containerd-50991e08c818c74bbd44bd704be25d60f5e0b0b141d4ab019b02bce9b3c75ef0.scope - libcontainer container 50991e08c818c74bbd44bd704be25d60f5e0b0b141d4ab019b02bce9b3c75ef0. Dec 13 02:40:59.143173 containerd[1466]: time="2024-12-13T02:40:59.143080866Z" level=info msg="StartContainer for \"50991e08c818c74bbd44bd704be25d60f5e0b0b141d4ab019b02bce9b3c75ef0\" returns successfully" Dec 13 02:40:59.152572 systemd[1]: cri-containerd-50991e08c818c74bbd44bd704be25d60f5e0b0b141d4ab019b02bce9b3c75ef0.scope: Deactivated successfully. Dec 13 02:40:59.296490 kubelet[1830]: E1213 02:40:59.296295 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:40:59.381144 containerd[1466]: time="2024-12-13T02:40:59.380999874Z" level=info msg="shim disconnected" id=50991e08c818c74bbd44bd704be25d60f5e0b0b141d4ab019b02bce9b3c75ef0 namespace=k8s.io Dec 13 02:40:59.381973 containerd[1466]: time="2024-12-13T02:40:59.381535956Z" level=warning msg="cleaning up after shim disconnected" id=50991e08c818c74bbd44bd704be25d60f5e0b0b141d4ab019b02bce9b3c75ef0 namespace=k8s.io Dec 13 02:40:59.381973 containerd[1466]: time="2024-12-13T02:40:59.381591166Z" level=info msg="cleaning up dead shim" namespace=k8s.io Dec 13 02:40:59.613625 kubelet[1830]: E1213 02:40:59.612999 1830 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-mn4q7" podUID="ed687165-b432-4311-a017-bee89dd86d6d" Dec 13 02:40:59.663044 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-50991e08c818c74bbd44bd704be25d60f5e0b0b141d4ab019b02bce9b3c75ef0-rootfs.mount: Deactivated successfully. Dec 13 02:41:00.298002 kubelet[1830]: E1213 02:41:00.297837 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:00.933975 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2817168286.mount: Deactivated successfully. Dec 13 02:41:01.298435 kubelet[1830]: E1213 02:41:01.298401 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:01.575395 containerd[1466]: time="2024-12-13T02:41:01.575145993Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.29.12\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 02:41:01.577186 containerd[1466]: time="2024-12-13T02:41:01.576917734Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.29.12: active requests=0, bytes read=28619966" Dec 13 02:41:01.578604 containerd[1466]: time="2024-12-13T02:41:01.578577240Z" level=info msg="ImageCreate event name:\"sha256:d699d5830022f9e67c3271d1c2af58eaede81e3567df82728b7d2a8bf12ed153\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 02:41:01.581986 containerd[1466]: time="2024-12-13T02:41:01.581926856Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:bc761494b78fa152a759457f42bc9b86ee9d18f5929bb127bd5f72f8e2112c39\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 02:41:01.583636 containerd[1466]: time="2024-12-13T02:41:01.583585550Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.29.12\" with image id \"sha256:d699d5830022f9e67c3271d1c2af58eaede81e3567df82728b7d2a8bf12ed153\", repo tag \"registry.k8s.io/kube-proxy:v1.29.12\", repo digest \"registry.k8s.io/kube-proxy@sha256:bc761494b78fa152a759457f42bc9b86ee9d18f5929bb127bd5f72f8e2112c39\", size \"28618977\" in 2.559708606s" Dec 13 02:41:01.583811 containerd[1466]: time="2024-12-13T02:41:01.583712073Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.29.12\" returns image reference \"sha256:d699d5830022f9e67c3271d1c2af58eaede81e3567df82728b7d2a8bf12ed153\"" Dec 13 02:41:01.586259 containerd[1466]: time="2024-12-13T02:41:01.586195921Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.1\"" Dec 13 02:41:01.588375 containerd[1466]: time="2024-12-13T02:41:01.588345786Z" level=info msg="CreateContainer within sandbox \"e46f0c1f0a29994af5782e2279cfd4f05f76ed72d0e3a686adb5696b628f8b59\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Dec 13 02:41:01.613210 kubelet[1830]: E1213 02:41:01.613123 1830 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-mn4q7" podUID="ed687165-b432-4311-a017-bee89dd86d6d" Dec 13 02:41:01.649284 containerd[1466]: time="2024-12-13T02:41:01.649150429Z" level=info msg="CreateContainer within sandbox \"e46f0c1f0a29994af5782e2279cfd4f05f76ed72d0e3a686adb5696b628f8b59\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"82f7ffc063606e2c231e8941d5bd1053a5193b28d11c21704304a1fc563f1c78\"" Dec 13 02:41:01.652030 containerd[1466]: time="2024-12-13T02:41:01.650243906Z" level=info msg="StartContainer for \"82f7ffc063606e2c231e8941d5bd1053a5193b28d11c21704304a1fc563f1c78\"" Dec 13 02:41:01.699094 systemd[1]: Started cri-containerd-82f7ffc063606e2c231e8941d5bd1053a5193b28d11c21704304a1fc563f1c78.scope - libcontainer container 82f7ffc063606e2c231e8941d5bd1053a5193b28d11c21704304a1fc563f1c78. Dec 13 02:41:01.744048 containerd[1466]: time="2024-12-13T02:41:01.743993186Z" level=info msg="StartContainer for \"82f7ffc063606e2c231e8941d5bd1053a5193b28d11c21704304a1fc563f1c78\" returns successfully" Dec 13 02:41:02.299710 kubelet[1830]: E1213 02:41:02.299615 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:02.688110 kubelet[1830]: I1213 02:41:02.687772 1830 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-proxy-j7n7v" podStartSLOduration=5.433871354 podStartE2EDuration="10.687627606s" podCreationTimestamp="2024-12-13 02:40:52 +0000 UTC" firstStartedPulling="2024-12-13 02:40:56.330635599 +0000 UTC m=+5.147671380" lastFinishedPulling="2024-12-13 02:41:01.58439184 +0000 UTC m=+10.401427632" observedRunningTime="2024-12-13 02:41:02.687311665 +0000 UTC m=+11.504347496" watchObservedRunningTime="2024-12-13 02:41:02.687627606 +0000 UTC m=+11.504663437" Dec 13 02:41:03.300648 kubelet[1830]: E1213 02:41:03.300555 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:03.614138 kubelet[1830]: E1213 02:41:03.613056 1830 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-mn4q7" podUID="ed687165-b432-4311-a017-bee89dd86d6d" Dec 13 02:41:04.301678 kubelet[1830]: E1213 02:41:04.301598 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:05.302389 kubelet[1830]: E1213 02:41:05.302264 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:05.613732 kubelet[1830]: E1213 02:41:05.613507 1830 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-mn4q7" podUID="ed687165-b432-4311-a017-bee89dd86d6d" Dec 13 02:41:06.303493 kubelet[1830]: E1213 02:41:06.303418 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:07.304335 kubelet[1830]: E1213 02:41:07.304025 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:07.613013 kubelet[1830]: E1213 02:41:07.612819 1830 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-mn4q7" podUID="ed687165-b432-4311-a017-bee89dd86d6d" Dec 13 02:41:08.305297 kubelet[1830]: E1213 02:41:08.305225 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:08.957704 containerd[1466]: time="2024-12-13T02:41:08.957571049Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 02:41:08.961283 containerd[1466]: time="2024-12-13T02:41:08.961151505Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.29.1: active requests=0, bytes read=96154154" Dec 13 02:41:08.967511 containerd[1466]: time="2024-12-13T02:41:08.967318642Z" level=info msg="ImageCreate event name:\"sha256:7dd6ea186aba0d7a1791a79d426fe854527ca95192b26bbd19e8baf8373f7d0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 02:41:08.978247 containerd[1466]: time="2024-12-13T02:41:08.978009569Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:21e759d51c90dfb34fc1397dc180dd3a3fb564c2b0580d2f61ffe108f2a3c94b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 02:41:08.981347 containerd[1466]: time="2024-12-13T02:41:08.979792884Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.29.1\" with image id \"sha256:7dd6ea186aba0d7a1791a79d426fe854527ca95192b26bbd19e8baf8373f7d0e\", repo tag \"ghcr.io/flatcar/calico/cni:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:21e759d51c90dfb34fc1397dc180dd3a3fb564c2b0580d2f61ffe108f2a3c94b\", size \"97647238\" in 7.393509583s" Dec 13 02:41:08.981347 containerd[1466]: time="2024-12-13T02:41:08.979886326Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.1\" returns image reference \"sha256:7dd6ea186aba0d7a1791a79d426fe854527ca95192b26bbd19e8baf8373f7d0e\"" Dec 13 02:41:08.985236 containerd[1466]: time="2024-12-13T02:41:08.985131771Z" level=info msg="CreateContainer within sandbox \"24813c3ebfa164c37d23c72368b37932d58214422a622d4b386bfe0684c58adb\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Dec 13 02:41:09.307023 kubelet[1830]: E1213 02:41:09.306851 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:09.378396 containerd[1466]: time="2024-12-13T02:41:09.378267892Z" level=info msg="CreateContainer within sandbox \"24813c3ebfa164c37d23c72368b37932d58214422a622d4b386bfe0684c58adb\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"601ff733b7d1bd3884274667cd6c3c742c1c2fcc347eb125f2155cf4c9ba15df\"" Dec 13 02:41:09.379221 containerd[1466]: time="2024-12-13T02:41:09.379070256Z" level=info msg="StartContainer for \"601ff733b7d1bd3884274667cd6c3c742c1c2fcc347eb125f2155cf4c9ba15df\"" Dec 13 02:41:09.444269 systemd[1]: Started cri-containerd-601ff733b7d1bd3884274667cd6c3c742c1c2fcc347eb125f2155cf4c9ba15df.scope - libcontainer container 601ff733b7d1bd3884274667cd6c3c742c1c2fcc347eb125f2155cf4c9ba15df. Dec 13 02:41:09.612871 kubelet[1830]: E1213 02:41:09.612639 1830 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-mn4q7" podUID="ed687165-b432-4311-a017-bee89dd86d6d" Dec 13 02:41:09.889815 containerd[1466]: time="2024-12-13T02:41:09.889436322Z" level=info msg="StartContainer for \"601ff733b7d1bd3884274667cd6c3c742c1c2fcc347eb125f2155cf4c9ba15df\" returns successfully" Dec 13 02:41:10.307752 kubelet[1830]: E1213 02:41:10.307656 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:11.308534 kubelet[1830]: E1213 02:41:11.308441 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:11.582598 update_engine[1445]: I20241213 02:41:11.581053 1445 update_attempter.cc:509] Updating boot flags... Dec 13 02:41:11.615000 kubelet[1830]: E1213 02:41:11.612811 1830 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-mn4q7" podUID="ed687165-b432-4311-a017-bee89dd86d6d" Dec 13 02:41:11.688394 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 41 scanned by (udev-worker) (2281) Dec 13 02:41:11.990987 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 41 scanned by (udev-worker) (2279) Dec 13 02:41:12.173986 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 41 scanned by (udev-worker) (2279) Dec 13 02:41:12.207079 systemd[1]: cri-containerd-601ff733b7d1bd3884274667cd6c3c742c1c2fcc347eb125f2155cf4c9ba15df.scope: Deactivated successfully. Dec 13 02:41:12.254217 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-601ff733b7d1bd3884274667cd6c3c742c1c2fcc347eb125f2155cf4c9ba15df-rootfs.mount: Deactivated successfully. Dec 13 02:41:12.259198 kubelet[1830]: I1213 02:41:12.257132 1830 kubelet_node_status.go:497] "Fast updating node status as it just became ready" Dec 13 02:41:12.290142 kubelet[1830]: E1213 02:41:12.290083 1830 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:12.631207 kubelet[1830]: E1213 02:41:12.308602 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:13.281668 containerd[1466]: time="2024-12-13T02:41:13.280304833Z" level=info msg="shim disconnected" id=601ff733b7d1bd3884274667cd6c3c742c1c2fcc347eb125f2155cf4c9ba15df namespace=k8s.io Dec 13 02:41:13.281668 containerd[1466]: time="2024-12-13T02:41:13.280537975Z" level=warning msg="cleaning up after shim disconnected" id=601ff733b7d1bd3884274667cd6c3c742c1c2fcc347eb125f2155cf4c9ba15df namespace=k8s.io Dec 13 02:41:13.281668 containerd[1466]: time="2024-12-13T02:41:13.280583580Z" level=info msg="cleaning up dead shim" namespace=k8s.io Dec 13 02:41:13.310874 kubelet[1830]: E1213 02:41:13.309675 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:13.627133 systemd[1]: Created slice kubepods-besteffort-poded687165_b432_4311_a017_bee89dd86d6d.slice - libcontainer container kubepods-besteffort-poded687165_b432_4311_a017_bee89dd86d6d.slice. Dec 13 02:41:13.634088 containerd[1466]: time="2024-12-13T02:41:13.633945249Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-mn4q7,Uid:ed687165-b432-4311-a017-bee89dd86d6d,Namespace:calico-system,Attempt:0,}" Dec 13 02:41:13.846589 containerd[1466]: time="2024-12-13T02:41:13.846528147Z" level=error msg="Failed to destroy network for sandbox \"7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 13 02:41:13.849803 containerd[1466]: time="2024-12-13T02:41:13.847233445Z" level=error msg="encountered an error cleaning up failed sandbox \"7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 13 02:41:13.849803 containerd[1466]: time="2024-12-13T02:41:13.849026510Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-mn4q7,Uid:ed687165-b432-4311-a017-bee89dd86d6d,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 13 02:41:13.850026 kubelet[1830]: E1213 02:41:13.849336 1830 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 13 02:41:13.850026 kubelet[1830]: E1213 02:41:13.849408 1830 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-mn4q7" Dec 13 02:41:13.850026 kubelet[1830]: E1213 02:41:13.849434 1830 kuberuntime_manager.go:1172] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-mn4q7" Dec 13 02:41:13.850359 kubelet[1830]: E1213 02:41:13.849500 1830 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-mn4q7_calico-system(ed687165-b432-4311-a017-bee89dd86d6d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-mn4q7_calico-system(ed687165-b432-4311-a017-bee89dd86d6d)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-mn4q7" podUID="ed687165-b432-4311-a017-bee89dd86d6d" Dec 13 02:41:13.850810 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c-shm.mount: Deactivated successfully. Dec 13 02:41:13.916721 containerd[1466]: time="2024-12-13T02:41:13.916145680Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.1\"" Dec 13 02:41:13.921957 kubelet[1830]: I1213 02:41:13.920633 1830 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c" Dec 13 02:41:13.923358 containerd[1466]: time="2024-12-13T02:41:13.922268913Z" level=info msg="StopPodSandbox for \"7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c\"" Dec 13 02:41:13.923358 containerd[1466]: time="2024-12-13T02:41:13.922669396Z" level=info msg="Ensure that sandbox 7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c in task-service has been cleanup successfully" Dec 13 02:41:13.996525 containerd[1466]: time="2024-12-13T02:41:13.996340836Z" level=error msg="StopPodSandbox for \"7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c\" failed" error="failed to destroy network for sandbox \"7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 13 02:41:13.997668 kubelet[1830]: E1213 02:41:13.997582 1830 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c" Dec 13 02:41:13.997864 kubelet[1830]: E1213 02:41:13.997695 1830 kuberuntime_manager.go:1381] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c"} Dec 13 02:41:13.997864 kubelet[1830]: E1213 02:41:13.997747 1830 kuberuntime_manager.go:1081] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"ed687165-b432-4311-a017-bee89dd86d6d\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Dec 13 02:41:13.997864 kubelet[1830]: E1213 02:41:13.997787 1830 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"ed687165-b432-4311-a017-bee89dd86d6d\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-mn4q7" podUID="ed687165-b432-4311-a017-bee89dd86d6d" Dec 13 02:41:14.310951 kubelet[1830]: E1213 02:41:14.310718 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:15.166965 kubelet[1830]: I1213 02:41:15.165741 1830 topology_manager.go:215] "Topology Admit Handler" podUID="1c2ae6bf-c9e7-4d97-8deb-2cd8f7181aa0" podNamespace="default" podName="nginx-deployment-6d5f899847-lblsv" Dec 13 02:41:15.181202 systemd[1]: Created slice kubepods-besteffort-pod1c2ae6bf_c9e7_4d97_8deb_2cd8f7181aa0.slice - libcontainer container kubepods-besteffort-pod1c2ae6bf_c9e7_4d97_8deb_2cd8f7181aa0.slice. Dec 13 02:41:15.311229 kubelet[1830]: E1213 02:41:15.311121 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:15.316666 kubelet[1830]: I1213 02:41:15.316558 1830 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6z9n\" (UniqueName: \"kubernetes.io/projected/1c2ae6bf-c9e7-4d97-8deb-2cd8f7181aa0-kube-api-access-c6z9n\") pod \"nginx-deployment-6d5f899847-lblsv\" (UID: \"1c2ae6bf-c9e7-4d97-8deb-2cd8f7181aa0\") " pod="default/nginx-deployment-6d5f899847-lblsv" Dec 13 02:41:15.489258 containerd[1466]: time="2024-12-13T02:41:15.489176890Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-6d5f899847-lblsv,Uid:1c2ae6bf-c9e7-4d97-8deb-2cd8f7181aa0,Namespace:default,Attempt:0,}" Dec 13 02:41:16.135100 containerd[1466]: time="2024-12-13T02:41:16.134980373Z" level=error msg="Failed to destroy network for sandbox \"3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 13 02:41:16.137246 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19-shm.mount: Deactivated successfully. Dec 13 02:41:16.137544 containerd[1466]: time="2024-12-13T02:41:16.137468564Z" level=error msg="encountered an error cleaning up failed sandbox \"3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 13 02:41:16.137687 containerd[1466]: time="2024-12-13T02:41:16.137625786Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-6d5f899847-lblsv,Uid:1c2ae6bf-c9e7-4d97-8deb-2cd8f7181aa0,Namespace:default,Attempt:0,} failed, error" error="failed to setup network for sandbox \"3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 13 02:41:16.138621 kubelet[1830]: E1213 02:41:16.138098 1830 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 13 02:41:16.138621 kubelet[1830]: E1213 02:41:16.138182 1830 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="default/nginx-deployment-6d5f899847-lblsv" Dec 13 02:41:16.138621 kubelet[1830]: E1213 02:41:16.138214 1830 kuberuntime_manager.go:1172] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="default/nginx-deployment-6d5f899847-lblsv" Dec 13 02:41:16.138777 kubelet[1830]: E1213 02:41:16.138293 1830 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"nginx-deployment-6d5f899847-lblsv_default(1c2ae6bf-c9e7-4d97-8deb-2cd8f7181aa0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"nginx-deployment-6d5f899847-lblsv_default(1c2ae6bf-c9e7-4d97-8deb-2cd8f7181aa0)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="default/nginx-deployment-6d5f899847-lblsv" podUID="1c2ae6bf-c9e7-4d97-8deb-2cd8f7181aa0" Dec 13 02:41:16.312545 kubelet[1830]: E1213 02:41:16.312474 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:16.933625 kubelet[1830]: I1213 02:41:16.932418 1830 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19" Dec 13 02:41:16.935963 containerd[1466]: time="2024-12-13T02:41:16.934309369Z" level=info msg="StopPodSandbox for \"3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19\"" Dec 13 02:41:16.935963 containerd[1466]: time="2024-12-13T02:41:16.934753904Z" level=info msg="Ensure that sandbox 3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19 in task-service has been cleanup successfully" Dec 13 02:41:17.018879 containerd[1466]: time="2024-12-13T02:41:17.018823058Z" level=error msg="StopPodSandbox for \"3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19\" failed" error="failed to destroy network for sandbox \"3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 13 02:41:17.019517 kubelet[1830]: E1213 02:41:17.019315 1830 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19" Dec 13 02:41:17.019517 kubelet[1830]: E1213 02:41:17.019377 1830 kuberuntime_manager.go:1381] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19"} Dec 13 02:41:17.019517 kubelet[1830]: E1213 02:41:17.019427 1830 kuberuntime_manager.go:1081] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"1c2ae6bf-c9e7-4d97-8deb-2cd8f7181aa0\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Dec 13 02:41:17.019517 kubelet[1830]: E1213 02:41:17.019471 1830 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"1c2ae6bf-c9e7-4d97-8deb-2cd8f7181aa0\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="default/nginx-deployment-6d5f899847-lblsv" podUID="1c2ae6bf-c9e7-4d97-8deb-2cd8f7181aa0" Dec 13 02:41:17.313300 kubelet[1830]: E1213 02:41:17.313242 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:18.315408 kubelet[1830]: E1213 02:41:18.315241 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:19.316967 kubelet[1830]: E1213 02:41:19.316665 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:20.317558 kubelet[1830]: E1213 02:41:20.317520 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:21.319158 kubelet[1830]: E1213 02:41:21.319101 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:22.320198 kubelet[1830]: E1213 02:41:22.320112 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:23.320748 kubelet[1830]: E1213 02:41:23.320698 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:24.059266 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1277923874.mount: Deactivated successfully. Dec 13 02:41:24.322395 kubelet[1830]: E1213 02:41:24.321879 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:24.740961 containerd[1466]: time="2024-12-13T02:41:24.739704464Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 02:41:24.742032 containerd[1466]: time="2024-12-13T02:41:24.741809404Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.29.1: active requests=0, bytes read=142742010" Dec 13 02:41:24.744060 containerd[1466]: time="2024-12-13T02:41:24.743837501Z" level=info msg="ImageCreate event name:\"sha256:feb26d4585d68e875d9bd9bd6c27ea9f2d5c9ed9ef70f8b8cb0ebb0559a1d664\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 02:41:24.752393 containerd[1466]: time="2024-12-13T02:41:24.750820035Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:99c3917516efe1f807a0cfdf2d14b628b7c5cc6bd8a9ee5a253154f31756bea1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 02:41:24.753020 containerd[1466]: time="2024-12-13T02:41:24.752944993Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.29.1\" with image id \"sha256:feb26d4585d68e875d9bd9bd6c27ea9f2d5c9ed9ef70f8b8cb0ebb0559a1d664\", repo tag \"ghcr.io/flatcar/calico/node:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/node@sha256:99c3917516efe1f807a0cfdf2d14b628b7c5cc6bd8a9ee5a253154f31756bea1\", size \"142741872\" in 10.836669883s" Dec 13 02:41:24.753220 containerd[1466]: time="2024-12-13T02:41:24.753179900Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.1\" returns image reference \"sha256:feb26d4585d68e875d9bd9bd6c27ea9f2d5c9ed9ef70f8b8cb0ebb0559a1d664\"" Dec 13 02:41:24.836105 containerd[1466]: time="2024-12-13T02:41:24.836015418Z" level=info msg="CreateContainer within sandbox \"24813c3ebfa164c37d23c72368b37932d58214422a622d4b386bfe0684c58adb\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Dec 13 02:41:24.930114 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1814069087.mount: Deactivated successfully. Dec 13 02:41:25.322738 kubelet[1830]: E1213 02:41:25.322565 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:25.484543 containerd[1466]: time="2024-12-13T02:41:25.484284919Z" level=info msg="CreateContainer within sandbox \"24813c3ebfa164c37d23c72368b37932d58214422a622d4b386bfe0684c58adb\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"760b3de3e986604fb19586c233138c3861d592556bc918f52971cb12dbbabab1\"" Dec 13 02:41:25.486293 containerd[1466]: time="2024-12-13T02:41:25.486236625Z" level=info msg="StartContainer for \"760b3de3e986604fb19586c233138c3861d592556bc918f52971cb12dbbabab1\"" Dec 13 02:41:25.715117 systemd[1]: run-containerd-runc-k8s.io-760b3de3e986604fb19586c233138c3861d592556bc918f52971cb12dbbabab1-runc.RT1Uo4.mount: Deactivated successfully. Dec 13 02:41:25.730050 systemd[1]: Started cri-containerd-760b3de3e986604fb19586c233138c3861d592556bc918f52971cb12dbbabab1.scope - libcontainer container 760b3de3e986604fb19586c233138c3861d592556bc918f52971cb12dbbabab1. Dec 13 02:41:25.819923 containerd[1466]: time="2024-12-13T02:41:25.819477459Z" level=info msg="StartContainer for \"760b3de3e986604fb19586c233138c3861d592556bc918f52971cb12dbbabab1\" returns successfully" Dec 13 02:41:25.873030 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Dec 13 02:41:25.873128 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Dec 13 02:41:26.323396 kubelet[1830]: E1213 02:41:26.323341 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:26.614850 containerd[1466]: time="2024-12-13T02:41:26.614658033Z" level=info msg="StopPodSandbox for \"7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c\"" Dec 13 02:41:26.771869 kubelet[1830]: I1213 02:41:26.771820 1830 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="calico-system/calico-node-n8hkc" podStartSLOduration=6.34002659 podStartE2EDuration="34.771736548s" podCreationTimestamp="2024-12-13 02:40:52 +0000 UTC" firstStartedPulling="2024-12-13 02:40:56.321939247 +0000 UTC m=+5.138975028" lastFinishedPulling="2024-12-13 02:41:24.753649154 +0000 UTC m=+33.570684986" observedRunningTime="2024-12-13 02:41:25.998839101 +0000 UTC m=+34.815874923" watchObservedRunningTime="2024-12-13 02:41:26.771736548 +0000 UTC m=+35.588772379" Dec 13 02:41:26.853797 containerd[1466]: 2024-12-13 02:41:26.770 [INFO][2544] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c" Dec 13 02:41:26.853797 containerd[1466]: 2024-12-13 02:41:26.770 [INFO][2544] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c" iface="eth0" netns="/var/run/netns/cni-4ce3cf32-cd70-e424-fc26-9dffa38f2829" Dec 13 02:41:26.853797 containerd[1466]: 2024-12-13 02:41:26.770 [INFO][2544] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c" iface="eth0" netns="/var/run/netns/cni-4ce3cf32-cd70-e424-fc26-9dffa38f2829" Dec 13 02:41:26.853797 containerd[1466]: 2024-12-13 02:41:26.775 [INFO][2544] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c" iface="eth0" netns="/var/run/netns/cni-4ce3cf32-cd70-e424-fc26-9dffa38f2829" Dec 13 02:41:26.853797 containerd[1466]: 2024-12-13 02:41:26.775 [INFO][2544] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c" Dec 13 02:41:26.853797 containerd[1466]: 2024-12-13 02:41:26.775 [INFO][2544] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c" Dec 13 02:41:26.853797 containerd[1466]: 2024-12-13 02:41:26.822 [INFO][2550] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c" HandleID="k8s-pod-network.7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c" Workload="172.24.4.241-k8s-csi--node--driver--mn4q7-eth0" Dec 13 02:41:26.853797 containerd[1466]: 2024-12-13 02:41:26.822 [INFO][2550] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Dec 13 02:41:26.853797 containerd[1466]: 2024-12-13 02:41:26.822 [INFO][2550] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Dec 13 02:41:26.853797 containerd[1466]: 2024-12-13 02:41:26.839 [WARNING][2550] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c" HandleID="k8s-pod-network.7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c" Workload="172.24.4.241-k8s-csi--node--driver--mn4q7-eth0" Dec 13 02:41:26.853797 containerd[1466]: 2024-12-13 02:41:26.840 [INFO][2550] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c" HandleID="k8s-pod-network.7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c" Workload="172.24.4.241-k8s-csi--node--driver--mn4q7-eth0" Dec 13 02:41:26.853797 containerd[1466]: 2024-12-13 02:41:26.846 [INFO][2550] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Dec 13 02:41:26.853797 containerd[1466]: 2024-12-13 02:41:26.851 [INFO][2544] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c" Dec 13 02:41:26.858316 containerd[1466]: time="2024-12-13T02:41:26.856122436Z" level=info msg="TearDown network for sandbox \"7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c\" successfully" Dec 13 02:41:26.858316 containerd[1466]: time="2024-12-13T02:41:26.856189981Z" level=info msg="StopPodSandbox for \"7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c\" returns successfully" Dec 13 02:41:26.859852 systemd[1]: run-netns-cni\x2d4ce3cf32\x2dcd70\x2de424\x2dfc26\x2d9dffa38f2829.mount: Deactivated successfully. Dec 13 02:41:26.860326 containerd[1466]: time="2024-12-13T02:41:26.858880524Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-mn4q7,Uid:ed687165-b432-4311-a017-bee89dd86d6d,Namespace:calico-system,Attempt:1,}" Dec 13 02:41:27.138429 systemd-networkd[1371]: cali2ea430d9eb4: Link UP Dec 13 02:41:27.145246 systemd-networkd[1371]: cali2ea430d9eb4: Gained carrier Dec 13 02:41:27.162485 containerd[1466]: 2024-12-13 02:41:26.942 [INFO][2557] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Dec 13 02:41:27.162485 containerd[1466]: 2024-12-13 02:41:26.967 [INFO][2557] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {172.24.4.241-k8s-csi--node--driver--mn4q7-eth0 csi-node-driver- calico-system ed687165-b432-4311-a017-bee89dd86d6d 1122 0 2024-12-13 02:40:52 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:55b695c467 k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s 172.24.4.241 csi-node-driver-mn4q7 eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] cali2ea430d9eb4 [] []}} ContainerID="54e65b34c2462000e6bc32fc82fa534166e6fd83d370edd3864a4686e91b65c8" Namespace="calico-system" Pod="csi-node-driver-mn4q7" WorkloadEndpoint="172.24.4.241-k8s-csi--node--driver--mn4q7-" Dec 13 02:41:27.162485 containerd[1466]: 2024-12-13 02:41:26.967 [INFO][2557] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="54e65b34c2462000e6bc32fc82fa534166e6fd83d370edd3864a4686e91b65c8" Namespace="calico-system" Pod="csi-node-driver-mn4q7" WorkloadEndpoint="172.24.4.241-k8s-csi--node--driver--mn4q7-eth0" Dec 13 02:41:27.162485 containerd[1466]: 2024-12-13 02:41:27.043 [INFO][2577] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="54e65b34c2462000e6bc32fc82fa534166e6fd83d370edd3864a4686e91b65c8" HandleID="k8s-pod-network.54e65b34c2462000e6bc32fc82fa534166e6fd83d370edd3864a4686e91b65c8" Workload="172.24.4.241-k8s-csi--node--driver--mn4q7-eth0" Dec 13 02:41:27.162485 containerd[1466]: 2024-12-13 02:41:27.065 [INFO][2577] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="54e65b34c2462000e6bc32fc82fa534166e6fd83d370edd3864a4686e91b65c8" HandleID="k8s-pod-network.54e65b34c2462000e6bc32fc82fa534166e6fd83d370edd3864a4686e91b65c8" Workload="172.24.4.241-k8s-csi--node--driver--mn4q7-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0003a7990), Attrs:map[string]string{"namespace":"calico-system", "node":"172.24.4.241", "pod":"csi-node-driver-mn4q7", "timestamp":"2024-12-13 02:41:27.042999586 +0000 UTC"}, Hostname:"172.24.4.241", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Dec 13 02:41:27.162485 containerd[1466]: 2024-12-13 02:41:27.066 [INFO][2577] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Dec 13 02:41:27.162485 containerd[1466]: 2024-12-13 02:41:27.066 [INFO][2577] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Dec 13 02:41:27.162485 containerd[1466]: 2024-12-13 02:41:27.066 [INFO][2577] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host '172.24.4.241' Dec 13 02:41:27.162485 containerd[1466]: 2024-12-13 02:41:27.069 [INFO][2577] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.54e65b34c2462000e6bc32fc82fa534166e6fd83d370edd3864a4686e91b65c8" host="172.24.4.241" Dec 13 02:41:27.162485 containerd[1466]: 2024-12-13 02:41:27.078 [INFO][2577] ipam/ipam.go 372: Looking up existing affinities for host host="172.24.4.241" Dec 13 02:41:27.162485 containerd[1466]: 2024-12-13 02:41:27.088 [INFO][2577] ipam/ipam.go 489: Trying affinity for 192.168.18.128/26 host="172.24.4.241" Dec 13 02:41:27.162485 containerd[1466]: 2024-12-13 02:41:27.092 [INFO][2577] ipam/ipam.go 155: Attempting to load block cidr=192.168.18.128/26 host="172.24.4.241" Dec 13 02:41:27.162485 containerd[1466]: 2024-12-13 02:41:27.097 [INFO][2577] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.18.128/26 host="172.24.4.241" Dec 13 02:41:27.162485 containerd[1466]: 2024-12-13 02:41:27.097 [INFO][2577] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.18.128/26 handle="k8s-pod-network.54e65b34c2462000e6bc32fc82fa534166e6fd83d370edd3864a4686e91b65c8" host="172.24.4.241" Dec 13 02:41:27.162485 containerd[1466]: 2024-12-13 02:41:27.101 [INFO][2577] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.54e65b34c2462000e6bc32fc82fa534166e6fd83d370edd3864a4686e91b65c8 Dec 13 02:41:27.162485 containerd[1466]: 2024-12-13 02:41:27.108 [INFO][2577] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.18.128/26 handle="k8s-pod-network.54e65b34c2462000e6bc32fc82fa534166e6fd83d370edd3864a4686e91b65c8" host="172.24.4.241" Dec 13 02:41:27.162485 containerd[1466]: 2024-12-13 02:41:27.121 [INFO][2577] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.18.129/26] block=192.168.18.128/26 handle="k8s-pod-network.54e65b34c2462000e6bc32fc82fa534166e6fd83d370edd3864a4686e91b65c8" host="172.24.4.241" Dec 13 02:41:27.162485 containerd[1466]: 2024-12-13 02:41:27.121 [INFO][2577] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.18.129/26] handle="k8s-pod-network.54e65b34c2462000e6bc32fc82fa534166e6fd83d370edd3864a4686e91b65c8" host="172.24.4.241" Dec 13 02:41:27.162485 containerd[1466]: 2024-12-13 02:41:27.121 [INFO][2577] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Dec 13 02:41:27.162485 containerd[1466]: 2024-12-13 02:41:27.122 [INFO][2577] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.18.129/26] IPv6=[] ContainerID="54e65b34c2462000e6bc32fc82fa534166e6fd83d370edd3864a4686e91b65c8" HandleID="k8s-pod-network.54e65b34c2462000e6bc32fc82fa534166e6fd83d370edd3864a4686e91b65c8" Workload="172.24.4.241-k8s-csi--node--driver--mn4q7-eth0" Dec 13 02:41:27.164066 containerd[1466]: 2024-12-13 02:41:27.124 [INFO][2557] cni-plugin/k8s.go 386: Populated endpoint ContainerID="54e65b34c2462000e6bc32fc82fa534166e6fd83d370edd3864a4686e91b65c8" Namespace="calico-system" Pod="csi-node-driver-mn4q7" WorkloadEndpoint="172.24.4.241-k8s-csi--node--driver--mn4q7-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172.24.4.241-k8s-csi--node--driver--mn4q7-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"ed687165-b432-4311-a017-bee89dd86d6d", ResourceVersion:"1122", Generation:0, CreationTimestamp:time.Date(2024, time.December, 13, 2, 40, 52, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"55b695c467", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172.24.4.241", ContainerID:"", Pod:"csi-node-driver-mn4q7", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.18.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali2ea430d9eb4", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Dec 13 02:41:27.164066 containerd[1466]: 2024-12-13 02:41:27.124 [INFO][2557] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.18.129/32] ContainerID="54e65b34c2462000e6bc32fc82fa534166e6fd83d370edd3864a4686e91b65c8" Namespace="calico-system" Pod="csi-node-driver-mn4q7" WorkloadEndpoint="172.24.4.241-k8s-csi--node--driver--mn4q7-eth0" Dec 13 02:41:27.164066 containerd[1466]: 2024-12-13 02:41:27.124 [INFO][2557] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali2ea430d9eb4 ContainerID="54e65b34c2462000e6bc32fc82fa534166e6fd83d370edd3864a4686e91b65c8" Namespace="calico-system" Pod="csi-node-driver-mn4q7" WorkloadEndpoint="172.24.4.241-k8s-csi--node--driver--mn4q7-eth0" Dec 13 02:41:27.164066 containerd[1466]: 2024-12-13 02:41:27.139 [INFO][2557] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="54e65b34c2462000e6bc32fc82fa534166e6fd83d370edd3864a4686e91b65c8" Namespace="calico-system" Pod="csi-node-driver-mn4q7" WorkloadEndpoint="172.24.4.241-k8s-csi--node--driver--mn4q7-eth0" Dec 13 02:41:27.164066 containerd[1466]: 2024-12-13 02:41:27.140 [INFO][2557] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="54e65b34c2462000e6bc32fc82fa534166e6fd83d370edd3864a4686e91b65c8" Namespace="calico-system" Pod="csi-node-driver-mn4q7" WorkloadEndpoint="172.24.4.241-k8s-csi--node--driver--mn4q7-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172.24.4.241-k8s-csi--node--driver--mn4q7-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"ed687165-b432-4311-a017-bee89dd86d6d", ResourceVersion:"1122", Generation:0, CreationTimestamp:time.Date(2024, time.December, 13, 2, 40, 52, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"55b695c467", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172.24.4.241", ContainerID:"54e65b34c2462000e6bc32fc82fa534166e6fd83d370edd3864a4686e91b65c8", Pod:"csi-node-driver-mn4q7", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.18.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali2ea430d9eb4", MAC:"36:e4:f1:a3:ef:44", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Dec 13 02:41:27.164066 containerd[1466]: 2024-12-13 02:41:27.159 [INFO][2557] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="54e65b34c2462000e6bc32fc82fa534166e6fd83d370edd3864a4686e91b65c8" Namespace="calico-system" Pod="csi-node-driver-mn4q7" WorkloadEndpoint="172.24.4.241-k8s-csi--node--driver--mn4q7-eth0" Dec 13 02:41:27.196805 containerd[1466]: time="2024-12-13T02:41:27.196124610Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 02:41:27.196805 containerd[1466]: time="2024-12-13T02:41:27.196332658Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 02:41:27.196805 containerd[1466]: time="2024-12-13T02:41:27.196382120Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 02:41:27.196805 containerd[1466]: time="2024-12-13T02:41:27.196564230Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 02:41:27.222109 systemd[1]: Started cri-containerd-54e65b34c2462000e6bc32fc82fa534166e6fd83d370edd3864a4686e91b65c8.scope - libcontainer container 54e65b34c2462000e6bc32fc82fa534166e6fd83d370edd3864a4686e91b65c8. Dec 13 02:41:27.252430 containerd[1466]: time="2024-12-13T02:41:27.252385954Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-mn4q7,Uid:ed687165-b432-4311-a017-bee89dd86d6d,Namespace:calico-system,Attempt:1,} returns sandbox id \"54e65b34c2462000e6bc32fc82fa534166e6fd83d370edd3864a4686e91b65c8\"" Dec 13 02:41:27.254973 containerd[1466]: time="2024-12-13T02:41:27.254932400Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.1\"" Dec 13 02:41:27.324078 kubelet[1830]: E1213 02:41:27.323991 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:27.775927 kernel: bpftool[2767]: memfd_create() called without MFD_EXEC or MFD_NOEXEC_SEAL set Dec 13 02:41:28.098335 systemd-networkd[1371]: vxlan.calico: Link UP Dec 13 02:41:28.098343 systemd-networkd[1371]: vxlan.calico: Gained carrier Dec 13 02:41:28.324589 kubelet[1830]: E1213 02:41:28.324509 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:28.616395 containerd[1466]: time="2024-12-13T02:41:28.614831054Z" level=info msg="StopPodSandbox for \"3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19\"" Dec 13 02:41:28.780045 containerd[1466]: 2024-12-13 02:41:28.718 [INFO][2847] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19" Dec 13 02:41:28.780045 containerd[1466]: 2024-12-13 02:41:28.719 [INFO][2847] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19" iface="eth0" netns="/var/run/netns/cni-13cfbfff-3eb1-ca8b-5aca-ca5f71e714e2" Dec 13 02:41:28.780045 containerd[1466]: 2024-12-13 02:41:28.719 [INFO][2847] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19" iface="eth0" netns="/var/run/netns/cni-13cfbfff-3eb1-ca8b-5aca-ca5f71e714e2" Dec 13 02:41:28.780045 containerd[1466]: 2024-12-13 02:41:28.720 [INFO][2847] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19" iface="eth0" netns="/var/run/netns/cni-13cfbfff-3eb1-ca8b-5aca-ca5f71e714e2" Dec 13 02:41:28.780045 containerd[1466]: 2024-12-13 02:41:28.720 [INFO][2847] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19" Dec 13 02:41:28.780045 containerd[1466]: 2024-12-13 02:41:28.720 [INFO][2847] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19" Dec 13 02:41:28.780045 containerd[1466]: 2024-12-13 02:41:28.750 [INFO][2853] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19" HandleID="k8s-pod-network.3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19" Workload="172.24.4.241-k8s-nginx--deployment--6d5f899847--lblsv-eth0" Dec 13 02:41:28.780045 containerd[1466]: 2024-12-13 02:41:28.750 [INFO][2853] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Dec 13 02:41:28.780045 containerd[1466]: 2024-12-13 02:41:28.750 [INFO][2853] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Dec 13 02:41:28.780045 containerd[1466]: 2024-12-13 02:41:28.768 [WARNING][2853] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19" HandleID="k8s-pod-network.3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19" Workload="172.24.4.241-k8s-nginx--deployment--6d5f899847--lblsv-eth0" Dec 13 02:41:28.780045 containerd[1466]: 2024-12-13 02:41:28.768 [INFO][2853] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19" HandleID="k8s-pod-network.3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19" Workload="172.24.4.241-k8s-nginx--deployment--6d5f899847--lblsv-eth0" Dec 13 02:41:28.780045 containerd[1466]: 2024-12-13 02:41:28.772 [INFO][2853] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Dec 13 02:41:28.780045 containerd[1466]: 2024-12-13 02:41:28.774 [INFO][2847] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19" Dec 13 02:41:28.781568 containerd[1466]: time="2024-12-13T02:41:28.781339262Z" level=info msg="TearDown network for sandbox \"3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19\" successfully" Dec 13 02:41:28.781568 containerd[1466]: time="2024-12-13T02:41:28.781401938Z" level=info msg="StopPodSandbox for \"3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19\" returns successfully" Dec 13 02:41:28.781784 systemd[1]: run-netns-cni\x2d13cfbfff\x2d3eb1\x2dca8b\x2d5aca\x2dca5f71e714e2.mount: Deactivated successfully. Dec 13 02:41:28.786616 containerd[1466]: time="2024-12-13T02:41:28.784718301Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-6d5f899847-lblsv,Uid:1c2ae6bf-c9e7-4d97-8deb-2cd8f7181aa0,Namespace:default,Attempt:1,}" Dec 13 02:41:28.988284 systemd-networkd[1371]: cali2ea430d9eb4: Gained IPv6LL Dec 13 02:41:29.070558 systemd-networkd[1371]: cali5ce7bd04b58: Link UP Dec 13 02:41:29.072222 systemd-networkd[1371]: cali5ce7bd04b58: Gained carrier Dec 13 02:41:29.095031 containerd[1466]: 2024-12-13 02:41:28.892 [INFO][2860] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {172.24.4.241-k8s-nginx--deployment--6d5f899847--lblsv-eth0 nginx-deployment-6d5f899847- default 1c2ae6bf-c9e7-4d97-8deb-2cd8f7181aa0 1132 0 2024-12-13 02:41:15 +0000 UTC map[app:nginx pod-template-hash:6d5f899847 projectcalico.org/namespace:default projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:default] map[] [] [] []} {k8s 172.24.4.241 nginx-deployment-6d5f899847-lblsv eth0 default [] [] [kns.default ksa.default.default] cali5ce7bd04b58 [] []}} ContainerID="f65a4a221942baeae7ec44e3c5c94868a4c59c70eb11907f41d3eb7c34f45808" Namespace="default" Pod="nginx-deployment-6d5f899847-lblsv" WorkloadEndpoint="172.24.4.241-k8s-nginx--deployment--6d5f899847--lblsv-" Dec 13 02:41:29.095031 containerd[1466]: 2024-12-13 02:41:28.892 [INFO][2860] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="f65a4a221942baeae7ec44e3c5c94868a4c59c70eb11907f41d3eb7c34f45808" Namespace="default" Pod="nginx-deployment-6d5f899847-lblsv" WorkloadEndpoint="172.24.4.241-k8s-nginx--deployment--6d5f899847--lblsv-eth0" Dec 13 02:41:29.095031 containerd[1466]: 2024-12-13 02:41:28.941 [INFO][2873] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="f65a4a221942baeae7ec44e3c5c94868a4c59c70eb11907f41d3eb7c34f45808" HandleID="k8s-pod-network.f65a4a221942baeae7ec44e3c5c94868a4c59c70eb11907f41d3eb7c34f45808" Workload="172.24.4.241-k8s-nginx--deployment--6d5f899847--lblsv-eth0" Dec 13 02:41:29.095031 containerd[1466]: 2024-12-13 02:41:28.970 [INFO][2873] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="f65a4a221942baeae7ec44e3c5c94868a4c59c70eb11907f41d3eb7c34f45808" HandleID="k8s-pod-network.f65a4a221942baeae7ec44e3c5c94868a4c59c70eb11907f41d3eb7c34f45808" Workload="172.24.4.241-k8s-nginx--deployment--6d5f899847--lblsv-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000319510), Attrs:map[string]string{"namespace":"default", "node":"172.24.4.241", "pod":"nginx-deployment-6d5f899847-lblsv", "timestamp":"2024-12-13 02:41:28.94123947 +0000 UTC"}, Hostname:"172.24.4.241", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Dec 13 02:41:29.095031 containerd[1466]: 2024-12-13 02:41:28.970 [INFO][2873] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Dec 13 02:41:29.095031 containerd[1466]: 2024-12-13 02:41:28.970 [INFO][2873] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Dec 13 02:41:29.095031 containerd[1466]: 2024-12-13 02:41:28.970 [INFO][2873] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host '172.24.4.241' Dec 13 02:41:29.095031 containerd[1466]: 2024-12-13 02:41:28.978 [INFO][2873] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.f65a4a221942baeae7ec44e3c5c94868a4c59c70eb11907f41d3eb7c34f45808" host="172.24.4.241" Dec 13 02:41:29.095031 containerd[1466]: 2024-12-13 02:41:28.990 [INFO][2873] ipam/ipam.go 372: Looking up existing affinities for host host="172.24.4.241" Dec 13 02:41:29.095031 containerd[1466]: 2024-12-13 02:41:29.004 [INFO][2873] ipam/ipam.go 489: Trying affinity for 192.168.18.128/26 host="172.24.4.241" Dec 13 02:41:29.095031 containerd[1466]: 2024-12-13 02:41:29.013 [INFO][2873] ipam/ipam.go 155: Attempting to load block cidr=192.168.18.128/26 host="172.24.4.241" Dec 13 02:41:29.095031 containerd[1466]: 2024-12-13 02:41:29.018 [INFO][2873] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.18.128/26 host="172.24.4.241" Dec 13 02:41:29.095031 containerd[1466]: 2024-12-13 02:41:29.018 [INFO][2873] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.18.128/26 handle="k8s-pod-network.f65a4a221942baeae7ec44e3c5c94868a4c59c70eb11907f41d3eb7c34f45808" host="172.24.4.241" Dec 13 02:41:29.095031 containerd[1466]: 2024-12-13 02:41:29.025 [INFO][2873] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.f65a4a221942baeae7ec44e3c5c94868a4c59c70eb11907f41d3eb7c34f45808 Dec 13 02:41:29.095031 containerd[1466]: 2024-12-13 02:41:29.040 [INFO][2873] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.18.128/26 handle="k8s-pod-network.f65a4a221942baeae7ec44e3c5c94868a4c59c70eb11907f41d3eb7c34f45808" host="172.24.4.241" Dec 13 02:41:29.095031 containerd[1466]: 2024-12-13 02:41:29.059 [INFO][2873] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.18.130/26] block=192.168.18.128/26 handle="k8s-pod-network.f65a4a221942baeae7ec44e3c5c94868a4c59c70eb11907f41d3eb7c34f45808" host="172.24.4.241" Dec 13 02:41:29.095031 containerd[1466]: 2024-12-13 02:41:29.060 [INFO][2873] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.18.130/26] handle="k8s-pod-network.f65a4a221942baeae7ec44e3c5c94868a4c59c70eb11907f41d3eb7c34f45808" host="172.24.4.241" Dec 13 02:41:29.095031 containerd[1466]: 2024-12-13 02:41:29.060 [INFO][2873] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Dec 13 02:41:29.095031 containerd[1466]: 2024-12-13 02:41:29.060 [INFO][2873] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.18.130/26] IPv6=[] ContainerID="f65a4a221942baeae7ec44e3c5c94868a4c59c70eb11907f41d3eb7c34f45808" HandleID="k8s-pod-network.f65a4a221942baeae7ec44e3c5c94868a4c59c70eb11907f41d3eb7c34f45808" Workload="172.24.4.241-k8s-nginx--deployment--6d5f899847--lblsv-eth0" Dec 13 02:41:29.096309 containerd[1466]: 2024-12-13 02:41:29.064 [INFO][2860] cni-plugin/k8s.go 386: Populated endpoint ContainerID="f65a4a221942baeae7ec44e3c5c94868a4c59c70eb11907f41d3eb7c34f45808" Namespace="default" Pod="nginx-deployment-6d5f899847-lblsv" WorkloadEndpoint="172.24.4.241-k8s-nginx--deployment--6d5f899847--lblsv-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172.24.4.241-k8s-nginx--deployment--6d5f899847--lblsv-eth0", GenerateName:"nginx-deployment-6d5f899847-", Namespace:"default", SelfLink:"", UID:"1c2ae6bf-c9e7-4d97-8deb-2cd8f7181aa0", ResourceVersion:"1132", Generation:0, CreationTimestamp:time.Date(2024, time.December, 13, 2, 41, 15, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app":"nginx", "pod-template-hash":"6d5f899847", "projectcalico.org/namespace":"default", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"default"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172.24.4.241", ContainerID:"", Pod:"nginx-deployment-6d5f899847-lblsv", Endpoint:"eth0", ServiceAccountName:"default", IPNetworks:[]string{"192.168.18.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.default", "ksa.default.default"}, InterfaceName:"cali5ce7bd04b58", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Dec 13 02:41:29.096309 containerd[1466]: 2024-12-13 02:41:29.064 [INFO][2860] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.18.130/32] ContainerID="f65a4a221942baeae7ec44e3c5c94868a4c59c70eb11907f41d3eb7c34f45808" Namespace="default" Pod="nginx-deployment-6d5f899847-lblsv" WorkloadEndpoint="172.24.4.241-k8s-nginx--deployment--6d5f899847--lblsv-eth0" Dec 13 02:41:29.096309 containerd[1466]: 2024-12-13 02:41:29.064 [INFO][2860] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali5ce7bd04b58 ContainerID="f65a4a221942baeae7ec44e3c5c94868a4c59c70eb11907f41d3eb7c34f45808" Namespace="default" Pod="nginx-deployment-6d5f899847-lblsv" WorkloadEndpoint="172.24.4.241-k8s-nginx--deployment--6d5f899847--lblsv-eth0" Dec 13 02:41:29.096309 containerd[1466]: 2024-12-13 02:41:29.070 [INFO][2860] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="f65a4a221942baeae7ec44e3c5c94868a4c59c70eb11907f41d3eb7c34f45808" Namespace="default" Pod="nginx-deployment-6d5f899847-lblsv" WorkloadEndpoint="172.24.4.241-k8s-nginx--deployment--6d5f899847--lblsv-eth0" Dec 13 02:41:29.096309 containerd[1466]: 2024-12-13 02:41:29.071 [INFO][2860] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="f65a4a221942baeae7ec44e3c5c94868a4c59c70eb11907f41d3eb7c34f45808" Namespace="default" Pod="nginx-deployment-6d5f899847-lblsv" WorkloadEndpoint="172.24.4.241-k8s-nginx--deployment--6d5f899847--lblsv-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172.24.4.241-k8s-nginx--deployment--6d5f899847--lblsv-eth0", GenerateName:"nginx-deployment-6d5f899847-", Namespace:"default", SelfLink:"", UID:"1c2ae6bf-c9e7-4d97-8deb-2cd8f7181aa0", ResourceVersion:"1132", Generation:0, CreationTimestamp:time.Date(2024, time.December, 13, 2, 41, 15, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app":"nginx", "pod-template-hash":"6d5f899847", "projectcalico.org/namespace":"default", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"default"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172.24.4.241", ContainerID:"f65a4a221942baeae7ec44e3c5c94868a4c59c70eb11907f41d3eb7c34f45808", Pod:"nginx-deployment-6d5f899847-lblsv", Endpoint:"eth0", ServiceAccountName:"default", IPNetworks:[]string{"192.168.18.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.default", "ksa.default.default"}, InterfaceName:"cali5ce7bd04b58", MAC:"0a:39:67:13:18:b0", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Dec 13 02:41:29.096309 containerd[1466]: 2024-12-13 02:41:29.084 [INFO][2860] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="f65a4a221942baeae7ec44e3c5c94868a4c59c70eb11907f41d3eb7c34f45808" Namespace="default" Pod="nginx-deployment-6d5f899847-lblsv" WorkloadEndpoint="172.24.4.241-k8s-nginx--deployment--6d5f899847--lblsv-eth0" Dec 13 02:41:29.152420 containerd[1466]: time="2024-12-13T02:41:29.151243495Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 02:41:29.152420 containerd[1466]: time="2024-12-13T02:41:29.151367577Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 02:41:29.152420 containerd[1466]: time="2024-12-13T02:41:29.151388015Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 02:41:29.152420 containerd[1466]: time="2024-12-13T02:41:29.151515091Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 02:41:29.194193 systemd[1]: Started cri-containerd-f65a4a221942baeae7ec44e3c5c94868a4c59c70eb11907f41d3eb7c34f45808.scope - libcontainer container f65a4a221942baeae7ec44e3c5c94868a4c59c70eb11907f41d3eb7c34f45808. Dec 13 02:41:29.265856 containerd[1466]: time="2024-12-13T02:41:29.265759118Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-6d5f899847-lblsv,Uid:1c2ae6bf-c9e7-4d97-8deb-2cd8f7181aa0,Namespace:default,Attempt:1,} returns sandbox id \"f65a4a221942baeae7ec44e3c5c94868a4c59c70eb11907f41d3eb7c34f45808\"" Dec 13 02:41:29.325586 kubelet[1830]: E1213 02:41:29.325529 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:29.355087 containerd[1466]: time="2024-12-13T02:41:29.355032984Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 02:41:29.356167 containerd[1466]: time="2024-12-13T02:41:29.355892267Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.29.1: active requests=0, bytes read=7902632" Dec 13 02:41:29.356945 containerd[1466]: time="2024-12-13T02:41:29.356861203Z" level=info msg="ImageCreate event name:\"sha256:bda8c42e04758c4f061339e213f50ccdc7502c4176fbf631aa12357e62b63540\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 02:41:29.361169 containerd[1466]: time="2024-12-13T02:41:29.361126016Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:eaa7e01fb16b603c155a67b81f16992281db7f831684c7b2081d3434587a7ff3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 02:41:29.361984 containerd[1466]: time="2024-12-13T02:41:29.361947818Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.29.1\" with image id \"sha256:bda8c42e04758c4f061339e213f50ccdc7502c4176fbf631aa12357e62b63540\", repo tag \"ghcr.io/flatcar/calico/csi:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:eaa7e01fb16b603c155a67b81f16992281db7f831684c7b2081d3434587a7ff3\", size \"9395716\" in 2.106971506s" Dec 13 02:41:29.362032 containerd[1466]: time="2024-12-13T02:41:29.361984016Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.1\" returns image reference \"sha256:bda8c42e04758c4f061339e213f50ccdc7502c4176fbf631aa12357e62b63540\"" Dec 13 02:41:29.362852 containerd[1466]: time="2024-12-13T02:41:29.362814535Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\"" Dec 13 02:41:29.365211 containerd[1466]: time="2024-12-13T02:41:29.365178061Z" level=info msg="CreateContainer within sandbox \"54e65b34c2462000e6bc32fc82fa534166e6fd83d370edd3864a4686e91b65c8\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Dec 13 02:41:29.403573 containerd[1466]: time="2024-12-13T02:41:29.403396566Z" level=info msg="CreateContainer within sandbox \"54e65b34c2462000e6bc32fc82fa534166e6fd83d370edd3864a4686e91b65c8\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"013e043102bd224865390965c8dac7a4e5658ca52f3dd335a7754cfd5b267e9f\"" Dec 13 02:41:29.405014 containerd[1466]: time="2024-12-13T02:41:29.404962796Z" level=info msg="StartContainer for \"013e043102bd224865390965c8dac7a4e5658ca52f3dd335a7754cfd5b267e9f\"" Dec 13 02:41:29.447085 systemd[1]: Started cri-containerd-013e043102bd224865390965c8dac7a4e5658ca52f3dd335a7754cfd5b267e9f.scope - libcontainer container 013e043102bd224865390965c8dac7a4e5658ca52f3dd335a7754cfd5b267e9f. Dec 13 02:41:29.491293 containerd[1466]: time="2024-12-13T02:41:29.491234575Z" level=info msg="StartContainer for \"013e043102bd224865390965c8dac7a4e5658ca52f3dd335a7754cfd5b267e9f\" returns successfully" Dec 13 02:41:30.076539 systemd-networkd[1371]: vxlan.calico: Gained IPv6LL Dec 13 02:41:30.326940 kubelet[1830]: E1213 02:41:30.326651 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:30.782131 systemd-networkd[1371]: cali5ce7bd04b58: Gained IPv6LL Dec 13 02:41:31.327172 kubelet[1830]: E1213 02:41:31.327105 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:32.289338 kubelet[1830]: E1213 02:41:32.289235 1830 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:32.327349 kubelet[1830]: E1213 02:41:32.327289 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:33.328200 kubelet[1830]: E1213 02:41:33.328088 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:33.479630 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount320103140.mount: Deactivated successfully. Dec 13 02:41:34.329241 kubelet[1830]: E1213 02:41:34.329170 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:34.995524 containerd[1466]: time="2024-12-13T02:41:34.995355515Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/nginx:latest\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 02:41:34.997876 containerd[1466]: time="2024-12-13T02:41:34.997799136Z" level=info msg="stop pulling image ghcr.io/flatcar/nginx:latest: active requests=0, bytes read=71036027" Dec 13 02:41:35.001628 containerd[1466]: time="2024-12-13T02:41:35.001536170Z" level=info msg="ImageCreate event name:\"sha256:fa0a8cea5e76ad962111c39c85bb312edaf5b89eccd8f404eeea66c9759641e3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 02:41:35.010470 containerd[1466]: time="2024-12-13T02:41:35.010407248Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/nginx@sha256:e04edf30a4ea4c5a4107110797c72d3ee8a654415f00acd4019be17218afd9a1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 02:41:35.015949 containerd[1466]: time="2024-12-13T02:41:35.015687564Z" level=info msg="Pulled image \"ghcr.io/flatcar/nginx:latest\" with image id \"sha256:fa0a8cea5e76ad962111c39c85bb312edaf5b89eccd8f404eeea66c9759641e3\", repo tag \"ghcr.io/flatcar/nginx:latest\", repo digest \"ghcr.io/flatcar/nginx@sha256:e04edf30a4ea4c5a4107110797c72d3ee8a654415f00acd4019be17218afd9a1\", size \"71035905\" in 5.652834207s" Dec 13 02:41:35.015949 containerd[1466]: time="2024-12-13T02:41:35.015755971Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\" returns image reference \"sha256:fa0a8cea5e76ad962111c39c85bb312edaf5b89eccd8f404eeea66c9759641e3\"" Dec 13 02:41:35.021070 containerd[1466]: time="2024-12-13T02:41:35.020732361Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\"" Dec 13 02:41:35.023575 containerd[1466]: time="2024-12-13T02:41:35.023010432Z" level=info msg="CreateContainer within sandbox \"f65a4a221942baeae7ec44e3c5c94868a4c59c70eb11907f41d3eb7c34f45808\" for container &ContainerMetadata{Name:nginx,Attempt:0,}" Dec 13 02:41:35.053739 containerd[1466]: time="2024-12-13T02:41:35.053586083Z" level=info msg="CreateContainer within sandbox \"f65a4a221942baeae7ec44e3c5c94868a4c59c70eb11907f41d3eb7c34f45808\" for &ContainerMetadata{Name:nginx,Attempt:0,} returns container id \"8783e74cce3cbb2a1c10f542cb4627fc1d97b0e34d3ab4b5da7ffaebecfc9466\"" Dec 13 02:41:35.055645 containerd[1466]: time="2024-12-13T02:41:35.054560632Z" level=info msg="StartContainer for \"8783e74cce3cbb2a1c10f542cb4627fc1d97b0e34d3ab4b5da7ffaebecfc9466\"" Dec 13 02:41:35.110135 systemd[1]: Started cri-containerd-8783e74cce3cbb2a1c10f542cb4627fc1d97b0e34d3ab4b5da7ffaebecfc9466.scope - libcontainer container 8783e74cce3cbb2a1c10f542cb4627fc1d97b0e34d3ab4b5da7ffaebecfc9466. Dec 13 02:41:35.143218 containerd[1466]: time="2024-12-13T02:41:35.143114490Z" level=info msg="StartContainer for \"8783e74cce3cbb2a1c10f542cb4627fc1d97b0e34d3ab4b5da7ffaebecfc9466\" returns successfully" Dec 13 02:41:35.331118 kubelet[1830]: E1213 02:41:35.330871 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:36.332143 kubelet[1830]: E1213 02:41:36.332016 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:37.333427 kubelet[1830]: E1213 02:41:37.333323 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:38.333652 kubelet[1830]: E1213 02:41:38.333566 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:38.611782 containerd[1466]: time="2024-12-13T02:41:38.611608405Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 02:41:38.617152 containerd[1466]: time="2024-12-13T02:41:38.617090362Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1: active requests=0, bytes read=10501081" Dec 13 02:41:38.620522 containerd[1466]: time="2024-12-13T02:41:38.619433377Z" level=info msg="ImageCreate event name:\"sha256:8b7d18f262d5cf6a6343578ad0db68a140c4c9989d9e02c58c27cb5d2c70320f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 02:41:38.621798 containerd[1466]: time="2024-12-13T02:41:38.621754080Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:a338da9488cbaa83c78457c3d7354d84149969c0480e88dd768e036632ff5b76\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 02:41:38.624933 containerd[1466]: time="2024-12-13T02:41:38.624505587Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\" with image id \"sha256:8b7d18f262d5cf6a6343578ad0db68a140c4c9989d9e02c58c27cb5d2c70320f\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:a338da9488cbaa83c78457c3d7354d84149969c0480e88dd768e036632ff5b76\", size \"11994117\" in 3.603718386s" Dec 13 02:41:38.624933 containerd[1466]: time="2024-12-13T02:41:38.624561191Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\" returns image reference \"sha256:8b7d18f262d5cf6a6343578ad0db68a140c4c9989d9e02c58c27cb5d2c70320f\"" Dec 13 02:41:38.628660 containerd[1466]: time="2024-12-13T02:41:38.628605353Z" level=info msg="CreateContainer within sandbox \"54e65b34c2462000e6bc32fc82fa534166e6fd83d370edd3864a4686e91b65c8\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Dec 13 02:41:38.655782 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3164217692.mount: Deactivated successfully. Dec 13 02:41:38.667145 containerd[1466]: time="2024-12-13T02:41:38.666998421Z" level=info msg="CreateContainer within sandbox \"54e65b34c2462000e6bc32fc82fa534166e6fd83d370edd3864a4686e91b65c8\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"a997d2e2bae31335b4cea3879286532a61b48143e40678c560633dc9ca048044\"" Dec 13 02:41:38.668254 containerd[1466]: time="2024-12-13T02:41:38.668098185Z" level=info msg="StartContainer for \"a997d2e2bae31335b4cea3879286532a61b48143e40678c560633dc9ca048044\"" Dec 13 02:41:38.702741 systemd[1]: run-containerd-runc-k8s.io-a997d2e2bae31335b4cea3879286532a61b48143e40678c560633dc9ca048044-runc.faoiXv.mount: Deactivated successfully. Dec 13 02:41:38.713090 systemd[1]: Started cri-containerd-a997d2e2bae31335b4cea3879286532a61b48143e40678c560633dc9ca048044.scope - libcontainer container a997d2e2bae31335b4cea3879286532a61b48143e40678c560633dc9ca048044. Dec 13 02:41:38.750769 containerd[1466]: time="2024-12-13T02:41:38.750720379Z" level=info msg="StartContainer for \"a997d2e2bae31335b4cea3879286532a61b48143e40678c560633dc9ca048044\" returns successfully" Dec 13 02:41:39.054620 kubelet[1830]: I1213 02:41:39.054095 1830 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="calico-system/csi-node-driver-mn4q7" podStartSLOduration=35.683063959 podStartE2EDuration="47.053998868s" podCreationTimestamp="2024-12-13 02:40:52 +0000 UTC" firstStartedPulling="2024-12-13 02:41:27.254608547 +0000 UTC m=+36.071644338" lastFinishedPulling="2024-12-13 02:41:38.625543466 +0000 UTC m=+47.442579247" observedRunningTime="2024-12-13 02:41:39.053742259 +0000 UTC m=+47.870778090" watchObservedRunningTime="2024-12-13 02:41:39.053998868 +0000 UTC m=+47.871034699" Dec 13 02:41:39.054620 kubelet[1830]: I1213 02:41:39.054423 1830 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="default/nginx-deployment-6d5f899847-lblsv" podStartSLOduration=18.307495327 podStartE2EDuration="24.054370011s" podCreationTimestamp="2024-12-13 02:41:15 +0000 UTC" firstStartedPulling="2024-12-13 02:41:29.270381447 +0000 UTC m=+38.087417229" lastFinishedPulling="2024-12-13 02:41:35.017256082 +0000 UTC m=+43.834291913" observedRunningTime="2024-12-13 02:41:36.053005764 +0000 UTC m=+44.870041595" watchObservedRunningTime="2024-12-13 02:41:39.054370011 +0000 UTC m=+47.871405842" Dec 13 02:41:39.334770 kubelet[1830]: E1213 02:41:39.334520 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:39.551317 kubelet[1830]: I1213 02:41:39.551238 1830 csi_plugin.go:99] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Dec 13 02:41:39.554770 kubelet[1830]: I1213 02:41:39.554689 1830 csi_plugin.go:112] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Dec 13 02:41:40.334885 kubelet[1830]: E1213 02:41:40.334750 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:41.335563 kubelet[1830]: E1213 02:41:41.335400 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:42.336615 kubelet[1830]: E1213 02:41:42.336499 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:43.337257 kubelet[1830]: E1213 02:41:43.337156 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:44.337865 kubelet[1830]: E1213 02:41:44.337761 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:45.338382 kubelet[1830]: E1213 02:41:45.338294 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:46.339180 kubelet[1830]: E1213 02:41:46.339094 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:47.339456 kubelet[1830]: E1213 02:41:47.339335 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:47.698029 kubelet[1830]: I1213 02:41:47.696682 1830 topology_manager.go:215] "Topology Admit Handler" podUID="c8676952-7320-49d9-8f64-e1c4f5119ad9" podNamespace="default" podName="nfs-server-provisioner-0" Dec 13 02:41:47.712004 systemd[1]: Created slice kubepods-besteffort-podc8676952_7320_49d9_8f64_e1c4f5119ad9.slice - libcontainer container kubepods-besteffort-podc8676952_7320_49d9_8f64_e1c4f5119ad9.slice. Dec 13 02:41:47.776673 kubelet[1830]: I1213 02:41:47.776600 1830 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data\" (UniqueName: \"kubernetes.io/empty-dir/c8676952-7320-49d9-8f64-e1c4f5119ad9-data\") pod \"nfs-server-provisioner-0\" (UID: \"c8676952-7320-49d9-8f64-e1c4f5119ad9\") " pod="default/nfs-server-provisioner-0" Dec 13 02:41:47.776892 kubelet[1830]: I1213 02:41:47.776706 1830 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8z98\" (UniqueName: \"kubernetes.io/projected/c8676952-7320-49d9-8f64-e1c4f5119ad9-kube-api-access-d8z98\") pod \"nfs-server-provisioner-0\" (UID: \"c8676952-7320-49d9-8f64-e1c4f5119ad9\") " pod="default/nfs-server-provisioner-0" Dec 13 02:41:48.021820 containerd[1466]: time="2024-12-13T02:41:48.021664989Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nfs-server-provisioner-0,Uid:c8676952-7320-49d9-8f64-e1c4f5119ad9,Namespace:default,Attempt:0,}" Dec 13 02:41:48.333589 systemd-networkd[1371]: cali60e51b789ff: Link UP Dec 13 02:41:48.336176 systemd-networkd[1371]: cali60e51b789ff: Gained carrier Dec 13 02:41:48.340669 kubelet[1830]: E1213 02:41:48.340614 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:48.359051 containerd[1466]: 2024-12-13 02:41:48.141 [INFO][3128] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {172.24.4.241-k8s-nfs--server--provisioner--0-eth0 nfs-server-provisioner- default c8676952-7320-49d9-8f64-e1c4f5119ad9 1207 0 2024-12-13 02:41:47 +0000 UTC map[app:nfs-server-provisioner apps.kubernetes.io/pod-index:0 chart:nfs-server-provisioner-1.8.0 controller-revision-hash:nfs-server-provisioner-d5cbb7f57 heritage:Helm projectcalico.org/namespace:default projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:nfs-server-provisioner release:nfs-server-provisioner statefulset.kubernetes.io/pod-name:nfs-server-provisioner-0] map[] [] [] []} {k8s 172.24.4.241 nfs-server-provisioner-0 eth0 nfs-server-provisioner [] [] [kns.default ksa.default.nfs-server-provisioner] cali60e51b789ff [{nfs TCP 2049 0 } {nfs-udp UDP 2049 0 } {nlockmgr TCP 32803 0 } {nlockmgr-udp UDP 32803 0 } {mountd TCP 20048 0 } {mountd-udp UDP 20048 0 } {rquotad TCP 875 0 } {rquotad-udp UDP 875 0 } {rpcbind TCP 111 0 } {rpcbind-udp UDP 111 0 } {statd TCP 662 0 } {statd-udp UDP 662 0 }] []}} ContainerID="ac4e4815181bfb0eb8425e35992b7bf8fd8759c2bac2353395a10a42bedf4da2" Namespace="default" Pod="nfs-server-provisioner-0" WorkloadEndpoint="172.24.4.241-k8s-nfs--server--provisioner--0-" Dec 13 02:41:48.359051 containerd[1466]: 2024-12-13 02:41:48.141 [INFO][3128] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="ac4e4815181bfb0eb8425e35992b7bf8fd8759c2bac2353395a10a42bedf4da2" Namespace="default" Pod="nfs-server-provisioner-0" WorkloadEndpoint="172.24.4.241-k8s-nfs--server--provisioner--0-eth0" Dec 13 02:41:48.359051 containerd[1466]: 2024-12-13 02:41:48.228 [INFO][3139] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="ac4e4815181bfb0eb8425e35992b7bf8fd8759c2bac2353395a10a42bedf4da2" HandleID="k8s-pod-network.ac4e4815181bfb0eb8425e35992b7bf8fd8759c2bac2353395a10a42bedf4da2" Workload="172.24.4.241-k8s-nfs--server--provisioner--0-eth0" Dec 13 02:41:48.359051 containerd[1466]: 2024-12-13 02:41:48.252 [INFO][3139] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="ac4e4815181bfb0eb8425e35992b7bf8fd8759c2bac2353395a10a42bedf4da2" HandleID="k8s-pod-network.ac4e4815181bfb0eb8425e35992b7bf8fd8759c2bac2353395a10a42bedf4da2" Workload="172.24.4.241-k8s-nfs--server--provisioner--0-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0003bc780), Attrs:map[string]string{"namespace":"default", "node":"172.24.4.241", "pod":"nfs-server-provisioner-0", "timestamp":"2024-12-13 02:41:48.228120696 +0000 UTC"}, Hostname:"172.24.4.241", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Dec 13 02:41:48.359051 containerd[1466]: 2024-12-13 02:41:48.252 [INFO][3139] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Dec 13 02:41:48.359051 containerd[1466]: 2024-12-13 02:41:48.252 [INFO][3139] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Dec 13 02:41:48.359051 containerd[1466]: 2024-12-13 02:41:48.253 [INFO][3139] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host '172.24.4.241' Dec 13 02:41:48.359051 containerd[1466]: 2024-12-13 02:41:48.257 [INFO][3139] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.ac4e4815181bfb0eb8425e35992b7bf8fd8759c2bac2353395a10a42bedf4da2" host="172.24.4.241" Dec 13 02:41:48.359051 containerd[1466]: 2024-12-13 02:41:48.271 [INFO][3139] ipam/ipam.go 372: Looking up existing affinities for host host="172.24.4.241" Dec 13 02:41:48.359051 containerd[1466]: 2024-12-13 02:41:48.282 [INFO][3139] ipam/ipam.go 489: Trying affinity for 192.168.18.128/26 host="172.24.4.241" Dec 13 02:41:48.359051 containerd[1466]: 2024-12-13 02:41:48.287 [INFO][3139] ipam/ipam.go 155: Attempting to load block cidr=192.168.18.128/26 host="172.24.4.241" Dec 13 02:41:48.359051 containerd[1466]: 2024-12-13 02:41:48.294 [INFO][3139] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.18.128/26 host="172.24.4.241" Dec 13 02:41:48.359051 containerd[1466]: 2024-12-13 02:41:48.295 [INFO][3139] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.18.128/26 handle="k8s-pod-network.ac4e4815181bfb0eb8425e35992b7bf8fd8759c2bac2353395a10a42bedf4da2" host="172.24.4.241" Dec 13 02:41:48.359051 containerd[1466]: 2024-12-13 02:41:48.299 [INFO][3139] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.ac4e4815181bfb0eb8425e35992b7bf8fd8759c2bac2353395a10a42bedf4da2 Dec 13 02:41:48.359051 containerd[1466]: 2024-12-13 02:41:48.310 [INFO][3139] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.18.128/26 handle="k8s-pod-network.ac4e4815181bfb0eb8425e35992b7bf8fd8759c2bac2353395a10a42bedf4da2" host="172.24.4.241" Dec 13 02:41:48.359051 containerd[1466]: 2024-12-13 02:41:48.322 [INFO][3139] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.18.131/26] block=192.168.18.128/26 handle="k8s-pod-network.ac4e4815181bfb0eb8425e35992b7bf8fd8759c2bac2353395a10a42bedf4da2" host="172.24.4.241" Dec 13 02:41:48.359051 containerd[1466]: 2024-12-13 02:41:48.322 [INFO][3139] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.18.131/26] handle="k8s-pod-network.ac4e4815181bfb0eb8425e35992b7bf8fd8759c2bac2353395a10a42bedf4da2" host="172.24.4.241" Dec 13 02:41:48.359051 containerd[1466]: 2024-12-13 02:41:48.322 [INFO][3139] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Dec 13 02:41:48.359051 containerd[1466]: 2024-12-13 02:41:48.322 [INFO][3139] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.18.131/26] IPv6=[] ContainerID="ac4e4815181bfb0eb8425e35992b7bf8fd8759c2bac2353395a10a42bedf4da2" HandleID="k8s-pod-network.ac4e4815181bfb0eb8425e35992b7bf8fd8759c2bac2353395a10a42bedf4da2" Workload="172.24.4.241-k8s-nfs--server--provisioner--0-eth0" Dec 13 02:41:48.360771 containerd[1466]: 2024-12-13 02:41:48.326 [INFO][3128] cni-plugin/k8s.go 386: Populated endpoint ContainerID="ac4e4815181bfb0eb8425e35992b7bf8fd8759c2bac2353395a10a42bedf4da2" Namespace="default" Pod="nfs-server-provisioner-0" WorkloadEndpoint="172.24.4.241-k8s-nfs--server--provisioner--0-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172.24.4.241-k8s-nfs--server--provisioner--0-eth0", GenerateName:"nfs-server-provisioner-", Namespace:"default", SelfLink:"", UID:"c8676952-7320-49d9-8f64-e1c4f5119ad9", ResourceVersion:"1207", Generation:0, CreationTimestamp:time.Date(2024, time.December, 13, 2, 41, 47, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app":"nfs-server-provisioner", "apps.kubernetes.io/pod-index":"0", "chart":"nfs-server-provisioner-1.8.0", "controller-revision-hash":"nfs-server-provisioner-d5cbb7f57", "heritage":"Helm", "projectcalico.org/namespace":"default", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"nfs-server-provisioner", "release":"nfs-server-provisioner", "statefulset.kubernetes.io/pod-name":"nfs-server-provisioner-0"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172.24.4.241", ContainerID:"", Pod:"nfs-server-provisioner-0", Endpoint:"eth0", ServiceAccountName:"nfs-server-provisioner", IPNetworks:[]string{"192.168.18.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.default", "ksa.default.nfs-server-provisioner"}, InterfaceName:"cali60e51b789ff", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"nfs", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x801, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"nfs-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x801, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"nlockmgr", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x8023, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"nlockmgr-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x8023, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"mountd", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x4e50, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"mountd-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x4e50, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"rquotad", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x36b, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"rquotad-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x36b, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"rpcbind", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x6f, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"rpcbind-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x6f, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"statd", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x296, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"statd-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x296, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Dec 13 02:41:48.360771 containerd[1466]: 2024-12-13 02:41:48.326 [INFO][3128] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.18.131/32] ContainerID="ac4e4815181bfb0eb8425e35992b7bf8fd8759c2bac2353395a10a42bedf4da2" Namespace="default" Pod="nfs-server-provisioner-0" WorkloadEndpoint="172.24.4.241-k8s-nfs--server--provisioner--0-eth0" Dec 13 02:41:48.360771 containerd[1466]: 2024-12-13 02:41:48.326 [INFO][3128] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali60e51b789ff ContainerID="ac4e4815181bfb0eb8425e35992b7bf8fd8759c2bac2353395a10a42bedf4da2" Namespace="default" Pod="nfs-server-provisioner-0" WorkloadEndpoint="172.24.4.241-k8s-nfs--server--provisioner--0-eth0" Dec 13 02:41:48.360771 containerd[1466]: 2024-12-13 02:41:48.335 [INFO][3128] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="ac4e4815181bfb0eb8425e35992b7bf8fd8759c2bac2353395a10a42bedf4da2" Namespace="default" Pod="nfs-server-provisioner-0" WorkloadEndpoint="172.24.4.241-k8s-nfs--server--provisioner--0-eth0" Dec 13 02:41:48.362136 containerd[1466]: 2024-12-13 02:41:48.336 [INFO][3128] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="ac4e4815181bfb0eb8425e35992b7bf8fd8759c2bac2353395a10a42bedf4da2" Namespace="default" Pod="nfs-server-provisioner-0" WorkloadEndpoint="172.24.4.241-k8s-nfs--server--provisioner--0-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172.24.4.241-k8s-nfs--server--provisioner--0-eth0", GenerateName:"nfs-server-provisioner-", Namespace:"default", SelfLink:"", UID:"c8676952-7320-49d9-8f64-e1c4f5119ad9", ResourceVersion:"1207", Generation:0, CreationTimestamp:time.Date(2024, time.December, 13, 2, 41, 47, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app":"nfs-server-provisioner", "apps.kubernetes.io/pod-index":"0", "chart":"nfs-server-provisioner-1.8.0", "controller-revision-hash":"nfs-server-provisioner-d5cbb7f57", "heritage":"Helm", "projectcalico.org/namespace":"default", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"nfs-server-provisioner", "release":"nfs-server-provisioner", "statefulset.kubernetes.io/pod-name":"nfs-server-provisioner-0"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172.24.4.241", ContainerID:"ac4e4815181bfb0eb8425e35992b7bf8fd8759c2bac2353395a10a42bedf4da2", Pod:"nfs-server-provisioner-0", Endpoint:"eth0", ServiceAccountName:"nfs-server-provisioner", IPNetworks:[]string{"192.168.18.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.default", "ksa.default.nfs-server-provisioner"}, InterfaceName:"cali60e51b789ff", MAC:"62:47:26:ab:19:f1", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"nfs", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x801, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"nfs-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x801, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"nlockmgr", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x8023, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"nlockmgr-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x8023, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"mountd", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x4e50, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"mountd-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x4e50, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"rquotad", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x36b, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"rquotad-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x36b, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"rpcbind", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x6f, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"rpcbind-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x6f, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"statd", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x296, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"statd-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x296, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Dec 13 02:41:48.362136 containerd[1466]: 2024-12-13 02:41:48.354 [INFO][3128] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="ac4e4815181bfb0eb8425e35992b7bf8fd8759c2bac2353395a10a42bedf4da2" Namespace="default" Pod="nfs-server-provisioner-0" WorkloadEndpoint="172.24.4.241-k8s-nfs--server--provisioner--0-eth0" Dec 13 02:41:48.408520 containerd[1466]: time="2024-12-13T02:41:48.407736587Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 02:41:48.412041 containerd[1466]: time="2024-12-13T02:41:48.410924714Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 02:41:48.412041 containerd[1466]: time="2024-12-13T02:41:48.410950993Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 02:41:48.412041 containerd[1466]: time="2024-12-13T02:41:48.411058524Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 02:41:48.436139 systemd[1]: Started cri-containerd-ac4e4815181bfb0eb8425e35992b7bf8fd8759c2bac2353395a10a42bedf4da2.scope - libcontainer container ac4e4815181bfb0eb8425e35992b7bf8fd8759c2bac2353395a10a42bedf4da2. Dec 13 02:41:48.491459 containerd[1466]: time="2024-12-13T02:41:48.491163281Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nfs-server-provisioner-0,Uid:c8676952-7320-49d9-8f64-e1c4f5119ad9,Namespace:default,Attempt:0,} returns sandbox id \"ac4e4815181bfb0eb8425e35992b7bf8fd8759c2bac2353395a10a42bedf4da2\"" Dec 13 02:41:48.493388 containerd[1466]: time="2024-12-13T02:41:48.493362010Z" level=info msg="PullImage \"registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8\"" Dec 13 02:41:49.342053 kubelet[1830]: E1213 02:41:49.341986 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:49.404103 systemd-networkd[1371]: cali60e51b789ff: Gained IPv6LL Dec 13 02:41:50.342557 kubelet[1830]: E1213 02:41:50.342491 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:51.344098 kubelet[1830]: E1213 02:41:51.344009 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:52.289368 kubelet[1830]: E1213 02:41:52.289281 1830 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:52.341065 containerd[1466]: time="2024-12-13T02:41:52.340880784Z" level=info msg="StopPodSandbox for \"3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19\"" Dec 13 02:41:52.344736 kubelet[1830]: E1213 02:41:52.344696 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:52.500516 containerd[1466]: 2024-12-13 02:41:52.446 [WARNING][3219] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172.24.4.241-k8s-nginx--deployment--6d5f899847--lblsv-eth0", GenerateName:"nginx-deployment-6d5f899847-", Namespace:"default", SelfLink:"", UID:"1c2ae6bf-c9e7-4d97-8deb-2cd8f7181aa0", ResourceVersion:"1157", Generation:0, CreationTimestamp:time.Date(2024, time.December, 13, 2, 41, 15, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app":"nginx", "pod-template-hash":"6d5f899847", "projectcalico.org/namespace":"default", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"default"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172.24.4.241", ContainerID:"f65a4a221942baeae7ec44e3c5c94868a4c59c70eb11907f41d3eb7c34f45808", Pod:"nginx-deployment-6d5f899847-lblsv", Endpoint:"eth0", ServiceAccountName:"default", IPNetworks:[]string{"192.168.18.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.default", "ksa.default.default"}, InterfaceName:"cali5ce7bd04b58", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Dec 13 02:41:52.500516 containerd[1466]: 2024-12-13 02:41:52.446 [INFO][3219] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19" Dec 13 02:41:52.500516 containerd[1466]: 2024-12-13 02:41:52.446 [INFO][3219] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19" iface="eth0" netns="" Dec 13 02:41:52.500516 containerd[1466]: 2024-12-13 02:41:52.446 [INFO][3219] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19" Dec 13 02:41:52.500516 containerd[1466]: 2024-12-13 02:41:52.446 [INFO][3219] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19" Dec 13 02:41:52.500516 containerd[1466]: 2024-12-13 02:41:52.475 [INFO][3226] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19" HandleID="k8s-pod-network.3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19" Workload="172.24.4.241-k8s-nginx--deployment--6d5f899847--lblsv-eth0" Dec 13 02:41:52.500516 containerd[1466]: 2024-12-13 02:41:52.475 [INFO][3226] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Dec 13 02:41:52.500516 containerd[1466]: 2024-12-13 02:41:52.476 [INFO][3226] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Dec 13 02:41:52.500516 containerd[1466]: 2024-12-13 02:41:52.494 [WARNING][3226] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19" HandleID="k8s-pod-network.3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19" Workload="172.24.4.241-k8s-nginx--deployment--6d5f899847--lblsv-eth0" Dec 13 02:41:52.500516 containerd[1466]: 2024-12-13 02:41:52.494 [INFO][3226] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19" HandleID="k8s-pod-network.3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19" Workload="172.24.4.241-k8s-nginx--deployment--6d5f899847--lblsv-eth0" Dec 13 02:41:52.500516 containerd[1466]: 2024-12-13 02:41:52.497 [INFO][3226] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Dec 13 02:41:52.500516 containerd[1466]: 2024-12-13 02:41:52.499 [INFO][3219] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19" Dec 13 02:41:52.501163 containerd[1466]: time="2024-12-13T02:41:52.500555340Z" level=info msg="TearDown network for sandbox \"3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19\" successfully" Dec 13 02:41:52.501163 containerd[1466]: time="2024-12-13T02:41:52.500583884Z" level=info msg="StopPodSandbox for \"3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19\" returns successfully" Dec 13 02:41:52.527788 containerd[1466]: time="2024-12-13T02:41:52.527342189Z" level=info msg="RemovePodSandbox for \"3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19\"" Dec 13 02:41:52.527788 containerd[1466]: time="2024-12-13T02:41:52.527398224Z" level=info msg="Forcibly stopping sandbox \"3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19\"" Dec 13 02:41:52.674767 containerd[1466]: 2024-12-13 02:41:52.614 [WARNING][3245] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172.24.4.241-k8s-nginx--deployment--6d5f899847--lblsv-eth0", GenerateName:"nginx-deployment-6d5f899847-", Namespace:"default", SelfLink:"", UID:"1c2ae6bf-c9e7-4d97-8deb-2cd8f7181aa0", ResourceVersion:"1157", Generation:0, CreationTimestamp:time.Date(2024, time.December, 13, 2, 41, 15, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app":"nginx", "pod-template-hash":"6d5f899847", "projectcalico.org/namespace":"default", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"default"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172.24.4.241", ContainerID:"f65a4a221942baeae7ec44e3c5c94868a4c59c70eb11907f41d3eb7c34f45808", Pod:"nginx-deployment-6d5f899847-lblsv", Endpoint:"eth0", ServiceAccountName:"default", IPNetworks:[]string{"192.168.18.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.default", "ksa.default.default"}, InterfaceName:"cali5ce7bd04b58", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Dec 13 02:41:52.674767 containerd[1466]: 2024-12-13 02:41:52.614 [INFO][3245] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19" Dec 13 02:41:52.674767 containerd[1466]: 2024-12-13 02:41:52.614 [INFO][3245] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19" iface="eth0" netns="" Dec 13 02:41:52.674767 containerd[1466]: 2024-12-13 02:41:52.614 [INFO][3245] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19" Dec 13 02:41:52.674767 containerd[1466]: 2024-12-13 02:41:52.614 [INFO][3245] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19" Dec 13 02:41:52.674767 containerd[1466]: 2024-12-13 02:41:52.652 [INFO][3251] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19" HandleID="k8s-pod-network.3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19" Workload="172.24.4.241-k8s-nginx--deployment--6d5f899847--lblsv-eth0" Dec 13 02:41:52.674767 containerd[1466]: 2024-12-13 02:41:52.652 [INFO][3251] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Dec 13 02:41:52.674767 containerd[1466]: 2024-12-13 02:41:52.652 [INFO][3251] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Dec 13 02:41:52.674767 containerd[1466]: 2024-12-13 02:41:52.669 [WARNING][3251] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19" HandleID="k8s-pod-network.3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19" Workload="172.24.4.241-k8s-nginx--deployment--6d5f899847--lblsv-eth0" Dec 13 02:41:52.674767 containerd[1466]: 2024-12-13 02:41:52.669 [INFO][3251] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19" HandleID="k8s-pod-network.3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19" Workload="172.24.4.241-k8s-nginx--deployment--6d5f899847--lblsv-eth0" Dec 13 02:41:52.674767 containerd[1466]: 2024-12-13 02:41:52.672 [INFO][3251] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Dec 13 02:41:52.674767 containerd[1466]: 2024-12-13 02:41:52.673 [INFO][3245] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19" Dec 13 02:41:52.674767 containerd[1466]: time="2024-12-13T02:41:52.674683503Z" level=info msg="TearDown network for sandbox \"3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19\" successfully" Dec 13 02:41:52.755580 containerd[1466]: time="2024-12-13T02:41:52.755507376Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Dec 13 02:41:52.756091 containerd[1466]: time="2024-12-13T02:41:52.755604287Z" level=info msg="RemovePodSandbox \"3c802c07a17f7455b98ba7d9dde51fb4949dffa9634cd8b6eb4ef3c287a08c19\" returns successfully" Dec 13 02:41:52.756369 containerd[1466]: time="2024-12-13T02:41:52.756290179Z" level=info msg="StopPodSandbox for \"7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c\"" Dec 13 02:41:52.803576 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount45186563.mount: Deactivated successfully. Dec 13 02:41:52.935206 containerd[1466]: 2024-12-13 02:41:52.844 [WARNING][3271] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172.24.4.241-k8s-csi--node--driver--mn4q7-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"ed687165-b432-4311-a017-bee89dd86d6d", ResourceVersion:"1170", Generation:0, CreationTimestamp:time.Date(2024, time.December, 13, 2, 40, 52, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"55b695c467", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172.24.4.241", ContainerID:"54e65b34c2462000e6bc32fc82fa534166e6fd83d370edd3864a4686e91b65c8", Pod:"csi-node-driver-mn4q7", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.18.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali2ea430d9eb4", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Dec 13 02:41:52.935206 containerd[1466]: 2024-12-13 02:41:52.844 [INFO][3271] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c" Dec 13 02:41:52.935206 containerd[1466]: 2024-12-13 02:41:52.845 [INFO][3271] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c" iface="eth0" netns="" Dec 13 02:41:52.935206 containerd[1466]: 2024-12-13 02:41:52.845 [INFO][3271] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c" Dec 13 02:41:52.935206 containerd[1466]: 2024-12-13 02:41:52.845 [INFO][3271] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c" Dec 13 02:41:52.935206 containerd[1466]: 2024-12-13 02:41:52.909 [INFO][3278] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c" HandleID="k8s-pod-network.7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c" Workload="172.24.4.241-k8s-csi--node--driver--mn4q7-eth0" Dec 13 02:41:52.935206 containerd[1466]: 2024-12-13 02:41:52.909 [INFO][3278] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Dec 13 02:41:52.935206 containerd[1466]: 2024-12-13 02:41:52.909 [INFO][3278] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Dec 13 02:41:52.935206 containerd[1466]: 2024-12-13 02:41:52.926 [WARNING][3278] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c" HandleID="k8s-pod-network.7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c" Workload="172.24.4.241-k8s-csi--node--driver--mn4q7-eth0" Dec 13 02:41:52.935206 containerd[1466]: 2024-12-13 02:41:52.926 [INFO][3278] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c" HandleID="k8s-pod-network.7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c" Workload="172.24.4.241-k8s-csi--node--driver--mn4q7-eth0" Dec 13 02:41:52.935206 containerd[1466]: 2024-12-13 02:41:52.929 [INFO][3278] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Dec 13 02:41:52.935206 containerd[1466]: 2024-12-13 02:41:52.932 [INFO][3271] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c" Dec 13 02:41:52.937372 containerd[1466]: time="2024-12-13T02:41:52.936140486Z" level=info msg="TearDown network for sandbox \"7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c\" successfully" Dec 13 02:41:52.937372 containerd[1466]: time="2024-12-13T02:41:52.936195039Z" level=info msg="StopPodSandbox for \"7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c\" returns successfully" Dec 13 02:41:52.937372 containerd[1466]: time="2024-12-13T02:41:52.937356420Z" level=info msg="RemovePodSandbox for \"7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c\"" Dec 13 02:41:52.937567 containerd[1466]: time="2024-12-13T02:41:52.937400311Z" level=info msg="Forcibly stopping sandbox \"7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c\"" Dec 13 02:41:53.054190 containerd[1466]: 2024-12-13 02:41:52.995 [WARNING][3301] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172.24.4.241-k8s-csi--node--driver--mn4q7-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"ed687165-b432-4311-a017-bee89dd86d6d", ResourceVersion:"1170", Generation:0, CreationTimestamp:time.Date(2024, time.December, 13, 2, 40, 52, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"55b695c467", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172.24.4.241", ContainerID:"54e65b34c2462000e6bc32fc82fa534166e6fd83d370edd3864a4686e91b65c8", Pod:"csi-node-driver-mn4q7", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.18.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali2ea430d9eb4", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Dec 13 02:41:53.054190 containerd[1466]: 2024-12-13 02:41:52.996 [INFO][3301] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c" Dec 13 02:41:53.054190 containerd[1466]: 2024-12-13 02:41:52.996 [INFO][3301] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c" iface="eth0" netns="" Dec 13 02:41:53.054190 containerd[1466]: 2024-12-13 02:41:52.996 [INFO][3301] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c" Dec 13 02:41:53.054190 containerd[1466]: 2024-12-13 02:41:52.996 [INFO][3301] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c" Dec 13 02:41:53.054190 containerd[1466]: 2024-12-13 02:41:53.028 [INFO][3307] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c" HandleID="k8s-pod-network.7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c" Workload="172.24.4.241-k8s-csi--node--driver--mn4q7-eth0" Dec 13 02:41:53.054190 containerd[1466]: 2024-12-13 02:41:53.028 [INFO][3307] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Dec 13 02:41:53.054190 containerd[1466]: 2024-12-13 02:41:53.028 [INFO][3307] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Dec 13 02:41:53.054190 containerd[1466]: 2024-12-13 02:41:53.047 [WARNING][3307] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c" HandleID="k8s-pod-network.7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c" Workload="172.24.4.241-k8s-csi--node--driver--mn4q7-eth0" Dec 13 02:41:53.054190 containerd[1466]: 2024-12-13 02:41:53.047 [INFO][3307] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c" HandleID="k8s-pod-network.7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c" Workload="172.24.4.241-k8s-csi--node--driver--mn4q7-eth0" Dec 13 02:41:53.054190 containerd[1466]: 2024-12-13 02:41:53.050 [INFO][3307] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Dec 13 02:41:53.054190 containerd[1466]: 2024-12-13 02:41:53.051 [INFO][3301] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c" Dec 13 02:41:53.054190 containerd[1466]: time="2024-12-13T02:41:53.054093534Z" level=info msg="TearDown network for sandbox \"7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c\" successfully" Dec 13 02:41:53.058058 containerd[1466]: time="2024-12-13T02:41:53.057865554Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Dec 13 02:41:53.058058 containerd[1466]: time="2024-12-13T02:41:53.057939102Z" level=info msg="RemovePodSandbox \"7b224caca01d5db38c80d14654d24bedd2e95d0ae33142c6b55f3f311ef4759c\" returns successfully" Dec 13 02:41:53.345790 kubelet[1830]: E1213 02:41:53.345348 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:54.346047 kubelet[1830]: E1213 02:41:54.346003 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:55.347350 kubelet[1830]: E1213 02:41:55.347222 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:56.347527 kubelet[1830]: E1213 02:41:56.347448 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:56.386598 containerd[1466]: time="2024-12-13T02:41:56.386481057Z" level=info msg="ImageCreate event name:\"registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 02:41:56.391935 containerd[1466]: time="2024-12-13T02:41:56.391731312Z" level=info msg="stop pulling image registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8: active requests=0, bytes read=91039414" Dec 13 02:41:56.398161 containerd[1466]: time="2024-12-13T02:41:56.398063731Z" level=info msg="ImageCreate event name:\"sha256:fd0b16f70b66b72bcb2f91d556fa33eba02729c44ffc5f2c16130e7f9fbed3c4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 02:41:56.410001 containerd[1466]: time="2024-12-13T02:41:56.409142131Z" level=info msg="ImageCreate event name:\"registry.k8s.io/sig-storage/nfs-provisioner@sha256:c825f3d5e28bde099bd7a3daace28772d412c9157ad47fa752a9ad0baafc118d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 02:41:56.415534 containerd[1466]: time="2024-12-13T02:41:56.415425197Z" level=info msg="Pulled image \"registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8\" with image id \"sha256:fd0b16f70b66b72bcb2f91d556fa33eba02729c44ffc5f2c16130e7f9fbed3c4\", repo tag \"registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8\", repo digest \"registry.k8s.io/sig-storage/nfs-provisioner@sha256:c825f3d5e28bde099bd7a3daace28772d412c9157ad47fa752a9ad0baafc118d\", size \"91036984\" in 7.921899241s" Dec 13 02:41:56.415534 containerd[1466]: time="2024-12-13T02:41:56.415525715Z" level=info msg="PullImage \"registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8\" returns image reference \"sha256:fd0b16f70b66b72bcb2f91d556fa33eba02729c44ffc5f2c16130e7f9fbed3c4\"" Dec 13 02:41:56.424165 containerd[1466]: time="2024-12-13T02:41:56.423880174Z" level=info msg="CreateContainer within sandbox \"ac4e4815181bfb0eb8425e35992b7bf8fd8759c2bac2353395a10a42bedf4da2\" for container &ContainerMetadata{Name:nfs-server-provisioner,Attempt:0,}" Dec 13 02:41:56.461930 containerd[1466]: time="2024-12-13T02:41:56.461810523Z" level=info msg="CreateContainer within sandbox \"ac4e4815181bfb0eb8425e35992b7bf8fd8759c2bac2353395a10a42bedf4da2\" for &ContainerMetadata{Name:nfs-server-provisioner,Attempt:0,} returns container id \"1243453bb6c6dd47dc0404851c1ee30e19a508779957cec057565ab45e197153\"" Dec 13 02:41:56.465846 containerd[1466]: time="2024-12-13T02:41:56.462742284Z" level=info msg="StartContainer for \"1243453bb6c6dd47dc0404851c1ee30e19a508779957cec057565ab45e197153\"" Dec 13 02:41:56.525257 systemd[1]: Started cri-containerd-1243453bb6c6dd47dc0404851c1ee30e19a508779957cec057565ab45e197153.scope - libcontainer container 1243453bb6c6dd47dc0404851c1ee30e19a508779957cec057565ab45e197153. Dec 13 02:41:56.565251 containerd[1466]: time="2024-12-13T02:41:56.565193162Z" level=info msg="StartContainer for \"1243453bb6c6dd47dc0404851c1ee30e19a508779957cec057565ab45e197153\" returns successfully" Dec 13 02:41:57.208041 kubelet[1830]: I1213 02:41:57.207883 1830 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="default/nfs-server-provisioner-0" podStartSLOduration=2.284388071 podStartE2EDuration="10.207789409s" podCreationTimestamp="2024-12-13 02:41:47 +0000 UTC" firstStartedPulling="2024-12-13 02:41:48.492621176 +0000 UTC m=+57.309656967" lastFinishedPulling="2024-12-13 02:41:56.416022464 +0000 UTC m=+65.233058305" observedRunningTime="2024-12-13 02:41:57.207322687 +0000 UTC m=+66.024358518" watchObservedRunningTime="2024-12-13 02:41:57.207789409 +0000 UTC m=+66.024825240" Dec 13 02:41:57.350088 kubelet[1830]: E1213 02:41:57.349458 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:58.350790 kubelet[1830]: E1213 02:41:58.350650 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:41:59.351199 kubelet[1830]: E1213 02:41:59.351097 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:00.352202 kubelet[1830]: E1213 02:42:00.352097 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:01.352448 kubelet[1830]: E1213 02:42:01.352364 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:02.352699 kubelet[1830]: E1213 02:42:02.352587 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:03.353850 kubelet[1830]: E1213 02:42:03.353757 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:04.354186 kubelet[1830]: E1213 02:42:04.354063 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:05.355035 kubelet[1830]: E1213 02:42:05.354831 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:06.355821 kubelet[1830]: E1213 02:42:06.355727 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:07.356801 kubelet[1830]: E1213 02:42:07.356707 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:08.357585 kubelet[1830]: E1213 02:42:08.357454 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:09.358057 kubelet[1830]: E1213 02:42:09.357881 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:10.358476 kubelet[1830]: E1213 02:42:10.358359 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:11.359051 kubelet[1830]: E1213 02:42:11.358958 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:12.289157 kubelet[1830]: E1213 02:42:12.289074 1830 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:12.359697 kubelet[1830]: E1213 02:42:12.359599 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:13.360559 kubelet[1830]: E1213 02:42:13.360453 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:14.360972 kubelet[1830]: E1213 02:42:14.360844 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:15.361790 kubelet[1830]: E1213 02:42:15.361663 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:16.362543 kubelet[1830]: E1213 02:42:16.362440 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:17.363617 kubelet[1830]: E1213 02:42:17.363553 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:18.364483 kubelet[1830]: E1213 02:42:18.364358 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:19.364738 kubelet[1830]: E1213 02:42:19.364635 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:20.365030 kubelet[1830]: E1213 02:42:20.364886 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:21.365655 kubelet[1830]: E1213 02:42:21.365586 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:21.953314 kubelet[1830]: I1213 02:42:21.953226 1830 topology_manager.go:215] "Topology Admit Handler" podUID="f43148c4-a0df-4f20-8545-633e063ae8de" podNamespace="default" podName="test-pod-1" Dec 13 02:42:21.968271 systemd[1]: Created slice kubepods-besteffort-podf43148c4_a0df_4f20_8545_633e063ae8de.slice - libcontainer container kubepods-besteffort-podf43148c4_a0df_4f20_8545_633e063ae8de.slice. Dec 13 02:42:22.149503 kubelet[1830]: I1213 02:42:22.149340 1830 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8trsb\" (UniqueName: \"kubernetes.io/projected/f43148c4-a0df-4f20-8545-633e063ae8de-kube-api-access-8trsb\") pod \"test-pod-1\" (UID: \"f43148c4-a0df-4f20-8545-633e063ae8de\") " pod="default/test-pod-1" Dec 13 02:42:22.149503 kubelet[1830]: I1213 02:42:22.149497 1830 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ac99bb88-9207-4b22-9581-5ac34f309358\" (UniqueName: \"kubernetes.io/nfs/f43148c4-a0df-4f20-8545-633e063ae8de-pvc-ac99bb88-9207-4b22-9581-5ac34f309358\") pod \"test-pod-1\" (UID: \"f43148c4-a0df-4f20-8545-633e063ae8de\") " pod="default/test-pod-1" Dec 13 02:42:22.340233 kernel: FS-Cache: Loaded Dec 13 02:42:22.370257 kubelet[1830]: E1213 02:42:22.366482 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:22.432558 kernel: RPC: Registered named UNIX socket transport module. Dec 13 02:42:22.432763 kernel: RPC: Registered udp transport module. Dec 13 02:42:22.432791 kernel: RPC: Registered tcp transport module. Dec 13 02:42:22.432815 kernel: RPC: Registered tcp-with-tls transport module. Dec 13 02:42:22.432835 kernel: RPC: Registered tcp NFSv4.1 backchannel transport module. Dec 13 02:42:22.800001 kernel: NFS: Registering the id_resolver key type Dec 13 02:42:22.800116 kernel: Key type id_resolver registered Dec 13 02:42:22.801007 kernel: Key type id_legacy registered Dec 13 02:42:22.853987 nfsidmap[3456]: nss_getpwnam: name 'root@nfs-server-provisioner.default.svc.cluster.local' does not map into domain 'novalocal' Dec 13 02:42:22.864869 nfsidmap[3457]: nss_name_to_gid: name 'root@nfs-server-provisioner.default.svc.cluster.local' does not map into domain 'novalocal' Dec 13 02:42:23.177361 containerd[1466]: time="2024-12-13T02:42:23.177089444Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:test-pod-1,Uid:f43148c4-a0df-4f20-8545-633e063ae8de,Namespace:default,Attempt:0,}" Dec 13 02:42:23.367031 kubelet[1830]: E1213 02:42:23.366795 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:23.718088 systemd-networkd[1371]: cali5ec59c6bf6e: Link UP Dec 13 02:42:23.718588 systemd-networkd[1371]: cali5ec59c6bf6e: Gained carrier Dec 13 02:42:23.739275 containerd[1466]: 2024-12-13 02:42:23.508 [INFO][3460] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {172.24.4.241-k8s-test--pod--1-eth0 default f43148c4-a0df-4f20-8545-633e063ae8de 1320 0 2024-12-13 02:41:52 +0000 UTC map[projectcalico.org/namespace:default projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:default] map[] [] [] []} {k8s 172.24.4.241 test-pod-1 eth0 default [] [] [kns.default ksa.default.default] cali5ec59c6bf6e [] []}} ContainerID="18444eb05f148b55db98ec81835f491af649cc7bde5a0d6c6e9f2ed48c768c13" Namespace="default" Pod="test-pod-1" WorkloadEndpoint="172.24.4.241-k8s-test--pod--1-" Dec 13 02:42:23.739275 containerd[1466]: 2024-12-13 02:42:23.508 [INFO][3460] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="18444eb05f148b55db98ec81835f491af649cc7bde5a0d6c6e9f2ed48c768c13" Namespace="default" Pod="test-pod-1" WorkloadEndpoint="172.24.4.241-k8s-test--pod--1-eth0" Dec 13 02:42:23.739275 containerd[1466]: 2024-12-13 02:42:23.642 [INFO][3470] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="18444eb05f148b55db98ec81835f491af649cc7bde5a0d6c6e9f2ed48c768c13" HandleID="k8s-pod-network.18444eb05f148b55db98ec81835f491af649cc7bde5a0d6c6e9f2ed48c768c13" Workload="172.24.4.241-k8s-test--pod--1-eth0" Dec 13 02:42:23.739275 containerd[1466]: 2024-12-13 02:42:23.657 [INFO][3470] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="18444eb05f148b55db98ec81835f491af649cc7bde5a0d6c6e9f2ed48c768c13" HandleID="k8s-pod-network.18444eb05f148b55db98ec81835f491af649cc7bde5a0d6c6e9f2ed48c768c13" Workload="172.24.4.241-k8s-test--pod--1-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000336c80), Attrs:map[string]string{"namespace":"default", "node":"172.24.4.241", "pod":"test-pod-1", "timestamp":"2024-12-13 02:42:23.642887668 +0000 UTC"}, Hostname:"172.24.4.241", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Dec 13 02:42:23.739275 containerd[1466]: 2024-12-13 02:42:23.657 [INFO][3470] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Dec 13 02:42:23.739275 containerd[1466]: 2024-12-13 02:42:23.658 [INFO][3470] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Dec 13 02:42:23.739275 containerd[1466]: 2024-12-13 02:42:23.658 [INFO][3470] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host '172.24.4.241' Dec 13 02:42:23.739275 containerd[1466]: 2024-12-13 02:42:23.661 [INFO][3470] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.18444eb05f148b55db98ec81835f491af649cc7bde5a0d6c6e9f2ed48c768c13" host="172.24.4.241" Dec 13 02:42:23.739275 containerd[1466]: 2024-12-13 02:42:23.674 [INFO][3470] ipam/ipam.go 372: Looking up existing affinities for host host="172.24.4.241" Dec 13 02:42:23.739275 containerd[1466]: 2024-12-13 02:42:23.680 [INFO][3470] ipam/ipam.go 489: Trying affinity for 192.168.18.128/26 host="172.24.4.241" Dec 13 02:42:23.739275 containerd[1466]: 2024-12-13 02:42:23.683 [INFO][3470] ipam/ipam.go 155: Attempting to load block cidr=192.168.18.128/26 host="172.24.4.241" Dec 13 02:42:23.739275 containerd[1466]: 2024-12-13 02:42:23.687 [INFO][3470] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.18.128/26 host="172.24.4.241" Dec 13 02:42:23.739275 containerd[1466]: 2024-12-13 02:42:23.687 [INFO][3470] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.18.128/26 handle="k8s-pod-network.18444eb05f148b55db98ec81835f491af649cc7bde5a0d6c6e9f2ed48c768c13" host="172.24.4.241" Dec 13 02:42:23.739275 containerd[1466]: 2024-12-13 02:42:23.690 [INFO][3470] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.18444eb05f148b55db98ec81835f491af649cc7bde5a0d6c6e9f2ed48c768c13 Dec 13 02:42:23.739275 containerd[1466]: 2024-12-13 02:42:23.698 [INFO][3470] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.18.128/26 handle="k8s-pod-network.18444eb05f148b55db98ec81835f491af649cc7bde5a0d6c6e9f2ed48c768c13" host="172.24.4.241" Dec 13 02:42:23.739275 containerd[1466]: 2024-12-13 02:42:23.710 [INFO][3470] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.18.132/26] block=192.168.18.128/26 handle="k8s-pod-network.18444eb05f148b55db98ec81835f491af649cc7bde5a0d6c6e9f2ed48c768c13" host="172.24.4.241" Dec 13 02:42:23.739275 containerd[1466]: 2024-12-13 02:42:23.710 [INFO][3470] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.18.132/26] handle="k8s-pod-network.18444eb05f148b55db98ec81835f491af649cc7bde5a0d6c6e9f2ed48c768c13" host="172.24.4.241" Dec 13 02:42:23.739275 containerd[1466]: 2024-12-13 02:42:23.710 [INFO][3470] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Dec 13 02:42:23.739275 containerd[1466]: 2024-12-13 02:42:23.710 [INFO][3470] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.18.132/26] IPv6=[] ContainerID="18444eb05f148b55db98ec81835f491af649cc7bde5a0d6c6e9f2ed48c768c13" HandleID="k8s-pod-network.18444eb05f148b55db98ec81835f491af649cc7bde5a0d6c6e9f2ed48c768c13" Workload="172.24.4.241-k8s-test--pod--1-eth0" Dec 13 02:42:23.739275 containerd[1466]: 2024-12-13 02:42:23.713 [INFO][3460] cni-plugin/k8s.go 386: Populated endpoint ContainerID="18444eb05f148b55db98ec81835f491af649cc7bde5a0d6c6e9f2ed48c768c13" Namespace="default" Pod="test-pod-1" WorkloadEndpoint="172.24.4.241-k8s-test--pod--1-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172.24.4.241-k8s-test--pod--1-eth0", GenerateName:"", Namespace:"default", SelfLink:"", UID:"f43148c4-a0df-4f20-8545-633e063ae8de", ResourceVersion:"1320", Generation:0, CreationTimestamp:time.Date(2024, time.December, 13, 2, 41, 52, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"projectcalico.org/namespace":"default", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"default"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172.24.4.241", ContainerID:"", Pod:"test-pod-1", Endpoint:"eth0", ServiceAccountName:"default", IPNetworks:[]string{"192.168.18.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.default", "ksa.default.default"}, InterfaceName:"cali5ec59c6bf6e", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Dec 13 02:42:23.740006 containerd[1466]: 2024-12-13 02:42:23.713 [INFO][3460] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.18.132/32] ContainerID="18444eb05f148b55db98ec81835f491af649cc7bde5a0d6c6e9f2ed48c768c13" Namespace="default" Pod="test-pod-1" WorkloadEndpoint="172.24.4.241-k8s-test--pod--1-eth0" Dec 13 02:42:23.740006 containerd[1466]: 2024-12-13 02:42:23.713 [INFO][3460] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali5ec59c6bf6e ContainerID="18444eb05f148b55db98ec81835f491af649cc7bde5a0d6c6e9f2ed48c768c13" Namespace="default" Pod="test-pod-1" WorkloadEndpoint="172.24.4.241-k8s-test--pod--1-eth0" Dec 13 02:42:23.740006 containerd[1466]: 2024-12-13 02:42:23.719 [INFO][3460] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="18444eb05f148b55db98ec81835f491af649cc7bde5a0d6c6e9f2ed48c768c13" Namespace="default" Pod="test-pod-1" WorkloadEndpoint="172.24.4.241-k8s-test--pod--1-eth0" Dec 13 02:42:23.740006 containerd[1466]: 2024-12-13 02:42:23.719 [INFO][3460] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="18444eb05f148b55db98ec81835f491af649cc7bde5a0d6c6e9f2ed48c768c13" Namespace="default" Pod="test-pod-1" WorkloadEndpoint="172.24.4.241-k8s-test--pod--1-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172.24.4.241-k8s-test--pod--1-eth0", GenerateName:"", Namespace:"default", SelfLink:"", UID:"f43148c4-a0df-4f20-8545-633e063ae8de", ResourceVersion:"1320", Generation:0, CreationTimestamp:time.Date(2024, time.December, 13, 2, 41, 52, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"projectcalico.org/namespace":"default", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"default"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172.24.4.241", ContainerID:"18444eb05f148b55db98ec81835f491af649cc7bde5a0d6c6e9f2ed48c768c13", Pod:"test-pod-1", Endpoint:"eth0", ServiceAccountName:"default", IPNetworks:[]string{"192.168.18.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.default", "ksa.default.default"}, InterfaceName:"cali5ec59c6bf6e", MAC:"3e:14:6e:fc:19:fa", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Dec 13 02:42:23.740006 containerd[1466]: 2024-12-13 02:42:23.729 [INFO][3460] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="18444eb05f148b55db98ec81835f491af649cc7bde5a0d6c6e9f2ed48c768c13" Namespace="default" Pod="test-pod-1" WorkloadEndpoint="172.24.4.241-k8s-test--pod--1-eth0" Dec 13 02:42:23.781623 containerd[1466]: time="2024-12-13T02:42:23.781444683Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 02:42:23.781623 containerd[1466]: time="2024-12-13T02:42:23.781543370Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 02:42:23.781623 containerd[1466]: time="2024-12-13T02:42:23.781565091Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 02:42:23.781623 containerd[1466]: time="2024-12-13T02:42:23.781678575Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 02:42:23.821083 systemd[1]: Started cri-containerd-18444eb05f148b55db98ec81835f491af649cc7bde5a0d6c6e9f2ed48c768c13.scope - libcontainer container 18444eb05f148b55db98ec81835f491af649cc7bde5a0d6c6e9f2ed48c768c13. Dec 13 02:42:23.873142 containerd[1466]: time="2024-12-13T02:42:23.873019924Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:test-pod-1,Uid:f43148c4-a0df-4f20-8545-633e063ae8de,Namespace:default,Attempt:0,} returns sandbox id \"18444eb05f148b55db98ec81835f491af649cc7bde5a0d6c6e9f2ed48c768c13\"" Dec 13 02:42:23.875991 containerd[1466]: time="2024-12-13T02:42:23.875946788Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\"" Dec 13 02:42:24.295057 containerd[1466]: time="2024-12-13T02:42:24.294018259Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/nginx:latest\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 02:42:24.297556 containerd[1466]: time="2024-12-13T02:42:24.297457740Z" level=info msg="stop pulling image ghcr.io/flatcar/nginx:latest: active requests=0, bytes read=61" Dec 13 02:42:24.300958 containerd[1466]: time="2024-12-13T02:42:24.300869948Z" level=info msg="Pulled image \"ghcr.io/flatcar/nginx:latest\" with image id \"sha256:fa0a8cea5e76ad962111c39c85bb312edaf5b89eccd8f404eeea66c9759641e3\", repo tag \"ghcr.io/flatcar/nginx:latest\", repo digest \"ghcr.io/flatcar/nginx@sha256:e04edf30a4ea4c5a4107110797c72d3ee8a654415f00acd4019be17218afd9a1\", size \"71035905\" in 424.875982ms" Dec 13 02:42:24.300958 containerd[1466]: time="2024-12-13T02:42:24.300929581Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\" returns image reference \"sha256:fa0a8cea5e76ad962111c39c85bb312edaf5b89eccd8f404eeea66c9759641e3\"" Dec 13 02:42:24.302665 containerd[1466]: time="2024-12-13T02:42:24.302588792Z" level=info msg="CreateContainer within sandbox \"18444eb05f148b55db98ec81835f491af649cc7bde5a0d6c6e9f2ed48c768c13\" for container &ContainerMetadata{Name:test,Attempt:0,}" Dec 13 02:42:24.351427 containerd[1466]: time="2024-12-13T02:42:24.351285432Z" level=info msg="CreateContainer within sandbox \"18444eb05f148b55db98ec81835f491af649cc7bde5a0d6c6e9f2ed48c768c13\" for &ContainerMetadata{Name:test,Attempt:0,} returns container id \"8ada45ebb335c86b2de60e4c6e09bf9e0485e2c681f9a3dab4568c63efab9850\"" Dec 13 02:42:24.352684 containerd[1466]: time="2024-12-13T02:42:24.352586737Z" level=info msg="StartContainer for \"8ada45ebb335c86b2de60e4c6e09bf9e0485e2c681f9a3dab4568c63efab9850\"" Dec 13 02:42:24.368876 kubelet[1830]: E1213 02:42:24.368785 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:24.403227 systemd[1]: Started cri-containerd-8ada45ebb335c86b2de60e4c6e09bf9e0485e2c681f9a3dab4568c63efab9850.scope - libcontainer container 8ada45ebb335c86b2de60e4c6e09bf9e0485e2c681f9a3dab4568c63efab9850. Dec 13 02:42:24.465499 containerd[1466]: time="2024-12-13T02:42:24.464657694Z" level=info msg="StartContainer for \"8ada45ebb335c86b2de60e4c6e09bf9e0485e2c681f9a3dab4568c63efab9850\" returns successfully" Dec 13 02:42:25.369857 kubelet[1830]: E1213 02:42:25.369753 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:25.500348 systemd-networkd[1371]: cali5ec59c6bf6e: Gained IPv6LL Dec 13 02:42:25.813715 kubelet[1830]: I1213 02:42:25.813614 1830 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="default/test-pod-1" podStartSLOduration=33.386908395 podStartE2EDuration="33.812862451s" podCreationTimestamp="2024-12-13 02:41:52 +0000 UTC" firstStartedPulling="2024-12-13 02:42:23.875215618 +0000 UTC m=+92.692251409" lastFinishedPulling="2024-12-13 02:42:24.301169674 +0000 UTC m=+93.118205465" observedRunningTime="2024-12-13 02:42:25.811861773 +0000 UTC m=+94.628897604" watchObservedRunningTime="2024-12-13 02:42:25.812862451 +0000 UTC m=+94.629898372" Dec 13 02:42:26.370616 kubelet[1830]: E1213 02:42:26.370523 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:27.371682 kubelet[1830]: E1213 02:42:27.371473 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:28.372246 kubelet[1830]: E1213 02:42:28.372141 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:29.373309 kubelet[1830]: E1213 02:42:29.373208 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:30.374407 kubelet[1830]: E1213 02:42:30.374309 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:31.375458 kubelet[1830]: E1213 02:42:31.375405 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:32.289837 kubelet[1830]: E1213 02:42:32.289752 1830 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:32.376527 kubelet[1830]: E1213 02:42:32.376377 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:33.377662 kubelet[1830]: E1213 02:42:33.377556 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:34.377955 kubelet[1830]: E1213 02:42:34.377766 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:35.378594 kubelet[1830]: E1213 02:42:35.378490 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:36.379298 kubelet[1830]: E1213 02:42:36.379182 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:37.380098 kubelet[1830]: E1213 02:42:37.380011 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 02:42:38.380642 kubelet[1830]: E1213 02:42:38.380588 1830 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests"