Dec 13 01:27:33.958541 kernel: Linux version 6.6.65-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Thu Dec 12 23:15:00 -00 2024 Dec 13 01:27:33.958581 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=2fdbba50b59d8c8a9877a81151806ddc16f473fe99b9ba0d8825997d654583ff Dec 13 01:27:33.958596 kernel: BIOS-provided physical RAM map: Dec 13 01:27:33.958604 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable Dec 13 01:27:33.958612 kernel: BIOS-e820: [mem 0x0000000000100000-0x00000000007fffff] usable Dec 13 01:27:33.958620 kernel: BIOS-e820: [mem 0x0000000000800000-0x0000000000807fff] ACPI NVS Dec 13 01:27:33.958630 kernel: BIOS-e820: [mem 0x0000000000808000-0x000000000080afff] usable Dec 13 01:27:33.958650 kernel: BIOS-e820: [mem 0x000000000080b000-0x000000000080bfff] ACPI NVS Dec 13 01:27:33.958658 kernel: BIOS-e820: [mem 0x000000000080c000-0x000000000080ffff] usable Dec 13 01:27:33.958666 kernel: BIOS-e820: [mem 0x0000000000810000-0x00000000008fffff] ACPI NVS Dec 13 01:27:33.958682 kernel: BIOS-e820: [mem 0x0000000000900000-0x000000009c8eefff] usable Dec 13 01:27:33.958691 kernel: BIOS-e820: [mem 0x000000009c8ef000-0x000000009c9eefff] reserved Dec 13 01:27:33.958699 kernel: BIOS-e820: [mem 0x000000009c9ef000-0x000000009caeefff] type 20 Dec 13 01:27:33.958708 kernel: BIOS-e820: [mem 0x000000009caef000-0x000000009cb6efff] reserved Dec 13 01:27:33.958723 kernel: BIOS-e820: [mem 0x000000009cb6f000-0x000000009cb7efff] ACPI data Dec 13 01:27:33.958732 kernel: BIOS-e820: [mem 0x000000009cb7f000-0x000000009cbfefff] ACPI NVS Dec 13 01:27:33.958746 kernel: BIOS-e820: [mem 0x000000009cbff000-0x000000009cf3ffff] usable Dec 13 01:27:33.958755 kernel: BIOS-e820: [mem 0x000000009cf40000-0x000000009cf5ffff] reserved Dec 13 01:27:33.958764 kernel: BIOS-e820: [mem 0x000000009cf60000-0x000000009cffffff] ACPI NVS Dec 13 01:27:33.958773 kernel: BIOS-e820: [mem 0x00000000b0000000-0x00000000bfffffff] reserved Dec 13 01:27:33.958782 kernel: NX (Execute Disable) protection: active Dec 13 01:27:33.958792 kernel: APIC: Static calls initialized Dec 13 01:27:33.958801 kernel: efi: EFI v2.7 by EDK II Dec 13 01:27:33.958810 kernel: efi: SMBIOS=0x9c9ab000 ACPI=0x9cb7e000 ACPI 2.0=0x9cb7e014 MEMATTR=0x9b674118 Dec 13 01:27:33.958819 kernel: SMBIOS 2.8 present. Dec 13 01:27:33.958828 kernel: DMI: QEMU Standard PC (Q35 + ICH9, 2009), BIOS 0.0.0 02/06/2015 Dec 13 01:27:33.958837 kernel: Hypervisor detected: KVM Dec 13 01:27:33.958849 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Dec 13 01:27:33.958858 kernel: kvm-clock: using sched offset of 5660478622 cycles Dec 13 01:27:33.958868 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Dec 13 01:27:33.958878 kernel: tsc: Detected 2794.748 MHz processor Dec 13 01:27:33.958888 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Dec 13 01:27:33.958898 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Dec 13 01:27:33.958908 kernel: last_pfn = 0x9cf40 max_arch_pfn = 0x400000000 Dec 13 01:27:33.958917 kernel: MTRR map: 4 entries (2 fixed + 2 variable; max 18), built from 8 variable MTRRs Dec 13 01:27:33.958926 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Dec 13 01:27:33.958938 kernel: Using GB pages for direct mapping Dec 13 01:27:33.958948 kernel: Secure boot disabled Dec 13 01:27:33.958957 kernel: ACPI: Early table checksum verification disabled Dec 13 01:27:33.958967 kernel: ACPI: RSDP 0x000000009CB7E014 000024 (v02 BOCHS ) Dec 13 01:27:33.958985 kernel: ACPI: XSDT 0x000000009CB7D0E8 000054 (v01 BOCHS BXPC 00000001 01000013) Dec 13 01:27:33.958995 kernel: ACPI: FACP 0x000000009CB79000 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 01:27:33.959005 kernel: ACPI: DSDT 0x000000009CB7A000 0021A8 (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 01:27:33.959018 kernel: ACPI: FACS 0x000000009CBDD000 000040 Dec 13 01:27:33.959045 kernel: ACPI: APIC 0x000000009CB78000 000090 (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 01:27:33.959055 kernel: ACPI: HPET 0x000000009CB77000 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 01:27:33.959065 kernel: ACPI: MCFG 0x000000009CB76000 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 01:27:33.959074 kernel: ACPI: WAET 0x000000009CB75000 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 01:27:33.959084 kernel: ACPI: BGRT 0x000000009CB74000 000038 (v01 INTEL EDK2 00000002 01000013) Dec 13 01:27:33.959094 kernel: ACPI: Reserving FACP table memory at [mem 0x9cb79000-0x9cb790f3] Dec 13 01:27:33.959110 kernel: ACPI: Reserving DSDT table memory at [mem 0x9cb7a000-0x9cb7c1a7] Dec 13 01:27:33.959120 kernel: ACPI: Reserving FACS table memory at [mem 0x9cbdd000-0x9cbdd03f] Dec 13 01:27:33.959129 kernel: ACPI: Reserving APIC table memory at [mem 0x9cb78000-0x9cb7808f] Dec 13 01:27:33.959139 kernel: ACPI: Reserving HPET table memory at [mem 0x9cb77000-0x9cb77037] Dec 13 01:27:33.959148 kernel: ACPI: Reserving MCFG table memory at [mem 0x9cb76000-0x9cb7603b] Dec 13 01:27:33.959158 kernel: ACPI: Reserving WAET table memory at [mem 0x9cb75000-0x9cb75027] Dec 13 01:27:33.959167 kernel: ACPI: Reserving BGRT table memory at [mem 0x9cb74000-0x9cb74037] Dec 13 01:27:33.959182 kernel: No NUMA configuration found Dec 13 01:27:33.959192 kernel: Faking a node at [mem 0x0000000000000000-0x000000009cf3ffff] Dec 13 01:27:33.959206 kernel: NODE_DATA(0) allocated [mem 0x9cea6000-0x9ceabfff] Dec 13 01:27:33.959216 kernel: Zone ranges: Dec 13 01:27:33.959226 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Dec 13 01:27:33.959236 kernel: DMA32 [mem 0x0000000001000000-0x000000009cf3ffff] Dec 13 01:27:33.959246 kernel: Normal empty Dec 13 01:27:33.959256 kernel: Movable zone start for each node Dec 13 01:27:33.959266 kernel: Early memory node ranges Dec 13 01:27:33.959275 kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff] Dec 13 01:27:33.959285 kernel: node 0: [mem 0x0000000000100000-0x00000000007fffff] Dec 13 01:27:33.959318 kernel: node 0: [mem 0x0000000000808000-0x000000000080afff] Dec 13 01:27:33.959363 kernel: node 0: [mem 0x000000000080c000-0x000000000080ffff] Dec 13 01:27:33.959378 kernel: node 0: [mem 0x0000000000900000-0x000000009c8eefff] Dec 13 01:27:33.959388 kernel: node 0: [mem 0x000000009cbff000-0x000000009cf3ffff] Dec 13 01:27:33.959401 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000009cf3ffff] Dec 13 01:27:33.959411 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Dec 13 01:27:33.959421 kernel: On node 0, zone DMA: 96 pages in unavailable ranges Dec 13 01:27:33.959440 kernel: On node 0, zone DMA: 8 pages in unavailable ranges Dec 13 01:27:33.959456 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Dec 13 01:27:33.959465 kernel: On node 0, zone DMA: 240 pages in unavailable ranges Dec 13 01:27:33.959483 kernel: On node 0, zone DMA32: 784 pages in unavailable ranges Dec 13 01:27:33.959493 kernel: On node 0, zone DMA32: 12480 pages in unavailable ranges Dec 13 01:27:33.959503 kernel: ACPI: PM-Timer IO Port: 0x608 Dec 13 01:27:33.959513 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Dec 13 01:27:33.959523 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Dec 13 01:27:33.959533 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Dec 13 01:27:33.959543 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Dec 13 01:27:33.959553 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Dec 13 01:27:33.959563 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Dec 13 01:27:33.959576 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Dec 13 01:27:33.959585 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Dec 13 01:27:33.959595 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Dec 13 01:27:33.959605 kernel: TSC deadline timer available Dec 13 01:27:33.959614 kernel: smpboot: Allowing 4 CPUs, 0 hotplug CPUs Dec 13 01:27:33.959625 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Dec 13 01:27:33.959635 kernel: kvm-guest: KVM setup pv remote TLB flush Dec 13 01:27:33.959653 kernel: kvm-guest: setup PV sched yield Dec 13 01:27:33.959663 kernel: [mem 0xc0000000-0xffffffff] available for PCI devices Dec 13 01:27:33.959676 kernel: Booting paravirtualized kernel on KVM Dec 13 01:27:33.959686 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Dec 13 01:27:33.959696 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:4 nr_cpu_ids:4 nr_node_ids:1 Dec 13 01:27:33.959706 kernel: percpu: Embedded 58 pages/cpu s197032 r8192 d32344 u524288 Dec 13 01:27:33.959716 kernel: pcpu-alloc: s197032 r8192 d32344 u524288 alloc=1*2097152 Dec 13 01:27:33.959725 kernel: pcpu-alloc: [0] 0 1 2 3 Dec 13 01:27:33.959735 kernel: kvm-guest: PV spinlocks enabled Dec 13 01:27:33.959744 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Dec 13 01:27:33.959760 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=2fdbba50b59d8c8a9877a81151806ddc16f473fe99b9ba0d8825997d654583ff Dec 13 01:27:33.959774 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Dec 13 01:27:33.959784 kernel: random: crng init done Dec 13 01:27:33.959794 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Dec 13 01:27:33.959803 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Dec 13 01:27:33.959813 kernel: Fallback order for Node 0: 0 Dec 13 01:27:33.959823 kernel: Built 1 zonelists, mobility grouping on. Total pages: 629759 Dec 13 01:27:33.959833 kernel: Policy zone: DMA32 Dec 13 01:27:33.959842 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Dec 13 01:27:33.959856 kernel: Memory: 2395612K/2567000K available (12288K kernel code, 2299K rwdata, 22724K rodata, 42844K init, 2348K bss, 171128K reserved, 0K cma-reserved) Dec 13 01:27:33.959867 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Dec 13 01:27:33.959876 kernel: ftrace: allocating 37902 entries in 149 pages Dec 13 01:27:33.959885 kernel: ftrace: allocated 149 pages with 4 groups Dec 13 01:27:33.959895 kernel: Dynamic Preempt: voluntary Dec 13 01:27:33.959916 kernel: rcu: Preemptible hierarchical RCU implementation. Dec 13 01:27:33.959935 kernel: rcu: RCU event tracing is enabled. Dec 13 01:27:33.959946 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Dec 13 01:27:33.959957 kernel: Trampoline variant of Tasks RCU enabled. Dec 13 01:27:33.959967 kernel: Rude variant of Tasks RCU enabled. Dec 13 01:27:33.959978 kernel: Tracing variant of Tasks RCU enabled. Dec 13 01:27:33.959988 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Dec 13 01:27:33.960003 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Dec 13 01:27:33.960013 kernel: NR_IRQS: 33024, nr_irqs: 456, preallocated irqs: 16 Dec 13 01:27:33.960059 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Dec 13 01:27:33.960071 kernel: Console: colour dummy device 80x25 Dec 13 01:27:33.960081 kernel: printk: console [ttyS0] enabled Dec 13 01:27:33.960096 kernel: ACPI: Core revision 20230628 Dec 13 01:27:33.960107 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Dec 13 01:27:33.960117 kernel: APIC: Switch to symmetric I/O mode setup Dec 13 01:27:33.960128 kernel: x2apic enabled Dec 13 01:27:33.960138 kernel: APIC: Switched APIC routing to: physical x2apic Dec 13 01:27:33.960149 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask() Dec 13 01:27:33.960159 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself() Dec 13 01:27:33.960169 kernel: kvm-guest: setup PV IPIs Dec 13 01:27:33.960179 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Dec 13 01:27:33.960193 kernel: tsc: Marking TSC unstable due to TSCs unsynchronized Dec 13 01:27:33.960204 kernel: Calibrating delay loop (skipped) preset value.. 5589.49 BogoMIPS (lpj=2794748) Dec 13 01:27:33.960214 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Dec 13 01:27:33.960224 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Dec 13 01:27:33.960234 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Dec 13 01:27:33.960245 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Dec 13 01:27:33.960255 kernel: Spectre V2 : Mitigation: Retpolines Dec 13 01:27:33.960265 kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch Dec 13 01:27:33.960276 kernel: Spectre V2 : Spectre v2 / SpectreRSB : Filling RSB on VMEXIT Dec 13 01:27:33.960290 kernel: Spectre V2 : Enabling Speculation Barrier for firmware calls Dec 13 01:27:33.960301 kernel: RETBleed: Mitigation: untrained return thunk Dec 13 01:27:33.960316 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Dec 13 01:27:33.960327 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Dec 13 01:27:33.960337 kernel: Speculative Return Stack Overflow: IBPB-extending microcode not applied! Dec 13 01:27:33.960348 kernel: Speculative Return Stack Overflow: WARNING: See https://kernel.org/doc/html/latest/admin-guide/hw-vuln/srso.html for mitigation options. Dec 13 01:27:33.960359 kernel: Speculative Return Stack Overflow: Vulnerable: Safe RET, no microcode Dec 13 01:27:33.960370 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Dec 13 01:27:33.960384 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Dec 13 01:27:33.960394 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Dec 13 01:27:33.960405 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Dec 13 01:27:33.960415 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. Dec 13 01:27:33.960425 kernel: Freeing SMP alternatives memory: 32K Dec 13 01:27:33.960435 kernel: pid_max: default: 32768 minimum: 301 Dec 13 01:27:33.960446 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Dec 13 01:27:33.960456 kernel: landlock: Up and running. Dec 13 01:27:33.960467 kernel: SELinux: Initializing. Dec 13 01:27:33.960481 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Dec 13 01:27:33.960491 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Dec 13 01:27:33.960501 kernel: smpboot: CPU0: AMD EPYC 7402P 24-Core Processor (family: 0x17, model: 0x31, stepping: 0x0) Dec 13 01:27:33.960522 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Dec 13 01:27:33.960533 kernel: RCU Tasks Rude: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Dec 13 01:27:33.960544 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Dec 13 01:27:33.960554 kernel: Performance Events: Fam17h+ core perfctr, AMD PMU driver. Dec 13 01:27:33.960565 kernel: ... version: 0 Dec 13 01:27:33.960575 kernel: ... bit width: 48 Dec 13 01:27:33.960590 kernel: ... generic registers: 6 Dec 13 01:27:33.960600 kernel: ... value mask: 0000ffffffffffff Dec 13 01:27:33.960611 kernel: ... max period: 00007fffffffffff Dec 13 01:27:33.960620 kernel: ... fixed-purpose events: 0 Dec 13 01:27:33.960630 kernel: ... event mask: 000000000000003f Dec 13 01:27:33.960649 kernel: signal: max sigframe size: 1776 Dec 13 01:27:33.960660 kernel: rcu: Hierarchical SRCU implementation. Dec 13 01:27:33.960671 kernel: rcu: Max phase no-delay instances is 400. Dec 13 01:27:33.960682 kernel: smp: Bringing up secondary CPUs ... Dec 13 01:27:33.960696 kernel: smpboot: x86: Booting SMP configuration: Dec 13 01:27:33.960706 kernel: .... node #0, CPUs: #1 #2 #3 Dec 13 01:27:33.960716 kernel: smp: Brought up 1 node, 4 CPUs Dec 13 01:27:33.960726 kernel: smpboot: Max logical packages: 1 Dec 13 01:27:33.960736 kernel: smpboot: Total of 4 processors activated (22357.98 BogoMIPS) Dec 13 01:27:33.960747 kernel: devtmpfs: initialized Dec 13 01:27:33.960757 kernel: x86/mm: Memory block size: 128MB Dec 13 01:27:33.960769 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00800000-0x00807fff] (32768 bytes) Dec 13 01:27:33.960781 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x0080b000-0x0080bfff] (4096 bytes) Dec 13 01:27:33.960797 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00810000-0x008fffff] (983040 bytes) Dec 13 01:27:33.960808 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cb7f000-0x9cbfefff] (524288 bytes) Dec 13 01:27:33.960820 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cf60000-0x9cffffff] (655360 bytes) Dec 13 01:27:33.960830 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Dec 13 01:27:33.960841 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Dec 13 01:27:33.960851 kernel: pinctrl core: initialized pinctrl subsystem Dec 13 01:27:33.960862 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Dec 13 01:27:33.960872 kernel: audit: initializing netlink subsys (disabled) Dec 13 01:27:33.960881 kernel: audit: type=2000 audit(1734053252.477:1): state=initialized audit_enabled=0 res=1 Dec 13 01:27:33.960895 kernel: thermal_sys: Registered thermal governor 'step_wise' Dec 13 01:27:33.960905 kernel: thermal_sys: Registered thermal governor 'user_space' Dec 13 01:27:33.960915 kernel: cpuidle: using governor menu Dec 13 01:27:33.960926 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Dec 13 01:27:33.960937 kernel: dca service started, version 1.12.1 Dec 13 01:27:33.960948 kernel: PCI: MMCONFIG for domain 0000 [bus 00-ff] at [mem 0xb0000000-0xbfffffff] (base 0xb0000000) Dec 13 01:27:33.960958 kernel: PCI: MMCONFIG at [mem 0xb0000000-0xbfffffff] reserved as E820 entry Dec 13 01:27:33.960969 kernel: PCI: Using configuration type 1 for base access Dec 13 01:27:33.960980 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Dec 13 01:27:33.960993 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Dec 13 01:27:33.961003 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Dec 13 01:27:33.961014 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Dec 13 01:27:33.961024 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Dec 13 01:27:33.961058 kernel: ACPI: Added _OSI(Module Device) Dec 13 01:27:33.961068 kernel: ACPI: Added _OSI(Processor Device) Dec 13 01:27:33.961079 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Dec 13 01:27:33.961089 kernel: ACPI: Added _OSI(Processor Aggregator Device) Dec 13 01:27:33.961100 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Dec 13 01:27:33.961115 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Dec 13 01:27:33.961125 kernel: ACPI: Interpreter enabled Dec 13 01:27:33.961141 kernel: ACPI: PM: (supports S0 S3 S5) Dec 13 01:27:33.961167 kernel: ACPI: Using IOAPIC for interrupt routing Dec 13 01:27:33.961185 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Dec 13 01:27:33.961198 kernel: PCI: Using E820 reservations for host bridge windows Dec 13 01:27:33.961208 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Dec 13 01:27:33.961219 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Dec 13 01:27:33.961534 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Dec 13 01:27:33.961764 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Dec 13 01:27:33.961972 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Dec 13 01:27:33.961989 kernel: PCI host bridge to bus 0000:00 Dec 13 01:27:33.962321 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Dec 13 01:27:33.962486 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Dec 13 01:27:33.962726 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Dec 13 01:27:33.962970 kernel: pci_bus 0000:00: root bus resource [mem 0x9d000000-0xafffffff window] Dec 13 01:27:33.963182 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window] Dec 13 01:27:33.963382 kernel: pci_bus 0000:00: root bus resource [mem 0x800000000-0xfffffffff window] Dec 13 01:27:33.963681 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Dec 13 01:27:33.963986 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 Dec 13 01:27:33.964226 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 Dec 13 01:27:33.964441 kernel: pci 0000:00:01.0: reg 0x10: [mem 0xc0000000-0xc0ffffff pref] Dec 13 01:27:33.964612 kernel: pci 0000:00:01.0: reg 0x18: [mem 0xc1044000-0xc1044fff] Dec 13 01:27:33.964787 kernel: pci 0000:00:01.0: reg 0x30: [mem 0xffff0000-0xffffffff pref] Dec 13 01:27:33.964954 kernel: pci 0000:00:01.0: BAR 0: assigned to efifb Dec 13 01:27:33.965159 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Dec 13 01:27:33.965358 kernel: pci 0000:00:02.0: [1af4:1005] type 00 class 0x00ff00 Dec 13 01:27:33.965572 kernel: pci 0000:00:02.0: reg 0x10: [io 0x6100-0x611f] Dec 13 01:27:33.965796 kernel: pci 0000:00:02.0: reg 0x14: [mem 0xc1043000-0xc1043fff] Dec 13 01:27:33.965982 kernel: pci 0000:00:02.0: reg 0x20: [mem 0x800000000-0x800003fff 64bit pref] Dec 13 01:27:33.966226 kernel: pci 0000:00:03.0: [1af4:1001] type 00 class 0x010000 Dec 13 01:27:33.966401 kernel: pci 0000:00:03.0: reg 0x10: [io 0x6000-0x607f] Dec 13 01:27:33.966573 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xc1042000-0xc1042fff] Dec 13 01:27:33.966752 kernel: pci 0000:00:03.0: reg 0x20: [mem 0x800004000-0x800007fff 64bit pref] Dec 13 01:27:33.966997 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 Dec 13 01:27:33.967200 kernel: pci 0000:00:04.0: reg 0x10: [io 0x60e0-0x60ff] Dec 13 01:27:33.967371 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xc1041000-0xc1041fff] Dec 13 01:27:33.967540 kernel: pci 0000:00:04.0: reg 0x20: [mem 0x800008000-0x80000bfff 64bit pref] Dec 13 01:27:33.967724 kernel: pci 0000:00:04.0: reg 0x30: [mem 0xfffc0000-0xffffffff pref] Dec 13 01:27:33.967916 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 Dec 13 01:27:33.968102 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Dec 13 01:27:33.968300 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 Dec 13 01:27:33.968490 kernel: pci 0000:00:1f.2: reg 0x20: [io 0x60c0-0x60df] Dec 13 01:27:33.968668 kernel: pci 0000:00:1f.2: reg 0x24: [mem 0xc1040000-0xc1040fff] Dec 13 01:27:33.968878 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 Dec 13 01:27:33.969197 kernel: pci 0000:00:1f.3: reg 0x20: [io 0x6080-0x60bf] Dec 13 01:27:33.969239 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Dec 13 01:27:33.969248 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Dec 13 01:27:33.969257 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Dec 13 01:27:33.969287 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Dec 13 01:27:33.969296 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Dec 13 01:27:33.969304 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Dec 13 01:27:33.969312 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Dec 13 01:27:33.969320 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Dec 13 01:27:33.969328 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Dec 13 01:27:33.969337 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Dec 13 01:27:33.969345 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Dec 13 01:27:33.969354 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Dec 13 01:27:33.969366 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Dec 13 01:27:33.969374 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Dec 13 01:27:33.969382 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Dec 13 01:27:33.969390 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Dec 13 01:27:33.969398 kernel: iommu: Default domain type: Translated Dec 13 01:27:33.969407 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Dec 13 01:27:33.969415 kernel: efivars: Registered efivars operations Dec 13 01:27:33.969431 kernel: PCI: Using ACPI for IRQ routing Dec 13 01:27:33.969439 kernel: PCI: pci_cache_line_size set to 64 bytes Dec 13 01:27:33.969452 kernel: e820: reserve RAM buffer [mem 0x0080b000-0x008fffff] Dec 13 01:27:33.969460 kernel: e820: reserve RAM buffer [mem 0x00810000-0x008fffff] Dec 13 01:27:33.969468 kernel: e820: reserve RAM buffer [mem 0x9c8ef000-0x9fffffff] Dec 13 01:27:33.969476 kernel: e820: reserve RAM buffer [mem 0x9cf40000-0x9fffffff] Dec 13 01:27:33.969633 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Dec 13 01:27:33.969794 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Dec 13 01:27:33.969973 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Dec 13 01:27:33.969993 kernel: vgaarb: loaded Dec 13 01:27:33.970020 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Dec 13 01:27:33.971402 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Dec 13 01:27:33.971411 kernel: clocksource: Switched to clocksource kvm-clock Dec 13 01:27:33.971419 kernel: VFS: Disk quotas dquot_6.6.0 Dec 13 01:27:33.971428 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Dec 13 01:27:33.971436 kernel: pnp: PnP ACPI init Dec 13 01:27:33.971625 kernel: system 00:05: [mem 0xb0000000-0xbfffffff window] has been reserved Dec 13 01:27:33.971651 kernel: pnp: PnP ACPI: found 6 devices Dec 13 01:27:33.971661 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Dec 13 01:27:33.971676 kernel: NET: Registered PF_INET protocol family Dec 13 01:27:33.971685 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Dec 13 01:27:33.971693 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Dec 13 01:27:33.971702 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Dec 13 01:27:33.971710 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Dec 13 01:27:33.971718 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Dec 13 01:27:33.971726 kernel: TCP: Hash tables configured (established 32768 bind 32768) Dec 13 01:27:33.971734 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Dec 13 01:27:33.971742 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Dec 13 01:27:33.971752 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Dec 13 01:27:33.971760 kernel: NET: Registered PF_XDP protocol family Dec 13 01:27:33.971895 kernel: pci 0000:00:04.0: can't claim BAR 6 [mem 0xfffc0000-0xffffffff pref]: no compatible bridge window Dec 13 01:27:33.972044 kernel: pci 0000:00:04.0: BAR 6: assigned [mem 0x9d000000-0x9d03ffff pref] Dec 13 01:27:33.972176 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Dec 13 01:27:33.972292 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Dec 13 01:27:33.972406 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Dec 13 01:27:33.972521 kernel: pci_bus 0000:00: resource 7 [mem 0x9d000000-0xafffffff window] Dec 13 01:27:33.972655 kernel: pci_bus 0000:00: resource 8 [mem 0xc0000000-0xfebfffff window] Dec 13 01:27:33.972786 kernel: pci_bus 0000:00: resource 9 [mem 0x800000000-0xfffffffff window] Dec 13 01:27:33.972799 kernel: PCI: CLS 0 bytes, default 64 Dec 13 01:27:33.972807 kernel: Initialise system trusted keyrings Dec 13 01:27:33.972815 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Dec 13 01:27:33.972823 kernel: Key type asymmetric registered Dec 13 01:27:33.972831 kernel: Asymmetric key parser 'x509' registered Dec 13 01:27:33.972839 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Dec 13 01:27:33.972852 kernel: io scheduler mq-deadline registered Dec 13 01:27:33.972863 kernel: io scheduler kyber registered Dec 13 01:27:33.972871 kernel: io scheduler bfq registered Dec 13 01:27:33.972879 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Dec 13 01:27:33.972888 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Dec 13 01:27:33.972896 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Dec 13 01:27:33.972904 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 Dec 13 01:27:33.972912 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Dec 13 01:27:33.972921 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Dec 13 01:27:33.972931 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Dec 13 01:27:33.972939 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Dec 13 01:27:33.972947 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Dec 13 01:27:33.973133 kernel: rtc_cmos 00:04: RTC can wake from S4 Dec 13 01:27:33.973275 kernel: rtc_cmos 00:04: registered as rtc0 Dec 13 01:27:33.973287 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Dec 13 01:27:33.973404 kernel: rtc_cmos 00:04: setting system clock to 2024-12-13T01:27:33 UTC (1734053253) Dec 13 01:27:33.973536 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram, hpet irqs Dec 13 01:27:33.973554 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Dec 13 01:27:33.973563 kernel: efifb: probing for efifb Dec 13 01:27:33.973572 kernel: efifb: framebuffer at 0xc0000000, using 1408k, total 1408k Dec 13 01:27:33.973580 kernel: efifb: mode is 800x600x24, linelength=2400, pages=1 Dec 13 01:27:33.973588 kernel: efifb: scrolling: redraw Dec 13 01:27:33.973596 kernel: efifb: Truecolor: size=0:8:8:8, shift=0:16:8:0 Dec 13 01:27:33.973604 kernel: Console: switching to colour frame buffer device 100x37 Dec 13 01:27:33.973629 kernel: fb0: EFI VGA frame buffer device Dec 13 01:27:33.973650 kernel: pstore: Using crash dump compression: deflate Dec 13 01:27:33.973663 kernel: pstore: Registered efi_pstore as persistent store backend Dec 13 01:27:33.973671 kernel: NET: Registered PF_INET6 protocol family Dec 13 01:27:33.973679 kernel: Segment Routing with IPv6 Dec 13 01:27:33.973687 kernel: In-situ OAM (IOAM) with IPv6 Dec 13 01:27:33.973695 kernel: NET: Registered PF_PACKET protocol family Dec 13 01:27:33.973703 kernel: Key type dns_resolver registered Dec 13 01:27:33.973711 kernel: IPI shorthand broadcast: enabled Dec 13 01:27:33.973720 kernel: sched_clock: Marking stable (1271004231, 145398953)->(1530277477, -113874293) Dec 13 01:27:33.973728 kernel: registered taskstats version 1 Dec 13 01:27:33.973739 kernel: Loading compiled-in X.509 certificates Dec 13 01:27:33.973747 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.65-flatcar: c82d546f528d79a5758dcebbc47fb6daf92836a0' Dec 13 01:27:33.973755 kernel: Key type .fscrypt registered Dec 13 01:27:33.973763 kernel: Key type fscrypt-provisioning registered Dec 13 01:27:33.973772 kernel: ima: No TPM chip found, activating TPM-bypass! Dec 13 01:27:33.973780 kernel: ima: Allocated hash algorithm: sha1 Dec 13 01:27:33.973788 kernel: ima: No architecture policies found Dec 13 01:27:33.973796 kernel: clk: Disabling unused clocks Dec 13 01:27:33.973804 kernel: Freeing unused kernel image (initmem) memory: 42844K Dec 13 01:27:33.973814 kernel: Write protecting the kernel read-only data: 36864k Dec 13 01:27:33.973823 kernel: Freeing unused kernel image (rodata/data gap) memory: 1852K Dec 13 01:27:33.973831 kernel: Run /init as init process Dec 13 01:27:33.973840 kernel: with arguments: Dec 13 01:27:33.973848 kernel: /init Dec 13 01:27:33.973856 kernel: with environment: Dec 13 01:27:33.973864 kernel: HOME=/ Dec 13 01:27:33.973872 kernel: TERM=linux Dec 13 01:27:33.973880 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Dec 13 01:27:33.973899 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Dec 13 01:27:33.973910 systemd[1]: Detected virtualization kvm. Dec 13 01:27:33.973919 systemd[1]: Detected architecture x86-64. Dec 13 01:27:33.973928 systemd[1]: Running in initrd. Dec 13 01:27:33.973941 systemd[1]: No hostname configured, using default hostname. Dec 13 01:27:33.973949 systemd[1]: Hostname set to . Dec 13 01:27:33.973958 systemd[1]: Initializing machine ID from VM UUID. Dec 13 01:27:33.973966 systemd[1]: Queued start job for default target initrd.target. Dec 13 01:27:33.973975 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Dec 13 01:27:33.973984 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Dec 13 01:27:33.973993 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Dec 13 01:27:33.974002 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Dec 13 01:27:33.974013 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Dec 13 01:27:33.974022 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Dec 13 01:27:33.974059 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Dec 13 01:27:33.974068 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Dec 13 01:27:33.974077 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Dec 13 01:27:33.974085 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Dec 13 01:27:33.974094 systemd[1]: Reached target paths.target - Path Units. Dec 13 01:27:33.974117 systemd[1]: Reached target slices.target - Slice Units. Dec 13 01:27:33.974126 systemd[1]: Reached target swap.target - Swaps. Dec 13 01:27:33.974135 systemd[1]: Reached target timers.target - Timer Units. Dec 13 01:27:33.974143 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Dec 13 01:27:33.974152 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Dec 13 01:27:33.974163 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Dec 13 01:27:33.974172 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Dec 13 01:27:33.974181 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Dec 13 01:27:33.974193 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Dec 13 01:27:33.974201 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Dec 13 01:27:33.974210 systemd[1]: Reached target sockets.target - Socket Units. Dec 13 01:27:33.974219 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Dec 13 01:27:33.974235 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Dec 13 01:27:33.974252 systemd[1]: Finished network-cleanup.service - Network Cleanup. Dec 13 01:27:33.974261 systemd[1]: Starting systemd-fsck-usr.service... Dec 13 01:27:33.974270 systemd[1]: Starting systemd-journald.service - Journal Service... Dec 13 01:27:33.974279 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Dec 13 01:27:33.974291 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Dec 13 01:27:33.974299 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Dec 13 01:27:33.974308 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Dec 13 01:27:33.974316 systemd[1]: Finished systemd-fsck-usr.service. Dec 13 01:27:33.974325 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Dec 13 01:27:33.974361 systemd-journald[193]: Collecting audit messages is disabled. Dec 13 01:27:33.974384 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Dec 13 01:27:33.974394 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Dec 13 01:27:33.974405 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Dec 13 01:27:33.974414 systemd-journald[193]: Journal started Dec 13 01:27:33.974433 systemd-journald[193]: Runtime Journal (/run/log/journal/a7d347ba4b9b4307afea67f08a96d540) is 6.0M, max 48.3M, 42.2M free. Dec 13 01:27:33.979428 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Dec 13 01:27:33.955690 systemd-modules-load[195]: Inserted module 'overlay' Dec 13 01:27:33.984053 systemd[1]: Started systemd-journald.service - Journal Service. Dec 13 01:27:33.987060 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Dec 13 01:27:33.989304 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Dec 13 01:27:33.993761 kernel: Bridge firewalling registered Dec 13 01:27:33.992175 systemd-modules-load[195]: Inserted module 'br_netfilter' Dec 13 01:27:33.994737 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Dec 13 01:27:33.998473 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Dec 13 01:27:34.003563 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Dec 13 01:27:34.049644 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Dec 13 01:27:34.051935 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Dec 13 01:27:34.057967 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Dec 13 01:27:34.060800 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Dec 13 01:27:34.069096 dracut-cmdline[225]: dracut-dracut-053 Dec 13 01:27:34.069168 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Dec 13 01:27:34.083957 dracut-cmdline[225]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=2fdbba50b59d8c8a9877a81151806ddc16f473fe99b9ba0d8825997d654583ff Dec 13 01:27:34.116270 systemd-resolved[233]: Positive Trust Anchors: Dec 13 01:27:34.116291 systemd-resolved[233]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Dec 13 01:27:34.116322 systemd-resolved[233]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Dec 13 01:27:34.127359 systemd-resolved[233]: Defaulting to hostname 'linux'. Dec 13 01:27:34.129540 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Dec 13 01:27:34.131775 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Dec 13 01:27:34.172059 kernel: SCSI subsystem initialized Dec 13 01:27:34.181048 kernel: Loading iSCSI transport class v2.0-870. Dec 13 01:27:34.192054 kernel: iscsi: registered transport (tcp) Dec 13 01:27:34.213051 kernel: iscsi: registered transport (qla4xxx) Dec 13 01:27:34.213076 kernel: QLogic iSCSI HBA Driver Dec 13 01:27:34.267930 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Dec 13 01:27:34.286191 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Dec 13 01:27:34.314156 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Dec 13 01:27:34.314212 kernel: device-mapper: uevent: version 1.0.3 Dec 13 01:27:34.315235 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Dec 13 01:27:34.358078 kernel: raid6: avx2x4 gen() 30222 MB/s Dec 13 01:27:34.375060 kernel: raid6: avx2x2 gen() 30747 MB/s Dec 13 01:27:34.392172 kernel: raid6: avx2x1 gen() 25132 MB/s Dec 13 01:27:34.392229 kernel: raid6: using algorithm avx2x2 gen() 30747 MB/s Dec 13 01:27:34.410340 kernel: raid6: .... xor() 18437 MB/s, rmw enabled Dec 13 01:27:34.410433 kernel: raid6: using avx2x2 recovery algorithm Dec 13 01:27:34.433073 kernel: xor: automatically using best checksumming function avx Dec 13 01:27:34.615073 kernel: Btrfs loaded, zoned=no, fsverity=no Dec 13 01:27:34.630039 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Dec 13 01:27:34.642186 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Dec 13 01:27:34.654567 systemd-udevd[414]: Using default interface naming scheme 'v255'. Dec 13 01:27:34.659857 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Dec 13 01:27:34.667433 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Dec 13 01:27:34.683062 dracut-pre-trigger[417]: rd.md=0: removing MD RAID activation Dec 13 01:27:34.717483 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Dec 13 01:27:34.727184 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Dec 13 01:27:34.796508 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Dec 13 01:27:34.806380 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Dec 13 01:27:34.820163 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Dec 13 01:27:34.821119 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Dec 13 01:27:34.824849 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Dec 13 01:27:34.824960 systemd[1]: Reached target remote-fs.target - Remote File Systems. Dec 13 01:27:34.837316 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Dec 13 01:27:34.847061 kernel: virtio_blk virtio1: 4/0/0 default/read/poll queues Dec 13 01:27:34.888568 kernel: cryptd: max_cpu_qlen set to 1000 Dec 13 01:27:34.888599 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Dec 13 01:27:34.888825 kernel: AVX2 version of gcm_enc/dec engaged. Dec 13 01:27:34.888843 kernel: AES CTR mode by8 optimization enabled Dec 13 01:27:34.888859 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Dec 13 01:27:34.888883 kernel: GPT:9289727 != 19775487 Dec 13 01:27:34.888899 kernel: GPT:Alternate GPT header not at the end of the disk. Dec 13 01:27:34.888914 kernel: GPT:9289727 != 19775487 Dec 13 01:27:34.888928 kernel: GPT: Use GNU Parted to correct GPT errors. Dec 13 01:27:34.888943 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Dec 13 01:27:34.888958 kernel: libata version 3.00 loaded. Dec 13 01:27:34.850183 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Dec 13 01:27:34.874709 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Dec 13 01:27:34.874906 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Dec 13 01:27:34.876696 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Dec 13 01:27:34.877888 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Dec 13 01:27:34.878223 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Dec 13 01:27:34.879515 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Dec 13 01:27:34.891087 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Dec 13 01:27:34.904830 kernel: ahci 0000:00:1f.2: version 3.0 Dec 13 01:27:34.927204 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Dec 13 01:27:34.927228 kernel: ahci 0000:00:1f.2: AHCI 0001.0000 32 slots 6 ports 1.5 Gbps 0x3f impl SATA mode Dec 13 01:27:34.927452 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Dec 13 01:27:34.927658 kernel: scsi host0: ahci Dec 13 01:27:34.927876 kernel: scsi host1: ahci Dec 13 01:27:34.928112 kernel: scsi host2: ahci Dec 13 01:27:34.928332 kernel: scsi host3: ahci Dec 13 01:27:34.928537 kernel: scsi host4: ahci Dec 13 01:27:34.928773 kernel: scsi host5: ahci Dec 13 01:27:34.928977 kernel: ata1: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040100 irq 34 Dec 13 01:27:34.929002 kernel: ata2: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040180 irq 34 Dec 13 01:27:34.929018 kernel: ata3: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040200 irq 34 Dec 13 01:27:34.929116 kernel: ata4: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040280 irq 34 Dec 13 01:27:34.929133 kernel: ata5: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040300 irq 34 Dec 13 01:27:34.929145 kernel: ata6: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040380 irq 34 Dec 13 01:27:34.910641 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Dec 13 01:27:34.921235 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Dec 13 01:27:34.934114 kernel: BTRFS: device fsid c3b72f8a-27ca-4d37-9d0e-1ec3c4bdc3be devid 1 transid 41 /dev/vda3 scanned by (udev-worker) (465) Dec 13 01:27:34.937074 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/vda6 scanned by (udev-worker) (475) Dec 13 01:27:34.953604 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Dec 13 01:27:34.961575 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Dec 13 01:27:34.964311 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Dec 13 01:27:34.975117 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Dec 13 01:27:34.977808 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Dec 13 01:27:34.985687 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Dec 13 01:27:35.032186 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Dec 13 01:27:35.082555 disk-uuid[575]: Primary Header is updated. Dec 13 01:27:35.082555 disk-uuid[575]: Secondary Entries is updated. Dec 13 01:27:35.082555 disk-uuid[575]: Secondary Header is updated. Dec 13 01:27:35.087053 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Dec 13 01:27:35.091051 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Dec 13 01:27:35.095065 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Dec 13 01:27:35.097051 kernel: block device autoloading is deprecated and will be removed. Dec 13 01:27:35.235624 kernel: ata2: SATA link down (SStatus 0 SControl 300) Dec 13 01:27:35.235717 kernel: ata1: SATA link down (SStatus 0 SControl 300) Dec 13 01:27:35.235729 kernel: ata5: SATA link down (SStatus 0 SControl 300) Dec 13 01:27:35.237075 kernel: ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Dec 13 01:27:35.238059 kernel: ata6: SATA link down (SStatus 0 SControl 300) Dec 13 01:27:35.239051 kernel: ata4: SATA link down (SStatus 0 SControl 300) Dec 13 01:27:35.239076 kernel: ata3.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Dec 13 01:27:35.240488 kernel: ata3.00: applying bridge limits Dec 13 01:27:35.240500 kernel: ata3.00: configured for UDMA/100 Dec 13 01:27:35.241051 kernel: scsi 2:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Dec 13 01:27:35.290074 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Dec 13 01:27:35.305815 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Dec 13 01:27:35.305837 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 Dec 13 01:27:36.124057 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Dec 13 01:27:36.124523 disk-uuid[576]: The operation has completed successfully. Dec 13 01:27:36.159173 systemd[1]: disk-uuid.service: Deactivated successfully. Dec 13 01:27:36.159367 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Dec 13 01:27:36.191316 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Dec 13 01:27:36.197807 sh[596]: Success Dec 13 01:27:36.211082 kernel: device-mapper: verity: sha256 using implementation "sha256-ni" Dec 13 01:27:36.259138 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Dec 13 01:27:36.285769 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Dec 13 01:27:36.288899 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Dec 13 01:27:36.304133 kernel: BTRFS info (device dm-0): first mount of filesystem c3b72f8a-27ca-4d37-9d0e-1ec3c4bdc3be Dec 13 01:27:36.304231 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Dec 13 01:27:36.304246 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Dec 13 01:27:36.306476 kernel: BTRFS info (device dm-0): disabling log replay at mount time Dec 13 01:27:36.306511 kernel: BTRFS info (device dm-0): using free space tree Dec 13 01:27:36.312485 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Dec 13 01:27:36.315822 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Dec 13 01:27:36.333321 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Dec 13 01:27:36.336956 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Dec 13 01:27:36.347003 kernel: BTRFS info (device vda6): first mount of filesystem db063747-cac8-4176-8963-c216c1b11dcb Dec 13 01:27:36.347080 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Dec 13 01:27:36.347093 kernel: BTRFS info (device vda6): using free space tree Dec 13 01:27:36.351085 kernel: BTRFS info (device vda6): auto enabling async discard Dec 13 01:27:36.364132 systemd[1]: mnt-oem.mount: Deactivated successfully. Dec 13 01:27:36.366133 kernel: BTRFS info (device vda6): last unmount of filesystem db063747-cac8-4176-8963-c216c1b11dcb Dec 13 01:27:36.377492 systemd[1]: Finished ignition-setup.service - Ignition (setup). Dec 13 01:27:36.384386 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Dec 13 01:27:36.455953 ignition[682]: Ignition 2.19.0 Dec 13 01:27:36.456588 ignition[682]: Stage: fetch-offline Dec 13 01:27:36.456674 ignition[682]: no configs at "/usr/lib/ignition/base.d" Dec 13 01:27:36.456691 ignition[682]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Dec 13 01:27:36.456840 ignition[682]: parsed url from cmdline: "" Dec 13 01:27:36.456846 ignition[682]: no config URL provided Dec 13 01:27:36.456854 ignition[682]: reading system config file "/usr/lib/ignition/user.ign" Dec 13 01:27:36.456868 ignition[682]: no config at "/usr/lib/ignition/user.ign" Dec 13 01:27:36.456906 ignition[682]: op(1): [started] loading QEMU firmware config module Dec 13 01:27:36.456914 ignition[682]: op(1): executing: "modprobe" "qemu_fw_cfg" Dec 13 01:27:36.467700 ignition[682]: op(1): [finished] loading QEMU firmware config module Dec 13 01:27:36.467731 ignition[682]: QEMU firmware config was not found. Ignoring... Dec 13 01:27:36.470263 ignition[682]: parsing config with SHA512: 2f56b93993a3915584e291a1ea09b74f8b48d505fecc01a99905ba6888265756a33147f7e5789b50388c6d0599c59adf7a47c3a1b59a1dce1900473fa3442c00 Dec 13 01:27:36.475614 unknown[682]: fetched base config from "system" Dec 13 01:27:36.475630 unknown[682]: fetched user config from "qemu" Dec 13 01:27:36.476065 ignition[682]: fetch-offline: fetch-offline passed Dec 13 01:27:36.476154 ignition[682]: Ignition finished successfully Dec 13 01:27:36.479185 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Dec 13 01:27:36.504102 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Dec 13 01:27:36.518292 systemd[1]: Starting systemd-networkd.service - Network Configuration... Dec 13 01:27:36.551620 systemd-networkd[785]: lo: Link UP Dec 13 01:27:36.551637 systemd-networkd[785]: lo: Gained carrier Dec 13 01:27:36.553914 systemd-networkd[785]: Enumeration completed Dec 13 01:27:36.554054 systemd[1]: Started systemd-networkd.service - Network Configuration. Dec 13 01:27:36.554613 systemd-networkd[785]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 13 01:27:36.554619 systemd-networkd[785]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Dec 13 01:27:36.555882 systemd-networkd[785]: eth0: Link UP Dec 13 01:27:36.555887 systemd-networkd[785]: eth0: Gained carrier Dec 13 01:27:36.555897 systemd-networkd[785]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 13 01:27:36.556315 systemd[1]: Reached target network.target - Network. Dec 13 01:27:36.558331 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Dec 13 01:27:36.567283 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Dec 13 01:27:36.581204 systemd-networkd[785]: eth0: DHCPv4 address 10.0.0.50/16, gateway 10.0.0.1 acquired from 10.0.0.1 Dec 13 01:27:36.588011 ignition[787]: Ignition 2.19.0 Dec 13 01:27:36.588057 ignition[787]: Stage: kargs Dec 13 01:27:36.588360 ignition[787]: no configs at "/usr/lib/ignition/base.d" Dec 13 01:27:36.588379 ignition[787]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Dec 13 01:27:36.589410 ignition[787]: kargs: kargs passed Dec 13 01:27:36.593168 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Dec 13 01:27:36.589487 ignition[787]: Ignition finished successfully Dec 13 01:27:36.605545 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Dec 13 01:27:36.621483 ignition[796]: Ignition 2.19.0 Dec 13 01:27:36.621503 ignition[796]: Stage: disks Dec 13 01:27:36.621773 ignition[796]: no configs at "/usr/lib/ignition/base.d" Dec 13 01:27:36.621794 ignition[796]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Dec 13 01:27:36.625582 systemd[1]: Finished ignition-disks.service - Ignition (disks). Dec 13 01:27:36.622831 ignition[796]: disks: disks passed Dec 13 01:27:36.627779 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Dec 13 01:27:36.622900 ignition[796]: Ignition finished successfully Dec 13 01:27:36.630321 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Dec 13 01:27:36.632694 systemd[1]: Reached target local-fs.target - Local File Systems. Dec 13 01:27:36.634023 systemd[1]: Reached target sysinit.target - System Initialization. Dec 13 01:27:36.636591 systemd[1]: Reached target basic.target - Basic System. Dec 13 01:27:36.649411 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Dec 13 01:27:36.668695 systemd-fsck[808]: ROOT: clean, 14/553520 files, 52654/553472 blocks Dec 13 01:27:36.676497 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Dec 13 01:27:36.686331 systemd[1]: Mounting sysroot.mount - /sysroot... Dec 13 01:27:36.785077 kernel: EXT4-fs (vda9): mounted filesystem 390119fa-ab9c-4f50-b046-3b5c76c46193 r/w with ordered data mode. Quota mode: none. Dec 13 01:27:36.786055 systemd[1]: Mounted sysroot.mount - /sysroot. Dec 13 01:27:36.788690 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Dec 13 01:27:36.806302 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Dec 13 01:27:36.809813 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Dec 13 01:27:36.812780 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Dec 13 01:27:36.812847 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Dec 13 01:27:36.822817 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/vda6 scanned by mount (816) Dec 13 01:27:36.822847 kernel: BTRFS info (device vda6): first mount of filesystem db063747-cac8-4176-8963-c216c1b11dcb Dec 13 01:27:36.822859 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Dec 13 01:27:36.822870 kernel: BTRFS info (device vda6): using free space tree Dec 13 01:27:36.812883 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Dec 13 01:27:36.825054 kernel: BTRFS info (device vda6): auto enabling async discard Dec 13 01:27:36.826771 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Dec 13 01:27:36.828972 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Dec 13 01:27:36.842230 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Dec 13 01:27:36.876321 systemd-resolved[233]: Detected conflict on linux IN A 10.0.0.50 Dec 13 01:27:36.876339 systemd-resolved[233]: Hostname conflict, changing published hostname from 'linux' to 'linux9'. Dec 13 01:27:36.881674 initrd-setup-root[840]: cut: /sysroot/etc/passwd: No such file or directory Dec 13 01:27:36.886459 initrd-setup-root[847]: cut: /sysroot/etc/group: No such file or directory Dec 13 01:27:36.890451 initrd-setup-root[854]: cut: /sysroot/etc/shadow: No such file or directory Dec 13 01:27:36.894871 initrd-setup-root[861]: cut: /sysroot/etc/gshadow: No such file or directory Dec 13 01:27:36.991504 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Dec 13 01:27:37.008168 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Dec 13 01:27:37.011153 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Dec 13 01:27:37.021057 kernel: BTRFS info (device vda6): last unmount of filesystem db063747-cac8-4176-8963-c216c1b11dcb Dec 13 01:27:37.041152 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Dec 13 01:27:37.080998 ignition[933]: INFO : Ignition 2.19.0 Dec 13 01:27:37.080998 ignition[933]: INFO : Stage: mount Dec 13 01:27:37.082901 ignition[933]: INFO : no configs at "/usr/lib/ignition/base.d" Dec 13 01:27:37.082901 ignition[933]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Dec 13 01:27:37.082901 ignition[933]: INFO : mount: mount passed Dec 13 01:27:37.082901 ignition[933]: INFO : Ignition finished successfully Dec 13 01:27:37.085505 systemd[1]: Finished ignition-mount.service - Ignition (mount). Dec 13 01:27:37.109339 systemd[1]: Starting ignition-files.service - Ignition (files)... Dec 13 01:27:37.303063 systemd[1]: sysroot-oem.mount: Deactivated successfully. Dec 13 01:27:37.316232 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Dec 13 01:27:37.323077 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/vda6 scanned by mount (942) Dec 13 01:27:37.323143 kernel: BTRFS info (device vda6): first mount of filesystem db063747-cac8-4176-8963-c216c1b11dcb Dec 13 01:27:37.325237 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Dec 13 01:27:37.325262 kernel: BTRFS info (device vda6): using free space tree Dec 13 01:27:37.329050 kernel: BTRFS info (device vda6): auto enabling async discard Dec 13 01:27:37.329950 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Dec 13 01:27:37.355974 ignition[959]: INFO : Ignition 2.19.0 Dec 13 01:27:37.355974 ignition[959]: INFO : Stage: files Dec 13 01:27:37.358196 ignition[959]: INFO : no configs at "/usr/lib/ignition/base.d" Dec 13 01:27:37.358196 ignition[959]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Dec 13 01:27:37.358196 ignition[959]: DEBUG : files: compiled without relabeling support, skipping Dec 13 01:27:37.358196 ignition[959]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Dec 13 01:27:37.358196 ignition[959]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Dec 13 01:27:37.365843 ignition[959]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Dec 13 01:27:37.365843 ignition[959]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Dec 13 01:27:37.365843 ignition[959]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Dec 13 01:27:37.365843 ignition[959]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/etc/flatcar-cgroupv1" Dec 13 01:27:37.365843 ignition[959]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/etc/flatcar-cgroupv1" Dec 13 01:27:37.365843 ignition[959]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Dec 13 01:27:37.365843 ignition[959]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Dec 13 01:27:37.365843 ignition[959]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/etc/flatcar/update.conf" Dec 13 01:27:37.365843 ignition[959]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/etc/flatcar/update.conf" Dec 13 01:27:37.365843 ignition[959]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.29.2-x86-64.raw" Dec 13 01:27:37.365843 ignition[959]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.29.2-x86-64.raw" Dec 13 01:27:37.365843 ignition[959]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.29.2-x86-64.raw" Dec 13 01:27:37.365843 ignition[959]: INFO : files: createFilesystemsFiles: createFiles: op(7): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.29.2-x86-64.raw: attempt #1 Dec 13 01:27:37.361383 unknown[959]: wrote ssh authorized keys file for user: core Dec 13 01:27:37.707370 systemd-networkd[785]: eth0: Gained IPv6LL Dec 13 01:27:37.731724 ignition[959]: INFO : files: createFilesystemsFiles: createFiles: op(7): GET result: OK Dec 13 01:27:38.672721 ignition[959]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.29.2-x86-64.raw" Dec 13 01:27:38.672721 ignition[959]: INFO : files: op(8): [started] processing unit "containerd.service" Dec 13 01:27:38.677105 ignition[959]: INFO : files: op(8): op(9): [started] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Dec 13 01:27:38.679706 ignition[959]: INFO : files: op(8): op(9): [finished] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Dec 13 01:27:38.679706 ignition[959]: INFO : files: op(8): [finished] processing unit "containerd.service" Dec 13 01:27:38.679706 ignition[959]: INFO : files: op(a): [started] processing unit "coreos-metadata.service" Dec 13 01:27:38.679706 ignition[959]: INFO : files: op(a): op(b): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Dec 13 01:27:38.679706 ignition[959]: INFO : files: op(a): op(b): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Dec 13 01:27:38.679706 ignition[959]: INFO : files: op(a): [finished] processing unit "coreos-metadata.service" Dec 13 01:27:38.679706 ignition[959]: INFO : files: op(c): [started] setting preset to disabled for "coreos-metadata.service" Dec 13 01:27:38.736559 ignition[959]: INFO : files: op(c): op(d): [started] removing enablement symlink(s) for "coreos-metadata.service" Dec 13 01:27:38.747381 ignition[959]: INFO : files: op(c): op(d): [finished] removing enablement symlink(s) for "coreos-metadata.service" Dec 13 01:27:38.749784 ignition[959]: INFO : files: op(c): [finished] setting preset to disabled for "coreos-metadata.service" Dec 13 01:27:38.749784 ignition[959]: INFO : files: createResultFile: createFiles: op(e): [started] writing file "/sysroot/etc/.ignition-result.json" Dec 13 01:27:38.749784 ignition[959]: INFO : files: createResultFile: createFiles: op(e): [finished] writing file "/sysroot/etc/.ignition-result.json" Dec 13 01:27:38.749784 ignition[959]: INFO : files: files passed Dec 13 01:27:38.749784 ignition[959]: INFO : Ignition finished successfully Dec 13 01:27:38.754767 systemd[1]: Finished ignition-files.service - Ignition (files). Dec 13 01:27:38.771320 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Dec 13 01:27:38.778339 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Dec 13 01:27:38.781935 systemd[1]: ignition-quench.service: Deactivated successfully. Dec 13 01:27:38.783230 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Dec 13 01:27:38.795518 initrd-setup-root-after-ignition[987]: grep: /sysroot/oem/oem-release: No such file or directory Dec 13 01:27:38.802227 initrd-setup-root-after-ignition[989]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Dec 13 01:27:38.802227 initrd-setup-root-after-ignition[989]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Dec 13 01:27:38.806962 initrd-setup-root-after-ignition[993]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Dec 13 01:27:38.809591 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Dec 13 01:27:38.813674 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Dec 13 01:27:38.836395 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Dec 13 01:27:38.874455 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Dec 13 01:27:38.874626 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Dec 13 01:27:38.882250 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Dec 13 01:27:38.885299 systemd[1]: Reached target initrd.target - Initrd Default Target. Dec 13 01:27:38.888054 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Dec 13 01:27:38.897294 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Dec 13 01:27:38.918215 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Dec 13 01:27:38.922316 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Dec 13 01:27:38.942192 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Dec 13 01:27:38.943749 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Dec 13 01:27:38.946607 systemd[1]: Stopped target timers.target - Timer Units. Dec 13 01:27:38.949729 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Dec 13 01:27:38.949923 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Dec 13 01:27:38.952319 systemd[1]: Stopped target initrd.target - Initrd Default Target. Dec 13 01:27:38.954794 systemd[1]: Stopped target basic.target - Basic System. Dec 13 01:27:38.957558 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Dec 13 01:27:38.959570 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Dec 13 01:27:38.962088 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Dec 13 01:27:38.964239 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Dec 13 01:27:38.966756 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Dec 13 01:27:38.969606 systemd[1]: Stopped target sysinit.target - System Initialization. Dec 13 01:27:38.972252 systemd[1]: Stopped target local-fs.target - Local File Systems. Dec 13 01:27:38.974746 systemd[1]: Stopped target swap.target - Swaps. Dec 13 01:27:38.977384 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Dec 13 01:27:38.977681 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Dec 13 01:27:38.981182 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Dec 13 01:27:38.983418 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Dec 13 01:27:38.985159 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Dec 13 01:27:38.985540 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Dec 13 01:27:38.988539 systemd[1]: dracut-initqueue.service: Deactivated successfully. Dec 13 01:27:38.988735 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Dec 13 01:27:38.991496 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Dec 13 01:27:38.991645 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Dec 13 01:27:38.993913 systemd[1]: Stopped target paths.target - Path Units. Dec 13 01:27:38.996469 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Dec 13 01:27:38.996800 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Dec 13 01:27:38.999751 systemd[1]: Stopped target slices.target - Slice Units. Dec 13 01:27:39.001437 systemd[1]: Stopped target sockets.target - Socket Units. Dec 13 01:27:39.004728 systemd[1]: iscsid.socket: Deactivated successfully. Dec 13 01:27:39.004886 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Dec 13 01:27:39.007839 systemd[1]: iscsiuio.socket: Deactivated successfully. Dec 13 01:27:39.008011 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Dec 13 01:27:39.009088 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Dec 13 01:27:39.009259 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Dec 13 01:27:39.014868 systemd[1]: ignition-files.service: Deactivated successfully. Dec 13 01:27:39.015118 systemd[1]: Stopped ignition-files.service - Ignition (files). Dec 13 01:27:39.034417 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Dec 13 01:27:39.038588 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Dec 13 01:27:39.040725 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Dec 13 01:27:39.040997 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Dec 13 01:27:39.042787 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Dec 13 01:27:39.042954 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Dec 13 01:27:39.053158 systemd[1]: initrd-cleanup.service: Deactivated successfully. Dec 13 01:27:39.053314 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Dec 13 01:27:39.057491 ignition[1013]: INFO : Ignition 2.19.0 Dec 13 01:27:39.057491 ignition[1013]: INFO : Stage: umount Dec 13 01:27:39.057491 ignition[1013]: INFO : no configs at "/usr/lib/ignition/base.d" Dec 13 01:27:39.057491 ignition[1013]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Dec 13 01:27:39.059349 ignition[1013]: INFO : umount: umount passed Dec 13 01:27:39.059349 ignition[1013]: INFO : Ignition finished successfully Dec 13 01:27:39.062635 systemd[1]: ignition-mount.service: Deactivated successfully. Dec 13 01:27:39.062870 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Dec 13 01:27:39.063780 systemd[1]: Stopped target network.target - Network. Dec 13 01:27:39.064385 systemd[1]: ignition-disks.service: Deactivated successfully. Dec 13 01:27:39.064463 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Dec 13 01:27:39.065329 systemd[1]: ignition-kargs.service: Deactivated successfully. Dec 13 01:27:39.065402 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Dec 13 01:27:39.065650 systemd[1]: ignition-setup.service: Deactivated successfully. Dec 13 01:27:39.065699 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Dec 13 01:27:39.065992 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Dec 13 01:27:39.066096 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Dec 13 01:27:39.067055 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Dec 13 01:27:39.067669 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Dec 13 01:27:39.096662 systemd[1]: systemd-resolved.service: Deactivated successfully. Dec 13 01:27:39.096901 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Dec 13 01:27:39.096920 systemd-networkd[785]: eth0: DHCPv6 lease lost Dec 13 01:27:39.100982 systemd[1]: systemd-networkd.service: Deactivated successfully. Dec 13 01:27:39.101246 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Dec 13 01:27:39.103598 systemd[1]: systemd-networkd.socket: Deactivated successfully. Dec 13 01:27:39.103680 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Dec 13 01:27:39.114235 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Dec 13 01:27:39.115375 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Dec 13 01:27:39.115481 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Dec 13 01:27:39.118351 systemd[1]: systemd-sysctl.service: Deactivated successfully. Dec 13 01:27:39.118455 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Dec 13 01:27:39.120604 systemd[1]: systemd-modules-load.service: Deactivated successfully. Dec 13 01:27:39.120686 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Dec 13 01:27:39.123667 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Dec 13 01:27:39.123740 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Dec 13 01:27:39.126621 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Dec 13 01:27:39.146981 systemd[1]: sysroot-boot.mount: Deactivated successfully. Dec 13 01:27:39.149545 systemd[1]: systemd-udevd.service: Deactivated successfully. Dec 13 01:27:39.149849 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Dec 13 01:27:39.153981 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Dec 13 01:27:39.154091 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Dec 13 01:27:39.156346 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Dec 13 01:27:39.156411 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Dec 13 01:27:39.159187 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Dec 13 01:27:39.159293 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Dec 13 01:27:39.162167 systemd[1]: dracut-cmdline.service: Deactivated successfully. Dec 13 01:27:39.162241 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Dec 13 01:27:39.164825 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Dec 13 01:27:39.164952 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Dec 13 01:27:39.169806 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Dec 13 01:27:39.171467 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Dec 13 01:27:39.171586 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Dec 13 01:27:39.174290 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Dec 13 01:27:39.174384 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Dec 13 01:27:39.176930 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Dec 13 01:27:39.177056 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Dec 13 01:27:39.181736 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Dec 13 01:27:39.181818 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Dec 13 01:27:39.188367 systemd[1]: network-cleanup.service: Deactivated successfully. Dec 13 01:27:39.188561 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Dec 13 01:27:39.211757 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Dec 13 01:27:39.211968 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Dec 13 01:27:39.323370 systemd[1]: sysroot-boot.service: Deactivated successfully. Dec 13 01:27:39.323594 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Dec 13 01:27:39.330161 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Dec 13 01:27:39.336333 systemd[1]: initrd-setup-root.service: Deactivated successfully. Dec 13 01:27:39.336504 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Dec 13 01:27:39.354445 systemd[1]: Starting initrd-switch-root.service - Switch Root... Dec 13 01:27:39.365246 systemd[1]: Switching root. Dec 13 01:27:39.403208 systemd-journald[193]: Journal stopped Dec 13 01:27:40.971332 systemd-journald[193]: Received SIGTERM from PID 1 (systemd). Dec 13 01:27:40.971405 kernel: SELinux: policy capability network_peer_controls=1 Dec 13 01:27:40.971424 kernel: SELinux: policy capability open_perms=1 Dec 13 01:27:40.971444 kernel: SELinux: policy capability extended_socket_class=1 Dec 13 01:27:40.971468 kernel: SELinux: policy capability always_check_network=0 Dec 13 01:27:40.971480 kernel: SELinux: policy capability cgroup_seclabel=1 Dec 13 01:27:40.971501 kernel: SELinux: policy capability nnp_nosuid_transition=1 Dec 13 01:27:40.971513 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Dec 13 01:27:40.971524 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Dec 13 01:27:40.971536 kernel: audit: type=1403 audit(1734053260.017:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Dec 13 01:27:40.971549 systemd[1]: Successfully loaded SELinux policy in 52.547ms. Dec 13 01:27:40.971572 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 13.377ms. Dec 13 01:27:40.971590 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Dec 13 01:27:40.971603 systemd[1]: Detected virtualization kvm. Dec 13 01:27:40.971615 systemd[1]: Detected architecture x86-64. Dec 13 01:27:40.971627 systemd[1]: Detected first boot. Dec 13 01:27:40.971639 systemd[1]: Initializing machine ID from VM UUID. Dec 13 01:27:40.971652 zram_generator::config[1074]: No configuration found. Dec 13 01:27:40.971665 systemd[1]: Populated /etc with preset unit settings. Dec 13 01:27:40.971678 systemd[1]: Queued start job for default target multi-user.target. Dec 13 01:27:40.971695 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Dec 13 01:27:40.971709 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Dec 13 01:27:40.971725 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Dec 13 01:27:40.971737 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Dec 13 01:27:40.971749 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Dec 13 01:27:40.971762 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Dec 13 01:27:40.971774 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Dec 13 01:27:40.971786 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Dec 13 01:27:40.971803 systemd[1]: Created slice user.slice - User and Session Slice. Dec 13 01:27:40.971816 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Dec 13 01:27:40.971828 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Dec 13 01:27:40.971840 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Dec 13 01:27:40.971858 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Dec 13 01:27:40.971873 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Dec 13 01:27:40.971885 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Dec 13 01:27:40.971898 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Dec 13 01:27:40.971910 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Dec 13 01:27:40.971927 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Dec 13 01:27:40.971940 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Dec 13 01:27:40.971952 systemd[1]: Reached target remote-fs.target - Remote File Systems. Dec 13 01:27:40.971964 systemd[1]: Reached target slices.target - Slice Units. Dec 13 01:27:40.971976 systemd[1]: Reached target swap.target - Swaps. Dec 13 01:27:40.971989 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Dec 13 01:27:40.972001 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Dec 13 01:27:40.972013 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Dec 13 01:27:40.972043 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Dec 13 01:27:40.972055 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Dec 13 01:27:40.972067 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Dec 13 01:27:40.972079 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Dec 13 01:27:40.972092 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Dec 13 01:27:40.972104 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Dec 13 01:27:40.972116 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Dec 13 01:27:40.972128 systemd[1]: Mounting media.mount - External Media Directory... Dec 13 01:27:40.972145 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 13 01:27:40.972157 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Dec 13 01:27:40.972174 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Dec 13 01:27:40.972187 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Dec 13 01:27:40.972199 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Dec 13 01:27:40.972212 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Dec 13 01:27:40.972224 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Dec 13 01:27:40.972236 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Dec 13 01:27:40.972248 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Dec 13 01:27:40.972260 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Dec 13 01:27:40.972278 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Dec 13 01:27:40.972290 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Dec 13 01:27:40.972302 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Dec 13 01:27:40.972315 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Dec 13 01:27:40.972327 systemd[1]: systemd-journald.service: unit configures an IP firewall, but the local system does not support BPF/cgroup firewalling. Dec 13 01:27:40.972340 systemd[1]: systemd-journald.service: (This warning is only shown for the first unit using IP firewalling.) Dec 13 01:27:40.972352 systemd[1]: Starting systemd-journald.service - Journal Service... Dec 13 01:27:40.972364 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Dec 13 01:27:40.972381 kernel: fuse: init (API version 7.39) Dec 13 01:27:40.972393 kernel: loop: module loaded Dec 13 01:27:40.972404 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Dec 13 01:27:40.972420 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Dec 13 01:27:40.972432 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Dec 13 01:27:40.972453 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 13 01:27:40.972487 systemd-journald[1151]: Collecting audit messages is disabled. Dec 13 01:27:40.972509 systemd-journald[1151]: Journal started Dec 13 01:27:40.972537 systemd-journald[1151]: Runtime Journal (/run/log/journal/a7d347ba4b9b4307afea67f08a96d540) is 6.0M, max 48.3M, 42.2M free. Dec 13 01:27:40.982111 kernel: ACPI: bus type drm_connector registered Dec 13 01:27:40.987464 systemd[1]: Started systemd-journald.service - Journal Service. Dec 13 01:27:40.989835 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Dec 13 01:27:40.991055 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Dec 13 01:27:40.992316 systemd[1]: Mounted media.mount - External Media Directory. Dec 13 01:27:40.993420 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Dec 13 01:27:40.996196 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Dec 13 01:27:40.997570 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Dec 13 01:27:40.999099 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Dec 13 01:27:41.001083 systemd[1]: modprobe@configfs.service: Deactivated successfully. Dec 13 01:27:41.001310 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Dec 13 01:27:41.002882 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Dec 13 01:27:41.003125 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Dec 13 01:27:41.004822 systemd[1]: modprobe@drm.service: Deactivated successfully. Dec 13 01:27:41.005053 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Dec 13 01:27:41.006709 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Dec 13 01:27:41.006925 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Dec 13 01:27:41.008497 systemd[1]: modprobe@fuse.service: Deactivated successfully. Dec 13 01:27:41.008717 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Dec 13 01:27:41.010175 systemd[1]: modprobe@loop.service: Deactivated successfully. Dec 13 01:27:41.010391 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Dec 13 01:27:41.012161 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Dec 13 01:27:41.013815 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Dec 13 01:27:41.015506 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Dec 13 01:27:41.017423 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Dec 13 01:27:41.033486 systemd[1]: Reached target network-pre.target - Preparation for Network. Dec 13 01:27:41.050199 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Dec 13 01:27:41.053083 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Dec 13 01:27:41.054331 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Dec 13 01:27:41.059204 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Dec 13 01:27:41.062069 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Dec 13 01:27:41.064345 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Dec 13 01:27:41.070301 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Dec 13 01:27:41.071623 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Dec 13 01:27:41.077510 systemd-journald[1151]: Time spent on flushing to /var/log/journal/a7d347ba4b9b4307afea67f08a96d540 is 25.534ms for 968 entries. Dec 13 01:27:41.077510 systemd-journald[1151]: System Journal (/var/log/journal/a7d347ba4b9b4307afea67f08a96d540) is 8.0M, max 195.6M, 187.6M free. Dec 13 01:27:41.117231 systemd-journald[1151]: Received client request to flush runtime journal. Dec 13 01:27:41.078673 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Dec 13 01:27:41.088360 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Dec 13 01:27:41.094924 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Dec 13 01:27:41.096740 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Dec 13 01:27:41.098409 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Dec 13 01:27:41.114368 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Dec 13 01:27:41.116918 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Dec 13 01:27:41.120894 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Dec 13 01:27:41.129408 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Dec 13 01:27:41.132331 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Dec 13 01:27:41.135783 udevadm[1216]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in. Dec 13 01:27:41.141003 systemd-tmpfiles[1210]: ACLs are not supported, ignoring. Dec 13 01:27:41.141039 systemd-tmpfiles[1210]: ACLs are not supported, ignoring. Dec 13 01:27:41.148660 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Dec 13 01:27:41.156182 systemd[1]: Starting systemd-sysusers.service - Create System Users... Dec 13 01:27:41.184819 systemd[1]: Finished systemd-sysusers.service - Create System Users. Dec 13 01:27:41.193172 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Dec 13 01:27:41.215449 systemd-tmpfiles[1233]: ACLs are not supported, ignoring. Dec 13 01:27:41.215478 systemd-tmpfiles[1233]: ACLs are not supported, ignoring. Dec 13 01:27:41.249593 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Dec 13 01:27:42.017401 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Dec 13 01:27:42.035228 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Dec 13 01:27:42.067394 systemd-udevd[1239]: Using default interface naming scheme 'v255'. Dec 13 01:27:42.084331 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Dec 13 01:27:42.106215 systemd[1]: Starting systemd-networkd.service - Network Configuration... Dec 13 01:27:42.140120 kernel: BTRFS info: devid 1 device path /dev/mapper/usr changed to /dev/dm-0 scanned by (udev-worker) (1240) Dec 13 01:27:42.145067 kernel: BTRFS info: devid 1 device path /dev/dm-0 changed to /dev/mapper/usr scanned by (udev-worker) (1240) Dec 13 01:27:42.146332 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Dec 13 01:27:42.149275 systemd[1]: Found device dev-ttyS0.device - /dev/ttyS0. Dec 13 01:27:42.160059 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 41 scanned by (udev-worker) (1256) Dec 13 01:27:42.393059 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Dec 13 01:27:42.402015 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Dec 13 01:27:42.407064 kernel: ACPI: button: Power Button [PWRF] Dec 13 01:27:42.414050 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input4 Dec 13 01:27:42.420954 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Dec 13 01:27:42.421392 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Dec 13 01:27:42.426104 kernel: mousedev: PS/2 mouse device common for all mice Dec 13 01:27:42.426149 kernel: i801_smbus 0000:00:1f.3: Enabling SMBus device Dec 13 01:27:42.429187 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Dec 13 01:27:42.429360 kernel: i2c i2c-0: 1/1 memory slots populated (from DMI) Dec 13 01:27:42.429570 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Dec 13 01:27:42.431322 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Dec 13 01:27:42.442819 systemd[1]: Started systemd-userdbd.service - User Database Manager. Dec 13 01:27:42.566333 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Dec 13 01:27:42.609192 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Dec 13 01:27:42.631297 kernel: kvm_amd: TSC scaling supported Dec 13 01:27:42.631371 kernel: kvm_amd: Nested Virtualization enabled Dec 13 01:27:42.631386 kernel: kvm_amd: Nested Paging enabled Dec 13 01:27:42.631398 kernel: kvm_amd: LBR virtualization supported Dec 13 01:27:42.635050 kernel: kvm_amd: Virtual VMLOAD VMSAVE supported Dec 13 01:27:42.635085 kernel: kvm_amd: Virtual GIF supported Dec 13 01:27:42.657065 kernel: EDAC MC: Ver: 3.0.0 Dec 13 01:27:42.663181 systemd-networkd[1255]: lo: Link UP Dec 13 01:27:42.663195 systemd-networkd[1255]: lo: Gained carrier Dec 13 01:27:42.665004 systemd-networkd[1255]: Enumeration completed Dec 13 01:27:42.665199 systemd[1]: Started systemd-networkd.service - Network Configuration. Dec 13 01:27:42.665743 systemd-networkd[1255]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 13 01:27:42.665754 systemd-networkd[1255]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Dec 13 01:27:42.666609 systemd-networkd[1255]: eth0: Link UP Dec 13 01:27:42.666620 systemd-networkd[1255]: eth0: Gained carrier Dec 13 01:27:42.666632 systemd-networkd[1255]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 13 01:27:42.677502 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Dec 13 01:27:42.693177 systemd-networkd[1255]: eth0: DHCPv4 address 10.0.0.50/16, gateway 10.0.0.1 acquired from 10.0.0.1 Dec 13 01:27:42.703889 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Dec 13 01:27:42.717438 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Dec 13 01:27:42.729408 lvm[1289]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Dec 13 01:27:42.765904 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Dec 13 01:27:42.767625 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Dec 13 01:27:42.786436 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Dec 13 01:27:42.793000 lvm[1292]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Dec 13 01:27:42.831434 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Dec 13 01:27:42.833237 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Dec 13 01:27:42.834680 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Dec 13 01:27:42.834720 systemd[1]: Reached target local-fs.target - Local File Systems. Dec 13 01:27:42.835949 systemd[1]: Reached target machines.target - Containers. Dec 13 01:27:42.838285 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Dec 13 01:27:42.851190 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Dec 13 01:27:42.854922 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Dec 13 01:27:42.856243 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Dec 13 01:27:42.857501 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Dec 13 01:27:42.862199 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Dec 13 01:27:42.867339 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Dec 13 01:27:42.870319 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Dec 13 01:27:42.875079 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Dec 13 01:27:42.885052 kernel: loop0: detected capacity change from 0 to 211296 Dec 13 01:27:42.893761 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Dec 13 01:27:42.894732 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Dec 13 01:27:42.904103 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Dec 13 01:27:42.932060 kernel: loop1: detected capacity change from 0 to 140768 Dec 13 01:27:42.989110 kernel: loop2: detected capacity change from 0 to 142488 Dec 13 01:27:43.032068 kernel: loop3: detected capacity change from 0 to 211296 Dec 13 01:27:43.044060 kernel: loop4: detected capacity change from 0 to 140768 Dec 13 01:27:43.116063 kernel: loop5: detected capacity change from 0 to 142488 Dec 13 01:27:43.125281 (sd-merge)[1312]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'. Dec 13 01:27:43.125975 (sd-merge)[1312]: Merged extensions into '/usr'. Dec 13 01:27:43.130143 systemd[1]: Reloading requested from client PID 1300 ('systemd-sysext') (unit systemd-sysext.service)... Dec 13 01:27:43.130157 systemd[1]: Reloading... Dec 13 01:27:43.260059 zram_generator::config[1340]: No configuration found. Dec 13 01:27:43.342280 ldconfig[1296]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Dec 13 01:27:43.436359 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Dec 13 01:27:43.508407 systemd[1]: Reloading finished in 377 ms. Dec 13 01:27:43.530811 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Dec 13 01:27:43.532866 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Dec 13 01:27:43.553390 systemd[1]: Starting ensure-sysext.service... Dec 13 01:27:43.556542 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Dec 13 01:27:43.561432 systemd[1]: Reloading requested from client PID 1384 ('systemctl') (unit ensure-sysext.service)... Dec 13 01:27:43.561453 systemd[1]: Reloading... Dec 13 01:27:43.584365 systemd-tmpfiles[1385]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Dec 13 01:27:43.584768 systemd-tmpfiles[1385]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Dec 13 01:27:43.585877 systemd-tmpfiles[1385]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Dec 13 01:27:43.586329 systemd-tmpfiles[1385]: ACLs are not supported, ignoring. Dec 13 01:27:43.586454 systemd-tmpfiles[1385]: ACLs are not supported, ignoring. Dec 13 01:27:43.590405 systemd-tmpfiles[1385]: Detected autofs mount point /boot during canonicalization of boot. Dec 13 01:27:43.590419 systemd-tmpfiles[1385]: Skipping /boot Dec 13 01:27:43.608882 systemd-tmpfiles[1385]: Detected autofs mount point /boot during canonicalization of boot. Dec 13 01:27:43.608907 systemd-tmpfiles[1385]: Skipping /boot Dec 13 01:27:43.688075 zram_generator::config[1413]: No configuration found. Dec 13 01:27:43.855004 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Dec 13 01:27:43.923095 systemd[1]: Reloading finished in 361 ms. Dec 13 01:27:43.944023 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Dec 13 01:27:43.962818 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Dec 13 01:27:43.965679 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Dec 13 01:27:43.968296 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Dec 13 01:27:43.972514 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Dec 13 01:27:43.979342 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Dec 13 01:27:44.029455 augenrules[1479]: No rules Dec 13 01:27:44.093616 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Dec 13 01:27:44.095723 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Dec 13 01:27:44.097529 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Dec 13 01:27:44.121040 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 13 01:27:44.121273 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Dec 13 01:27:44.123086 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Dec 13 01:27:44.128084 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Dec 13 01:27:44.132394 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Dec 13 01:27:44.134304 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Dec 13 01:27:44.137806 systemd[1]: Starting systemd-update-done.service - Update is Completed... Dec 13 01:27:44.139083 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Dec 13 01:27:44.139474 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 13 01:27:44.141819 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Dec 13 01:27:44.144184 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Dec 13 01:27:44.144540 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Dec 13 01:27:44.147126 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Dec 13 01:27:44.147370 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Dec 13 01:27:44.149669 systemd[1]: modprobe@loop.service: Deactivated successfully. Dec 13 01:27:44.149958 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Dec 13 01:27:44.155548 systemd[1]: Finished systemd-update-done.service - Update is Completed. Dec 13 01:27:44.163069 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 13 01:27:44.163534 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Dec 13 01:27:44.180536 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Dec 13 01:27:44.183755 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Dec 13 01:27:44.187389 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Dec 13 01:27:44.188746 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Dec 13 01:27:44.188977 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Dec 13 01:27:44.189207 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 13 01:27:44.191224 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Dec 13 01:27:44.191515 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Dec 13 01:27:44.193670 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Dec 13 01:27:44.193939 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Dec 13 01:27:44.198924 systemd[1]: modprobe@loop.service: Deactivated successfully. Dec 13 01:27:44.199323 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Dec 13 01:27:44.205437 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 13 01:27:44.205803 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Dec 13 01:27:44.212401 systemd-resolved[1462]: Positive Trust Anchors: Dec 13 01:27:44.212416 systemd-resolved[1462]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Dec 13 01:27:44.212449 systemd-resolved[1462]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Dec 13 01:27:44.213461 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Dec 13 01:27:44.216658 systemd-resolved[1462]: Defaulting to hostname 'linux'. Dec 13 01:27:44.216917 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Dec 13 01:27:44.219857 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Dec 13 01:27:44.225622 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Dec 13 01:27:44.227323 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Dec 13 01:27:44.227910 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Dec 13 01:27:44.228186 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 13 01:27:44.230527 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Dec 13 01:27:44.233250 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Dec 13 01:27:44.233569 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Dec 13 01:27:44.235533 systemd[1]: modprobe@drm.service: Deactivated successfully. Dec 13 01:27:44.235850 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Dec 13 01:27:44.238457 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Dec 13 01:27:44.238772 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Dec 13 01:27:44.241153 systemd[1]: modprobe@loop.service: Deactivated successfully. Dec 13 01:27:44.241623 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Dec 13 01:27:44.247571 systemd[1]: Finished ensure-sysext.service. Dec 13 01:27:44.255365 systemd[1]: Reached target network.target - Network. Dec 13 01:27:44.256451 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Dec 13 01:27:44.257748 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Dec 13 01:27:44.257866 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Dec 13 01:27:44.271382 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Dec 13 01:27:44.341389 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Dec 13 01:27:44.342530 systemd-timesyncd[1530]: Contacted time server 10.0.0.1:123 (10.0.0.1). Dec 13 01:27:44.342579 systemd-timesyncd[1530]: Initial clock synchronization to Fri 2024-12-13 01:27:44.530213 UTC. Dec 13 01:27:44.343535 systemd[1]: Reached target sysinit.target - System Initialization. Dec 13 01:27:44.344784 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Dec 13 01:27:44.346146 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Dec 13 01:27:44.347486 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Dec 13 01:27:44.348831 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Dec 13 01:27:44.348876 systemd[1]: Reached target paths.target - Path Units. Dec 13 01:27:44.349865 systemd[1]: Reached target time-set.target - System Time Set. Dec 13 01:27:44.351215 systemd[1]: Started logrotate.timer - Daily rotation of log files. Dec 13 01:27:44.352550 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Dec 13 01:27:44.353906 systemd[1]: Reached target timers.target - Timer Units. Dec 13 01:27:44.356365 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Dec 13 01:27:44.360341 systemd[1]: Starting docker.socket - Docker Socket for the API... Dec 13 01:27:44.363726 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Dec 13 01:27:44.373295 systemd[1]: Listening on docker.socket - Docker Socket for the API. Dec 13 01:27:44.374755 systemd[1]: Reached target sockets.target - Socket Units. Dec 13 01:27:44.375889 systemd[1]: Reached target basic.target - Basic System. Dec 13 01:27:44.377308 systemd[1]: System is tainted: cgroupsv1 Dec 13 01:27:44.377386 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Dec 13 01:27:44.377429 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Dec 13 01:27:44.379836 systemd[1]: Starting containerd.service - containerd container runtime... Dec 13 01:27:44.383068 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Dec 13 01:27:44.386314 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Dec 13 01:27:44.392322 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Dec 13 01:27:44.393708 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Dec 13 01:27:44.396895 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Dec 13 01:27:44.401423 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Dec 13 01:27:44.404058 jq[1536]: false Dec 13 01:27:44.409282 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Dec 13 01:27:44.419305 systemd[1]: Starting systemd-logind.service - User Login Management... Dec 13 01:27:44.420870 extend-filesystems[1538]: Found loop3 Dec 13 01:27:44.420870 extend-filesystems[1538]: Found loop4 Dec 13 01:27:44.422989 extend-filesystems[1538]: Found loop5 Dec 13 01:27:44.422989 extend-filesystems[1538]: Found sr0 Dec 13 01:27:44.422989 extend-filesystems[1538]: Found vda Dec 13 01:27:44.422989 extend-filesystems[1538]: Found vda1 Dec 13 01:27:44.422989 extend-filesystems[1538]: Found vda2 Dec 13 01:27:44.422989 extend-filesystems[1538]: Found vda3 Dec 13 01:27:44.422989 extend-filesystems[1538]: Found usr Dec 13 01:27:44.422989 extend-filesystems[1538]: Found vda4 Dec 13 01:27:44.422989 extend-filesystems[1538]: Found vda6 Dec 13 01:27:44.422989 extend-filesystems[1538]: Found vda7 Dec 13 01:27:44.422989 extend-filesystems[1538]: Found vda9 Dec 13 01:27:44.422989 extend-filesystems[1538]: Checking size of /dev/vda9 Dec 13 01:27:44.421273 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Dec 13 01:27:44.422273 dbus-daemon[1535]: [system] SELinux support is enabled Dec 13 01:27:44.426419 systemd[1]: Starting update-engine.service - Update Engine... Dec 13 01:27:44.428520 systemd-networkd[1255]: eth0: Gained IPv6LL Dec 13 01:27:44.439159 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Dec 13 01:27:44.442076 systemd[1]: Started dbus.service - D-Bus System Message Bus. Dec 13 01:27:44.447006 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Dec 13 01:27:44.451481 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Dec 13 01:27:44.451895 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Dec 13 01:27:44.452389 systemd[1]: motdgen.service: Deactivated successfully. Dec 13 01:27:44.452778 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Dec 13 01:27:44.454970 jq[1555]: true Dec 13 01:27:44.456826 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Dec 13 01:27:44.457765 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Dec 13 01:27:44.470560 extend-filesystems[1538]: Resized partition /dev/vda9 Dec 13 01:27:44.471934 update_engine[1553]: I20241213 01:27:44.471830 1553 main.cc:92] Flatcar Update Engine starting Dec 13 01:27:44.475713 extend-filesystems[1566]: resize2fs 1.47.1 (20-May-2024) Dec 13 01:27:44.481068 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Dec 13 01:27:44.484069 update_engine[1553]: I20241213 01:27:44.482463 1553 update_check_scheduler.cc:74] Next update check in 4m40s Dec 13 01:27:44.489940 (ntainerd)[1572]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Dec 13 01:27:44.495957 jq[1564]: true Dec 13 01:27:44.503283 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 41 scanned by (udev-worker) (1244) Dec 13 01:27:44.512780 systemd[1]: Started update-engine.service - Update Engine. Dec 13 01:27:44.514928 systemd[1]: Reached target network-online.target - Network is Online. Dec 13 01:27:44.519221 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Dec 13 01:27:44.541075 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Dec 13 01:27:44.551358 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 13 01:27:44.562264 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Dec 13 01:27:44.564798 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Dec 13 01:27:44.564839 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Dec 13 01:27:44.569293 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Dec 13 01:27:44.569320 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Dec 13 01:27:44.625487 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Dec 13 01:27:44.628810 systemd[1]: Started locksmithd.service - Cluster reboot manager. Dec 13 01:27:44.668816 systemd[1]: coreos-metadata.service: Deactivated successfully. Dec 13 01:27:44.669265 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Dec 13 01:27:44.670998 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Dec 13 01:27:44.675499 systemd-logind[1547]: Watching system buttons on /dev/input/event1 (Power Button) Dec 13 01:27:44.719901 extend-filesystems[1566]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Dec 13 01:27:44.719901 extend-filesystems[1566]: old_desc_blocks = 1, new_desc_blocks = 1 Dec 13 01:27:44.719901 extend-filesystems[1566]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Dec 13 01:27:44.724721 sshd_keygen[1558]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Dec 13 01:27:44.675540 systemd-logind[1547]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Dec 13 01:27:44.724960 extend-filesystems[1538]: Resized filesystem in /dev/vda9 Dec 13 01:27:44.718150 systemd-logind[1547]: New seat seat0. Dec 13 01:27:44.726875 systemd[1]: Started systemd-logind.service - User Login Management. Dec 13 01:27:44.736506 systemd[1]: extend-filesystems.service: Deactivated successfully. Dec 13 01:27:44.736843 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Dec 13 01:27:44.741225 locksmithd[1597]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Dec 13 01:27:44.742462 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Dec 13 01:27:44.751503 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Dec 13 01:27:44.759508 systemd[1]: Starting issuegen.service - Generate /run/issue... Dec 13 01:27:44.777673 systemd[1]: issuegen.service: Deactivated successfully. Dec 13 01:27:44.778363 systemd[1]: Finished issuegen.service - Generate /run/issue. Dec 13 01:27:44.798491 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Dec 13 01:27:44.845292 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Dec 13 01:27:44.857454 systemd[1]: Started getty@tty1.service - Getty on tty1. Dec 13 01:27:44.861370 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Dec 13 01:27:44.864218 systemd[1]: Reached target getty.target - Login Prompts. Dec 13 01:27:44.869497 bash[1596]: Updated "/home/core/.ssh/authorized_keys" Dec 13 01:27:44.871912 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Dec 13 01:27:44.874658 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Dec 13 01:27:45.020785 containerd[1572]: time="2024-12-13T01:27:45.020578023Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Dec 13 01:27:45.103292 containerd[1572]: time="2024-12-13T01:27:45.103217855Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Dec 13 01:27:45.108198 containerd[1572]: time="2024-12-13T01:27:45.106132249Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.65-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Dec 13 01:27:45.108198 containerd[1572]: time="2024-12-13T01:27:45.106180145Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Dec 13 01:27:45.108198 containerd[1572]: time="2024-12-13T01:27:45.106203617Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Dec 13 01:27:45.108198 containerd[1572]: time="2024-12-13T01:27:45.106522935Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Dec 13 01:27:45.108198 containerd[1572]: time="2024-12-13T01:27:45.106593432Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Dec 13 01:27:45.108198 containerd[1572]: time="2024-12-13T01:27:45.106734301Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Dec 13 01:27:45.108198 containerd[1572]: time="2024-12-13T01:27:45.106757168Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Dec 13 01:27:45.108198 containerd[1572]: time="2024-12-13T01:27:45.107205512Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Dec 13 01:27:45.108198 containerd[1572]: time="2024-12-13T01:27:45.107232142Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Dec 13 01:27:45.108198 containerd[1572]: time="2024-12-13T01:27:45.107252916Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Dec 13 01:27:45.108198 containerd[1572]: time="2024-12-13T01:27:45.107270450Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Dec 13 01:27:45.108489 containerd[1572]: time="2024-12-13T01:27:45.107476329Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Dec 13 01:27:45.108489 containerd[1572]: time="2024-12-13T01:27:45.107894291Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Dec 13 01:27:45.108489 containerd[1572]: time="2024-12-13T01:27:45.108195061Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Dec 13 01:27:45.108489 containerd[1572]: time="2024-12-13T01:27:45.108220430Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Dec 13 01:27:45.108489 containerd[1572]: time="2024-12-13T01:27:45.108391763Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Dec 13 01:27:45.108602 containerd[1572]: time="2024-12-13T01:27:45.108499101Z" level=info msg="metadata content store policy set" policy=shared Dec 13 01:27:45.128311 containerd[1572]: time="2024-12-13T01:27:45.128226594Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Dec 13 01:27:45.128444 containerd[1572]: time="2024-12-13T01:27:45.128329821Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Dec 13 01:27:45.128444 containerd[1572]: time="2024-12-13T01:27:45.128368284Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Dec 13 01:27:45.128444 containerd[1572]: time="2024-12-13T01:27:45.128388781Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Dec 13 01:27:45.128444 containerd[1572]: time="2024-12-13T01:27:45.128409238Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Dec 13 01:27:45.128659 containerd[1572]: time="2024-12-13T01:27:45.128621167Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Dec 13 01:27:45.131301 containerd[1572]: time="2024-12-13T01:27:45.131260960Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Dec 13 01:27:45.131499 containerd[1572]: time="2024-12-13T01:27:45.131469556Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Dec 13 01:27:45.131499 containerd[1572]: time="2024-12-13T01:27:45.131494976Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Dec 13 01:27:45.131599 containerd[1572]: time="2024-12-13T01:27:45.131511290Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Dec 13 01:27:45.131599 containerd[1572]: time="2024-12-13T01:27:45.131528712Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Dec 13 01:27:45.131599 containerd[1572]: time="2024-12-13T01:27:45.131545589Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Dec 13 01:27:45.131599 containerd[1572]: time="2024-12-13T01:27:45.131576003Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Dec 13 01:27:45.131599 containerd[1572]: time="2024-12-13T01:27:45.131593404Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Dec 13 01:27:45.131771 containerd[1572]: time="2024-12-13T01:27:45.131611943Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Dec 13 01:27:45.131771 containerd[1572]: time="2024-12-13T01:27:45.131642859Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Dec 13 01:27:45.131771 containerd[1572]: time="2024-12-13T01:27:45.131661921Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Dec 13 01:27:45.131771 containerd[1572]: time="2024-12-13T01:27:45.131677240Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Dec 13 01:27:45.131771 containerd[1572]: time="2024-12-13T01:27:45.131727146Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Dec 13 01:27:45.131771 containerd[1572]: time="2024-12-13T01:27:45.131746619Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Dec 13 01:27:45.131771 containerd[1572]: time="2024-12-13T01:27:45.131762102Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Dec 13 01:27:45.132009 containerd[1572]: time="2024-12-13T01:27:45.131777647Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Dec 13 01:27:45.132009 containerd[1572]: time="2024-12-13T01:27:45.131793623Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Dec 13 01:27:45.132009 containerd[1572]: time="2024-12-13T01:27:45.131810736Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Dec 13 01:27:45.132009 containerd[1572]: time="2024-12-13T01:27:45.131830066Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Dec 13 01:27:45.132009 containerd[1572]: time="2024-12-13T01:27:45.131905175Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Dec 13 01:27:45.132009 containerd[1572]: time="2024-12-13T01:27:45.131921612Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Dec 13 01:27:45.132009 containerd[1572]: time="2024-12-13T01:27:45.131940162Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Dec 13 01:27:45.132009 containerd[1572]: time="2024-12-13T01:27:45.131958127Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Dec 13 01:27:45.132009 containerd[1572]: time="2024-12-13T01:27:45.131974482Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Dec 13 01:27:45.132009 containerd[1572]: time="2024-12-13T01:27:45.131990068Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Dec 13 01:27:45.132009 containerd[1572]: time="2024-12-13T01:27:45.132010289Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Dec 13 01:27:45.132635 containerd[1572]: time="2024-12-13T01:27:45.132035401Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Dec 13 01:27:45.132635 containerd[1572]: time="2024-12-13T01:27:45.132067158Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Dec 13 01:27:45.132635 containerd[1572]: time="2024-12-13T01:27:45.132082015Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Dec 13 01:27:45.132635 containerd[1572]: time="2024-12-13T01:27:45.132154419Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Dec 13 01:27:45.132635 containerd[1572]: time="2024-12-13T01:27:45.132179725Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Dec 13 01:27:45.132635 containerd[1572]: time="2024-12-13T01:27:45.132194163Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Dec 13 01:27:45.132635 containerd[1572]: time="2024-12-13T01:27:45.132210785Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Dec 13 01:27:45.132635 containerd[1572]: time="2024-12-13T01:27:45.132226259Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Dec 13 01:27:45.132635 containerd[1572]: time="2024-12-13T01:27:45.132266803Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Dec 13 01:27:45.132635 containerd[1572]: time="2024-12-13T01:27:45.132286398Z" level=info msg="NRI interface is disabled by configuration." Dec 13 01:27:45.132635 containerd[1572]: time="2024-12-13T01:27:45.132301922Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Dec 13 01:27:45.132958 containerd[1572]: time="2024-12-13T01:27:45.132723689Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:false] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:false SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Dec 13 01:27:45.132958 containerd[1572]: time="2024-12-13T01:27:45.132806715Z" level=info msg="Connect containerd service" Dec 13 01:27:45.132958 containerd[1572]: time="2024-12-13T01:27:45.132848603Z" level=info msg="using legacy CRI server" Dec 13 01:27:45.132958 containerd[1572]: time="2024-12-13T01:27:45.132857893Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Dec 13 01:27:45.135510 containerd[1572]: time="2024-12-13T01:27:45.134605467Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Dec 13 01:27:45.136861 containerd[1572]: time="2024-12-13T01:27:45.136819741Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Dec 13 01:27:45.137155 containerd[1572]: time="2024-12-13T01:27:45.137104821Z" level=info msg="Start subscribing containerd event" Dec 13 01:27:45.137288 containerd[1572]: time="2024-12-13T01:27:45.137177019Z" level=info msg="Start recovering state" Dec 13 01:27:45.137341 containerd[1572]: time="2024-12-13T01:27:45.137290510Z" level=info msg="Start event monitor" Dec 13 01:27:45.137341 containerd[1572]: time="2024-12-13T01:27:45.137318022Z" level=info msg="Start snapshots syncer" Dec 13 01:27:45.137341 containerd[1572]: time="2024-12-13T01:27:45.137330265Z" level=info msg="Start cni network conf syncer for default" Dec 13 01:27:45.137428 containerd[1572]: time="2024-12-13T01:27:45.137344589Z" level=info msg="Start streaming server" Dec 13 01:27:45.137428 containerd[1572]: time="2024-12-13T01:27:45.137396229Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Dec 13 01:27:45.137512 containerd[1572]: time="2024-12-13T01:27:45.137476272Z" level=info msg=serving... address=/run/containerd/containerd.sock Dec 13 01:27:45.137963 containerd[1572]: time="2024-12-13T01:27:45.137575612Z" level=info msg="containerd successfully booted in 0.119801s" Dec 13 01:27:45.137811 systemd[1]: Started containerd.service - containerd container runtime. Dec 13 01:27:46.205152 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 13 01:27:46.207376 systemd[1]: Reached target multi-user.target - Multi-User System. Dec 13 01:27:46.211360 systemd[1]: Startup finished in 7.697s (kernel) + 6.245s (userspace) = 13.943s. Dec 13 01:27:46.212490 (kubelet)[1663]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 13 01:27:47.388635 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Dec 13 01:27:47.401533 systemd[1]: Started sshd@0-10.0.0.50:22-10.0.0.1:43390.service - OpenSSH per-connection server daemon (10.0.0.1:43390). Dec 13 01:27:47.431196 kubelet[1663]: E1213 01:27:47.431068 1663 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 13 01:27:47.436403 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 13 01:27:47.436734 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 13 01:27:47.457184 sshd[1675]: Accepted publickey for core from 10.0.0.1 port 43390 ssh2: RSA SHA256:x0r+OYSWSaRwllGtX4o4H8bWGnkqZzK3ZUwKdtfgOO0 Dec 13 01:27:47.459939 sshd[1675]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 01:27:47.470166 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Dec 13 01:27:47.482378 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Dec 13 01:27:47.484509 systemd-logind[1547]: New session 1 of user core. Dec 13 01:27:47.498775 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Dec 13 01:27:47.507597 systemd[1]: Starting user@500.service - User Manager for UID 500... Dec 13 01:27:47.512577 (systemd)[1683]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Dec 13 01:27:47.663588 systemd[1683]: Queued start job for default target default.target. Dec 13 01:27:47.664153 systemd[1683]: Created slice app.slice - User Application Slice. Dec 13 01:27:47.664175 systemd[1683]: Reached target paths.target - Paths. Dec 13 01:27:47.664189 systemd[1683]: Reached target timers.target - Timers. Dec 13 01:27:47.672303 systemd[1683]: Starting dbus.socket - D-Bus User Message Bus Socket... Dec 13 01:27:47.681542 systemd[1683]: Listening on dbus.socket - D-Bus User Message Bus Socket. Dec 13 01:27:47.681635 systemd[1683]: Reached target sockets.target - Sockets. Dec 13 01:27:47.681649 systemd[1683]: Reached target basic.target - Basic System. Dec 13 01:27:47.681710 systemd[1683]: Reached target default.target - Main User Target. Dec 13 01:27:47.681748 systemd[1683]: Startup finished in 157ms. Dec 13 01:27:47.683082 systemd[1]: Started user@500.service - User Manager for UID 500. Dec 13 01:27:47.693486 systemd[1]: Started session-1.scope - Session 1 of User core. Dec 13 01:27:47.757440 systemd[1]: Started sshd@1-10.0.0.50:22-10.0.0.1:43392.service - OpenSSH per-connection server daemon (10.0.0.1:43392). Dec 13 01:27:47.793578 sshd[1695]: Accepted publickey for core from 10.0.0.1 port 43392 ssh2: RSA SHA256:x0r+OYSWSaRwllGtX4o4H8bWGnkqZzK3ZUwKdtfgOO0 Dec 13 01:27:47.796088 sshd[1695]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 01:27:47.801328 systemd-logind[1547]: New session 2 of user core. Dec 13 01:27:47.811650 systemd[1]: Started session-2.scope - Session 2 of User core. Dec 13 01:27:47.870579 sshd[1695]: pam_unix(sshd:session): session closed for user core Dec 13 01:27:47.887581 systemd[1]: Started sshd@2-10.0.0.50:22-10.0.0.1:43394.service - OpenSSH per-connection server daemon (10.0.0.1:43394). Dec 13 01:27:47.888472 systemd[1]: sshd@1-10.0.0.50:22-10.0.0.1:43392.service: Deactivated successfully. Dec 13 01:27:47.891088 systemd[1]: session-2.scope: Deactivated successfully. Dec 13 01:27:47.892169 systemd-logind[1547]: Session 2 logged out. Waiting for processes to exit. Dec 13 01:27:47.894300 systemd-logind[1547]: Removed session 2. Dec 13 01:27:47.925744 sshd[1700]: Accepted publickey for core from 10.0.0.1 port 43394 ssh2: RSA SHA256:x0r+OYSWSaRwllGtX4o4H8bWGnkqZzK3ZUwKdtfgOO0 Dec 13 01:27:47.928576 sshd[1700]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 01:27:47.935769 systemd-logind[1547]: New session 3 of user core. Dec 13 01:27:47.949708 systemd[1]: Started session-3.scope - Session 3 of User core. Dec 13 01:27:48.006367 sshd[1700]: pam_unix(sshd:session): session closed for user core Dec 13 01:27:48.019568 systemd[1]: Started sshd@3-10.0.0.50:22-10.0.0.1:43402.service - OpenSSH per-connection server daemon (10.0.0.1:43402). Dec 13 01:27:48.020526 systemd[1]: sshd@2-10.0.0.50:22-10.0.0.1:43394.service: Deactivated successfully. Dec 13 01:27:48.022970 systemd[1]: session-3.scope: Deactivated successfully. Dec 13 01:27:48.023766 systemd-logind[1547]: Session 3 logged out. Waiting for processes to exit. Dec 13 01:27:48.025401 systemd-logind[1547]: Removed session 3. Dec 13 01:27:48.053384 sshd[1709]: Accepted publickey for core from 10.0.0.1 port 43402 ssh2: RSA SHA256:x0r+OYSWSaRwllGtX4o4H8bWGnkqZzK3ZUwKdtfgOO0 Dec 13 01:27:48.055169 sshd[1709]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 01:27:48.059641 systemd-logind[1547]: New session 4 of user core. Dec 13 01:27:48.069448 systemd[1]: Started session-4.scope - Session 4 of User core. Dec 13 01:27:48.126694 sshd[1709]: pam_unix(sshd:session): session closed for user core Dec 13 01:27:48.139403 systemd[1]: Started sshd@4-10.0.0.50:22-10.0.0.1:43408.service - OpenSSH per-connection server daemon (10.0.0.1:43408). Dec 13 01:27:48.140244 systemd[1]: sshd@3-10.0.0.50:22-10.0.0.1:43402.service: Deactivated successfully. Dec 13 01:27:48.142574 systemd[1]: session-4.scope: Deactivated successfully. Dec 13 01:27:48.143357 systemd-logind[1547]: Session 4 logged out. Waiting for processes to exit. Dec 13 01:27:48.144733 systemd-logind[1547]: Removed session 4. Dec 13 01:27:48.171737 sshd[1716]: Accepted publickey for core from 10.0.0.1 port 43408 ssh2: RSA SHA256:x0r+OYSWSaRwllGtX4o4H8bWGnkqZzK3ZUwKdtfgOO0 Dec 13 01:27:48.173423 sshd[1716]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 01:27:48.178086 systemd-logind[1547]: New session 5 of user core. Dec 13 01:27:48.192498 systemd[1]: Started session-5.scope - Session 5 of User core. Dec 13 01:27:48.252686 sudo[1723]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Dec 13 01:27:48.253069 sudo[1723]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Dec 13 01:27:48.266273 sudo[1723]: pam_unix(sudo:session): session closed for user root Dec 13 01:27:48.268443 sshd[1716]: pam_unix(sshd:session): session closed for user core Dec 13 01:27:48.277356 systemd[1]: Started sshd@5-10.0.0.50:22-10.0.0.1:43424.service - OpenSSH per-connection server daemon (10.0.0.1:43424). Dec 13 01:27:48.277898 systemd[1]: sshd@4-10.0.0.50:22-10.0.0.1:43408.service: Deactivated successfully. Dec 13 01:27:48.280151 systemd-logind[1547]: Session 5 logged out. Waiting for processes to exit. Dec 13 01:27:48.280935 systemd[1]: session-5.scope: Deactivated successfully. Dec 13 01:27:48.282523 systemd-logind[1547]: Removed session 5. Dec 13 01:27:48.311070 sshd[1725]: Accepted publickey for core from 10.0.0.1 port 43424 ssh2: RSA SHA256:x0r+OYSWSaRwllGtX4o4H8bWGnkqZzK3ZUwKdtfgOO0 Dec 13 01:27:48.312896 sshd[1725]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 01:27:48.318593 systemd-logind[1547]: New session 6 of user core. Dec 13 01:27:48.329358 systemd[1]: Started session-6.scope - Session 6 of User core. Dec 13 01:27:48.467886 sudo[1733]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Dec 13 01:27:48.468466 sudo[1733]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Dec 13 01:27:48.473412 sudo[1733]: pam_unix(sudo:session): session closed for user root Dec 13 01:27:48.481756 sudo[1732]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Dec 13 01:27:48.482156 sudo[1732]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Dec 13 01:27:48.509329 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Dec 13 01:27:48.511527 auditctl[1736]: No rules Dec 13 01:27:48.513228 systemd[1]: audit-rules.service: Deactivated successfully. Dec 13 01:27:48.513595 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Dec 13 01:27:48.515842 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Dec 13 01:27:48.557521 augenrules[1755]: No rules Dec 13 01:27:48.559887 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Dec 13 01:27:48.561386 sudo[1732]: pam_unix(sudo:session): session closed for user root Dec 13 01:27:48.563542 sshd[1725]: pam_unix(sshd:session): session closed for user core Dec 13 01:27:48.572332 systemd[1]: Started sshd@6-10.0.0.50:22-10.0.0.1:43440.service - OpenSSH per-connection server daemon (10.0.0.1:43440). Dec 13 01:27:48.572831 systemd[1]: sshd@5-10.0.0.50:22-10.0.0.1:43424.service: Deactivated successfully. Dec 13 01:27:48.575679 systemd-logind[1547]: Session 6 logged out. Waiting for processes to exit. Dec 13 01:27:48.576605 systemd[1]: session-6.scope: Deactivated successfully. Dec 13 01:27:48.577721 systemd-logind[1547]: Removed session 6. Dec 13 01:27:48.605332 sshd[1761]: Accepted publickey for core from 10.0.0.1 port 43440 ssh2: RSA SHA256:x0r+OYSWSaRwllGtX4o4H8bWGnkqZzK3ZUwKdtfgOO0 Dec 13 01:27:48.607215 sshd[1761]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 01:27:48.612524 systemd-logind[1547]: New session 7 of user core. Dec 13 01:27:48.628533 systemd[1]: Started session-7.scope - Session 7 of User core. Dec 13 01:27:48.685025 sudo[1768]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Dec 13 01:27:48.685404 sudo[1768]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Dec 13 01:27:48.712985 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Dec 13 01:27:48.737303 systemd[1]: coreos-metadata.service: Deactivated successfully. Dec 13 01:27:48.737669 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Dec 13 01:27:49.493511 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Dec 13 01:27:49.503313 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 13 01:27:49.546398 systemd[1]: Reloading requested from client PID 1820 ('systemctl') (unit session-7.scope)... Dec 13 01:27:49.546421 systemd[1]: Reloading... Dec 13 01:27:49.645294 zram_generator::config[1861]: No configuration found. Dec 13 01:27:50.108534 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Dec 13 01:27:50.184806 systemd[1]: Reloading finished in 637 ms. Dec 13 01:27:50.233379 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Dec 13 01:27:50.233503 systemd[1]: kubelet.service: Failed with result 'signal'. Dec 13 01:27:50.233919 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Dec 13 01:27:50.245683 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 13 01:27:50.406519 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 13 01:27:50.413333 (kubelet)[1918]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Dec 13 01:27:50.501060 kubelet[1918]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 13 01:27:50.501060 kubelet[1918]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 13 01:27:50.501060 kubelet[1918]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 13 01:27:50.502356 kubelet[1918]: I1213 01:27:50.502286 1918 server.go:204] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 13 01:27:50.736986 kubelet[1918]: I1213 01:27:50.736810 1918 server.go:487] "Kubelet version" kubeletVersion="v1.29.2" Dec 13 01:27:50.736986 kubelet[1918]: I1213 01:27:50.736862 1918 server.go:489] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 13 01:27:50.737170 kubelet[1918]: I1213 01:27:50.737131 1918 server.go:919] "Client rotation is on, will bootstrap in background" Dec 13 01:27:50.752223 kubelet[1918]: I1213 01:27:50.752156 1918 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Dec 13 01:27:50.771332 kubelet[1918]: I1213 01:27:50.771255 1918 server.go:745] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Dec 13 01:27:50.774616 kubelet[1918]: I1213 01:27:50.774518 1918 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 13 01:27:50.774794 kubelet[1918]: I1213 01:27:50.774771 1918 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} Dec 13 01:27:50.775410 kubelet[1918]: I1213 01:27:50.775373 1918 topology_manager.go:138] "Creating topology manager with none policy" Dec 13 01:27:50.775410 kubelet[1918]: I1213 01:27:50.775395 1918 container_manager_linux.go:301] "Creating device plugin manager" Dec 13 01:27:50.775613 kubelet[1918]: I1213 01:27:50.775591 1918 state_mem.go:36] "Initialized new in-memory state store" Dec 13 01:27:50.775759 kubelet[1918]: I1213 01:27:50.775742 1918 kubelet.go:396] "Attempting to sync node with API server" Dec 13 01:27:50.775838 kubelet[1918]: I1213 01:27:50.775768 1918 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 13 01:27:50.775838 kubelet[1918]: I1213 01:27:50.775820 1918 kubelet.go:312] "Adding apiserver pod source" Dec 13 01:27:50.775893 kubelet[1918]: I1213 01:27:50.775844 1918 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 13 01:27:50.776450 kubelet[1918]: E1213 01:27:50.776330 1918 file.go:98] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:27:50.776450 kubelet[1918]: E1213 01:27:50.776405 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:27:50.777646 kubelet[1918]: I1213 01:27:50.777609 1918 kuberuntime_manager.go:258] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Dec 13 01:27:50.780563 kubelet[1918]: I1213 01:27:50.780526 1918 kubelet.go:809] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 13 01:27:50.783478 kubelet[1918]: W1213 01:27:50.783428 1918 probe.go:268] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Dec 13 01:27:50.784727 kubelet[1918]: I1213 01:27:50.784511 1918 server.go:1256] "Started kubelet" Dec 13 01:27:50.784727 kubelet[1918]: I1213 01:27:50.784629 1918 server.go:162] "Starting to listen" address="0.0.0.0" port=10250 Dec 13 01:27:50.786426 kubelet[1918]: I1213 01:27:50.785834 1918 server.go:461] "Adding debug handlers to kubelet server" Dec 13 01:27:50.786426 kubelet[1918]: I1213 01:27:50.785854 1918 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 13 01:27:50.786594 kubelet[1918]: I1213 01:27:50.786436 1918 server.go:233] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 13 01:27:50.787110 kubelet[1918]: I1213 01:27:50.787085 1918 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 13 01:27:50.788283 kubelet[1918]: I1213 01:27:50.788164 1918 volume_manager.go:291] "Starting Kubelet Volume Manager" Dec 13 01:27:50.788334 kubelet[1918]: I1213 01:27:50.788288 1918 desired_state_of_world_populator.go:151] "Desired state populator starts to run" Dec 13 01:27:50.788896 kubelet[1918]: I1213 01:27:50.788381 1918 reconciler_new.go:29] "Reconciler: start to sync state" Dec 13 01:27:50.790601 kubelet[1918]: I1213 01:27:50.790553 1918 factory.go:221] Registration of the systemd container factory successfully Dec 13 01:27:50.790879 kubelet[1918]: I1213 01:27:50.790719 1918 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Dec 13 01:27:50.792797 kubelet[1918]: I1213 01:27:50.792559 1918 factory.go:221] Registration of the containerd container factory successfully Dec 13 01:27:50.792797 kubelet[1918]: E1213 01:27:50.792710 1918 kubelet.go:1462] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Dec 13 01:27:50.806968 kubelet[1918]: E1213 01:27:50.806905 1918 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"10.0.0.50\" not found" node="10.0.0.50" Dec 13 01:27:50.826795 kubelet[1918]: I1213 01:27:50.826756 1918 cpu_manager.go:214] "Starting CPU manager" policy="none" Dec 13 01:27:50.826795 kubelet[1918]: I1213 01:27:50.826789 1918 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Dec 13 01:27:50.826960 kubelet[1918]: I1213 01:27:50.826842 1918 state_mem.go:36] "Initialized new in-memory state store" Dec 13 01:27:50.889686 kubelet[1918]: I1213 01:27:50.889619 1918 kubelet_node_status.go:73] "Attempting to register node" node="10.0.0.50" Dec 13 01:27:50.957856 kubelet[1918]: I1213 01:27:50.957793 1918 kubelet_node_status.go:76] "Successfully registered node" node="10.0.0.50" Dec 13 01:27:51.184984 kubelet[1918]: E1213 01:27:51.184804 1918 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"10.0.0.50\" not found" Dec 13 01:27:51.228609 kubelet[1918]: I1213 01:27:51.228540 1918 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 13 01:27:51.230828 kubelet[1918]: I1213 01:27:51.230788 1918 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 13 01:27:51.230951 kubelet[1918]: I1213 01:27:51.230851 1918 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 13 01:27:51.230951 kubelet[1918]: I1213 01:27:51.230880 1918 kubelet.go:2329] "Starting kubelet main sync loop" Dec 13 01:27:51.231504 kubelet[1918]: E1213 01:27:51.231059 1918 kubelet.go:2353] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 13 01:27:51.234323 kubelet[1918]: I1213 01:27:51.234261 1918 policy_none.go:49] "None policy: Start" Dec 13 01:27:51.235329 kubelet[1918]: I1213 01:27:51.235278 1918 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 13 01:27:51.235329 kubelet[1918]: I1213 01:27:51.235331 1918 state_mem.go:35] "Initializing new in-memory state store" Dec 13 01:27:51.280437 kubelet[1918]: I1213 01:27:51.280388 1918 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 13 01:27:51.281733 kubelet[1918]: I1213 01:27:51.280798 1918 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 13 01:27:51.286273 kubelet[1918]: I1213 01:27:51.286229 1918 kuberuntime_manager.go:1529] "Updating runtime config through cri with podcidr" CIDR="192.168.1.0/24" Dec 13 01:27:51.286732 containerd[1572]: time="2024-12-13T01:27:51.286651950Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Dec 13 01:27:51.287225 kubelet[1918]: I1213 01:27:51.286991 1918 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.1.0/24" Dec 13 01:27:51.739160 kubelet[1918]: I1213 01:27:51.739077 1918 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Dec 13 01:27:51.739877 kubelet[1918]: W1213 01:27:51.739365 1918 reflector.go:462] vendor/k8s.io/client-go/informers/factory.go:159: watch of *v1.RuntimeClass ended with: very short watch: vendor/k8s.io/client-go/informers/factory.go:159: Unexpected watch close - watch lasted less than a second and no items received Dec 13 01:27:51.739877 kubelet[1918]: W1213 01:27:51.739421 1918 reflector.go:462] vendor/k8s.io/client-go/informers/factory.go:159: watch of *v1.Service ended with: very short watch: vendor/k8s.io/client-go/informers/factory.go:159: Unexpected watch close - watch lasted less than a second and no items received Dec 13 01:27:51.739877 kubelet[1918]: W1213 01:27:51.739458 1918 reflector.go:462] vendor/k8s.io/client-go/informers/factory.go:159: watch of *v1.CSIDriver ended with: very short watch: vendor/k8s.io/client-go/informers/factory.go:159: Unexpected watch close - watch lasted less than a second and no items received Dec 13 01:27:51.776731 kubelet[1918]: I1213 01:27:51.776651 1918 apiserver.go:52] "Watching apiserver" Dec 13 01:27:51.776891 kubelet[1918]: E1213 01:27:51.776564 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:27:51.782706 kubelet[1918]: I1213 01:27:51.782662 1918 topology_manager.go:215] "Topology Admit Handler" podUID="79a81a9d-aa4b-4c92-8850-8bfec6ad81b7" podNamespace="calico-system" podName="calico-node-qpvfj" Dec 13 01:27:51.782910 kubelet[1918]: I1213 01:27:51.782868 1918 topology_manager.go:215] "Topology Admit Handler" podUID="8e158755-5b09-4a6a-ba0b-890e0f7dd097" podNamespace="calico-system" podName="csi-node-driver-nmw4l" Dec 13 01:27:51.783016 kubelet[1918]: I1213 01:27:51.782946 1918 topology_manager.go:215] "Topology Admit Handler" podUID="9b32327b-9846-41c7-b0ae-3c59d791fdc9" podNamespace="kube-system" podName="kube-proxy-j2vgr" Dec 13 01:27:51.783535 kubelet[1918]: E1213 01:27:51.783504 1918 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-nmw4l" podUID="8e158755-5b09-4a6a-ba0b-890e0f7dd097" Dec 13 01:27:51.939623 kubelet[1918]: I1213 01:27:51.939245 1918 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/79a81a9d-aa4b-4c92-8850-8bfec6ad81b7-lib-modules\") pod \"calico-node-qpvfj\" (UID: \"79a81a9d-aa4b-4c92-8850-8bfec6ad81b7\") " pod="calico-system/calico-node-qpvfj" Dec 13 01:27:51.939777 kubelet[1918]: I1213 01:27:51.939736 1918 desired_state_of_world_populator.go:159] "Finished populating initial desired state of world" Dec 13 01:27:51.954401 sudo[1768]: pam_unix(sudo:session): session closed for user root Dec 13 01:27:51.956927 sshd[1761]: pam_unix(sshd:session): session closed for user core Dec 13 01:27:51.960359 systemd[1]: sshd@6-10.0.0.50:22-10.0.0.1:43440.service: Deactivated successfully. Dec 13 01:27:51.964884 systemd[1]: session-7.scope: Deactivated successfully. Dec 13 01:27:51.966302 systemd-logind[1547]: Session 7 logged out. Waiting for processes to exit. Dec 13 01:27:51.968450 systemd-logind[1547]: Removed session 7. Dec 13 01:27:52.040438 kubelet[1918]: I1213 01:27:52.040203 1918 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/79a81a9d-aa4b-4c92-8850-8bfec6ad81b7-xtables-lock\") pod \"calico-node-qpvfj\" (UID: \"79a81a9d-aa4b-4c92-8850-8bfec6ad81b7\") " pod="calico-system/calico-node-qpvfj" Dec 13 01:27:52.040438 kubelet[1918]: I1213 01:27:52.040288 1918 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8e158755-5b09-4a6a-ba0b-890e0f7dd097-kubelet-dir\") pod \"csi-node-driver-nmw4l\" (UID: \"8e158755-5b09-4a6a-ba0b-890e0f7dd097\") " pod="calico-system/csi-node-driver-nmw4l" Dec 13 01:27:52.040438 kubelet[1918]: I1213 01:27:52.040322 1918 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/8e158755-5b09-4a6a-ba0b-890e0f7dd097-socket-dir\") pod \"csi-node-driver-nmw4l\" (UID: \"8e158755-5b09-4a6a-ba0b-890e0f7dd097\") " pod="calico-system/csi-node-driver-nmw4l" Dec 13 01:27:52.040438 kubelet[1918]: I1213 01:27:52.040357 1918 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/79a81a9d-aa4b-4c92-8850-8bfec6ad81b7-tigera-ca-bundle\") pod \"calico-node-qpvfj\" (UID: \"79a81a9d-aa4b-4c92-8850-8bfec6ad81b7\") " pod="calico-system/calico-node-qpvfj" Dec 13 01:27:52.040438 kubelet[1918]: I1213 01:27:52.040422 1918 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/79a81a9d-aa4b-4c92-8850-8bfec6ad81b7-var-run-calico\") pod \"calico-node-qpvfj\" (UID: \"79a81a9d-aa4b-4c92-8850-8bfec6ad81b7\") " pod="calico-system/calico-node-qpvfj" Dec 13 01:27:52.040750 kubelet[1918]: I1213 01:27:52.040467 1918 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8lrs\" (UniqueName: \"kubernetes.io/projected/79a81a9d-aa4b-4c92-8850-8bfec6ad81b7-kube-api-access-p8lrs\") pod \"calico-node-qpvfj\" (UID: \"79a81a9d-aa4b-4c92-8850-8bfec6ad81b7\") " pod="calico-system/calico-node-qpvfj" Dec 13 01:27:52.040750 kubelet[1918]: I1213 01:27:52.040499 1918 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nfhg\" (UniqueName: \"kubernetes.io/projected/8e158755-5b09-4a6a-ba0b-890e0f7dd097-kube-api-access-2nfhg\") pod \"csi-node-driver-nmw4l\" (UID: \"8e158755-5b09-4a6a-ba0b-890e0f7dd097\") " pod="calico-system/csi-node-driver-nmw4l" Dec 13 01:27:52.040750 kubelet[1918]: I1213 01:27:52.040602 1918 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/79a81a9d-aa4b-4c92-8850-8bfec6ad81b7-cni-net-dir\") pod \"calico-node-qpvfj\" (UID: \"79a81a9d-aa4b-4c92-8850-8bfec6ad81b7\") " pod="calico-system/calico-node-qpvfj" Dec 13 01:27:52.040750 kubelet[1918]: I1213 01:27:52.040638 1918 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/79a81a9d-aa4b-4c92-8850-8bfec6ad81b7-cni-log-dir\") pod \"calico-node-qpvfj\" (UID: \"79a81a9d-aa4b-4c92-8850-8bfec6ad81b7\") " pod="calico-system/calico-node-qpvfj" Dec 13 01:27:52.040750 kubelet[1918]: I1213 01:27:52.040705 1918 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/79a81a9d-aa4b-4c92-8850-8bfec6ad81b7-flexvol-driver-host\") pod \"calico-node-qpvfj\" (UID: \"79a81a9d-aa4b-4c92-8850-8bfec6ad81b7\") " pod="calico-system/calico-node-qpvfj" Dec 13 01:27:52.040929 kubelet[1918]: I1213 01:27:52.040750 1918 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/8e158755-5b09-4a6a-ba0b-890e0f7dd097-varrun\") pod \"csi-node-driver-nmw4l\" (UID: \"8e158755-5b09-4a6a-ba0b-890e0f7dd097\") " pod="calico-system/csi-node-driver-nmw4l" Dec 13 01:27:52.040929 kubelet[1918]: I1213 01:27:52.040781 1918 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/8e158755-5b09-4a6a-ba0b-890e0f7dd097-registration-dir\") pod \"csi-node-driver-nmw4l\" (UID: \"8e158755-5b09-4a6a-ba0b-890e0f7dd097\") " pod="calico-system/csi-node-driver-nmw4l" Dec 13 01:27:52.040929 kubelet[1918]: I1213 01:27:52.040812 1918 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/9b32327b-9846-41c7-b0ae-3c59d791fdc9-xtables-lock\") pod \"kube-proxy-j2vgr\" (UID: \"9b32327b-9846-41c7-b0ae-3c59d791fdc9\") " pod="kube-system/kube-proxy-j2vgr" Dec 13 01:27:52.040929 kubelet[1918]: I1213 01:27:52.040846 1918 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/79a81a9d-aa4b-4c92-8850-8bfec6ad81b7-policysync\") pod \"calico-node-qpvfj\" (UID: \"79a81a9d-aa4b-4c92-8850-8bfec6ad81b7\") " pod="calico-system/calico-node-qpvfj" Dec 13 01:27:52.040929 kubelet[1918]: I1213 01:27:52.040876 1918 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/79a81a9d-aa4b-4c92-8850-8bfec6ad81b7-node-certs\") pod \"calico-node-qpvfj\" (UID: \"79a81a9d-aa4b-4c92-8850-8bfec6ad81b7\") " pod="calico-system/calico-node-qpvfj" Dec 13 01:27:52.041135 kubelet[1918]: I1213 01:27:52.040916 1918 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/79a81a9d-aa4b-4c92-8850-8bfec6ad81b7-var-lib-calico\") pod \"calico-node-qpvfj\" (UID: \"79a81a9d-aa4b-4c92-8850-8bfec6ad81b7\") " pod="calico-system/calico-node-qpvfj" Dec 13 01:27:52.041135 kubelet[1918]: I1213 01:27:52.040949 1918 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/79a81a9d-aa4b-4c92-8850-8bfec6ad81b7-cni-bin-dir\") pod \"calico-node-qpvfj\" (UID: \"79a81a9d-aa4b-4c92-8850-8bfec6ad81b7\") " pod="calico-system/calico-node-qpvfj" Dec 13 01:27:52.041135 kubelet[1918]: I1213 01:27:52.040980 1918 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/9b32327b-9846-41c7-b0ae-3c59d791fdc9-kube-proxy\") pod \"kube-proxy-j2vgr\" (UID: \"9b32327b-9846-41c7-b0ae-3c59d791fdc9\") " pod="kube-system/kube-proxy-j2vgr" Dec 13 01:27:52.041135 kubelet[1918]: I1213 01:27:52.041011 1918 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9b32327b-9846-41c7-b0ae-3c59d791fdc9-lib-modules\") pod \"kube-proxy-j2vgr\" (UID: \"9b32327b-9846-41c7-b0ae-3c59d791fdc9\") " pod="kube-system/kube-proxy-j2vgr" Dec 13 01:27:52.041135 kubelet[1918]: I1213 01:27:52.041063 1918 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dlp8\" (UniqueName: \"kubernetes.io/projected/9b32327b-9846-41c7-b0ae-3c59d791fdc9-kube-api-access-7dlp8\") pod \"kube-proxy-j2vgr\" (UID: \"9b32327b-9846-41c7-b0ae-3c59d791fdc9\") " pod="kube-system/kube-proxy-j2vgr" Dec 13 01:27:52.143958 kubelet[1918]: E1213 01:27:52.143858 1918 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 13 01:27:52.143958 kubelet[1918]: W1213 01:27:52.143894 1918 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 13 01:27:52.143958 kubelet[1918]: E1213 01:27:52.143925 1918 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 13 01:27:52.146412 kubelet[1918]: E1213 01:27:52.146357 1918 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 13 01:27:52.146412 kubelet[1918]: W1213 01:27:52.146389 1918 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 13 01:27:52.146412 kubelet[1918]: E1213 01:27:52.146416 1918 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 13 01:27:52.173527 kubelet[1918]: E1213 01:27:52.173377 1918 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 13 01:27:52.173527 kubelet[1918]: W1213 01:27:52.173403 1918 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 13 01:27:52.173527 kubelet[1918]: E1213 01:27:52.173428 1918 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 13 01:27:52.173856 kubelet[1918]: E1213 01:27:52.173827 1918 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 13 01:27:52.173902 kubelet[1918]: W1213 01:27:52.173868 1918 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 13 01:27:52.173925 kubelet[1918]: E1213 01:27:52.173903 1918 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 13 01:27:52.174441 kubelet[1918]: E1213 01:27:52.174391 1918 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 13 01:27:52.174441 kubelet[1918]: W1213 01:27:52.174422 1918 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 13 01:27:52.174441 kubelet[1918]: E1213 01:27:52.174453 1918 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 13 01:27:52.392327 kubelet[1918]: E1213 01:27:52.391276 1918 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 01:27:52.392327 kubelet[1918]: E1213 01:27:52.391708 1918 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 01:27:52.394513 containerd[1572]: time="2024-12-13T01:27:52.392694516Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-qpvfj,Uid:79a81a9d-aa4b-4c92-8850-8bfec6ad81b7,Namespace:calico-system,Attempt:0,}" Dec 13 01:27:52.394513 containerd[1572]: time="2024-12-13T01:27:52.392808479Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-j2vgr,Uid:9b32327b-9846-41c7-b0ae-3c59d791fdc9,Namespace:kube-system,Attempt:0,}" Dec 13 01:27:52.777701 kubelet[1918]: E1213 01:27:52.777477 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:27:53.231821 kubelet[1918]: E1213 01:27:53.231347 1918 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-nmw4l" podUID="8e158755-5b09-4a6a-ba0b-890e0f7dd097" Dec 13 01:27:53.675506 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2191435422.mount: Deactivated successfully. Dec 13 01:27:53.686366 containerd[1572]: time="2024-12-13T01:27:53.686287261Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Dec 13 01:27:53.687502 containerd[1572]: time="2024-12-13T01:27:53.687395873Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Dec 13 01:27:53.688822 containerd[1572]: time="2024-12-13T01:27:53.688735454Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Dec 13 01:27:53.689907 containerd[1572]: time="2024-12-13T01:27:53.689840057Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312056" Dec 13 01:27:53.690844 containerd[1572]: time="2024-12-13T01:27:53.690806155Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Dec 13 01:27:53.693894 containerd[1572]: time="2024-12-13T01:27:53.693855083Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Dec 13 01:27:53.694851 containerd[1572]: time="2024-12-13T01:27:53.694807828Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 1.301800291s" Dec 13 01:27:53.697093 containerd[1572]: time="2024-12-13T01:27:53.697047132Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 1.304224229s" Dec 13 01:27:53.777892 kubelet[1918]: E1213 01:27:53.777820 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:27:53.820639 containerd[1572]: time="2024-12-13T01:27:53.820391148Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 01:27:53.820639 containerd[1572]: time="2024-12-13T01:27:53.820441051Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 01:27:53.820639 containerd[1572]: time="2024-12-13T01:27:53.820452282Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 01:27:53.820639 containerd[1572]: time="2024-12-13T01:27:53.820531927Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 01:27:53.820639 containerd[1572]: time="2024-12-13T01:27:53.820345023Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 01:27:53.820639 containerd[1572]: time="2024-12-13T01:27:53.820403945Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 01:27:53.820639 containerd[1572]: time="2024-12-13T01:27:53.820418448Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 01:27:53.820639 containerd[1572]: time="2024-12-13T01:27:53.820521999Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 01:27:54.080786 containerd[1572]: time="2024-12-13T01:27:54.080335538Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-qpvfj,Uid:79a81a9d-aa4b-4c92-8850-8bfec6ad81b7,Namespace:calico-system,Attempt:0,} returns sandbox id \"56634de33987afcd6f9fcbc73b443bfb032f311c4af438bd0256fdd5ec9dc242\"" Dec 13 01:27:54.082156 containerd[1572]: time="2024-12-13T01:27:54.082104570Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-j2vgr,Uid:9b32327b-9846-41c7-b0ae-3c59d791fdc9,Namespace:kube-system,Attempt:0,} returns sandbox id \"97232f960b170bf74dab418bd9420743cb749c77fddabf9ed0c9b819d18b0788\"" Dec 13 01:27:54.082469 kubelet[1918]: E1213 01:27:54.082426 1918 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 01:27:54.083172 kubelet[1918]: E1213 01:27:54.083000 1918 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 01:27:54.084412 containerd[1572]: time="2024-12-13T01:27:54.084369420Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\"" Dec 13 01:27:54.779009 kubelet[1918]: E1213 01:27:54.778927 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:27:55.231874 kubelet[1918]: E1213 01:27:55.231712 1918 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-nmw4l" podUID="8e158755-5b09-4a6a-ba0b-890e0f7dd097" Dec 13 01:27:55.760973 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2918681207.mount: Deactivated successfully. Dec 13 01:27:55.779401 kubelet[1918]: E1213 01:27:55.779362 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:27:55.935657 containerd[1572]: time="2024-12-13T01:27:55.935598869Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 01:27:55.990766 containerd[1572]: time="2024-12-13T01:27:55.990652497Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1: active requests=0, bytes read=6855343" Dec 13 01:27:56.033612 containerd[1572]: time="2024-12-13T01:27:56.033468403Z" level=info msg="ImageCreate event name:\"sha256:2b7452b763ec8833ca0386ada5fd066e552a9b3b02b8538a5e34cc3d6d3840a6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 01:27:56.082440 containerd[1572]: time="2024-12-13T01:27:56.082380112Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:a63f8b4ff531912d12d143664eb263fdbc6cd7b3ff4aa777dfb6e318a090462c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 01:27:56.082884 containerd[1572]: time="2024-12-13T01:27:56.082851334Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\" with image id \"sha256:2b7452b763ec8833ca0386ada5fd066e552a9b3b02b8538a5e34cc3d6d3840a6\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:a63f8b4ff531912d12d143664eb263fdbc6cd7b3ff4aa777dfb6e318a090462c\", size \"6855165\" in 1.998205679s" Dec 13 01:27:56.082926 containerd[1572]: time="2024-12-13T01:27:56.082889622Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\" returns image reference \"sha256:2b7452b763ec8833ca0386ada5fd066e552a9b3b02b8538a5e34cc3d6d3840a6\"" Dec 13 01:27:56.083656 containerd[1572]: time="2024-12-13T01:27:56.083606642Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.29.12\"" Dec 13 01:27:56.084916 containerd[1572]: time="2024-12-13T01:27:56.084878873Z" level=info msg="CreateContainer within sandbox \"56634de33987afcd6f9fcbc73b443bfb032f311c4af438bd0256fdd5ec9dc242\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Dec 13 01:27:56.119888 containerd[1572]: time="2024-12-13T01:27:56.119831422Z" level=info msg="CreateContainer within sandbox \"56634de33987afcd6f9fcbc73b443bfb032f311c4af438bd0256fdd5ec9dc242\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"5f5356c2aff5789f0f14c412a44a63adc56ec36a31f81cb2ee6476e3f91a1ace\"" Dec 13 01:27:56.120608 containerd[1572]: time="2024-12-13T01:27:56.120580656Z" level=info msg="StartContainer for \"5f5356c2aff5789f0f14c412a44a63adc56ec36a31f81cb2ee6476e3f91a1ace\"" Dec 13 01:27:56.264694 containerd[1572]: time="2024-12-13T01:27:56.264646712Z" level=info msg="StartContainer for \"5f5356c2aff5789f0f14c412a44a63adc56ec36a31f81cb2ee6476e3f91a1ace\" returns successfully" Dec 13 01:27:56.550983 containerd[1572]: time="2024-12-13T01:27:56.550903256Z" level=info msg="shim disconnected" id=5f5356c2aff5789f0f14c412a44a63adc56ec36a31f81cb2ee6476e3f91a1ace namespace=k8s.io Dec 13 01:27:56.550983 containerd[1572]: time="2024-12-13T01:27:56.550970655Z" level=warning msg="cleaning up after shim disconnected" id=5f5356c2aff5789f0f14c412a44a63adc56ec36a31f81cb2ee6476e3f91a1ace namespace=k8s.io Dec 13 01:27:56.550983 containerd[1572]: time="2024-12-13T01:27:56.550981171Z" level=info msg="cleaning up dead shim" namespace=k8s.io Dec 13 01:27:56.737348 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-5f5356c2aff5789f0f14c412a44a63adc56ec36a31f81cb2ee6476e3f91a1ace-rootfs.mount: Deactivated successfully. Dec 13 01:27:56.779791 kubelet[1918]: E1213 01:27:56.779718 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:27:57.282747 kubelet[1918]: E1213 01:27:57.282702 1918 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-nmw4l" podUID="8e158755-5b09-4a6a-ba0b-890e0f7dd097" Dec 13 01:27:57.287704 kubelet[1918]: E1213 01:27:57.286675 1918 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 01:27:57.533680 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1119258929.mount: Deactivated successfully. Dec 13 01:27:57.780795 kubelet[1918]: E1213 01:27:57.780741 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:27:57.989830 containerd[1572]: time="2024-12-13T01:27:57.989690210Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.29.12\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 01:27:57.990708 containerd[1572]: time="2024-12-13T01:27:57.990668662Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.29.12: active requests=0, bytes read=28619958" Dec 13 01:27:57.991789 containerd[1572]: time="2024-12-13T01:27:57.991761081Z" level=info msg="ImageCreate event name:\"sha256:d699d5830022f9e67c3271d1c2af58eaede81e3567df82728b7d2a8bf12ed153\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 01:27:57.993954 containerd[1572]: time="2024-12-13T01:27:57.993907770Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:bc761494b78fa152a759457f42bc9b86ee9d18f5929bb127bd5f72f8e2112c39\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 01:27:57.994487 containerd[1572]: time="2024-12-13T01:27:57.994450889Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.29.12\" with image id \"sha256:d699d5830022f9e67c3271d1c2af58eaede81e3567df82728b7d2a8bf12ed153\", repo tag \"registry.k8s.io/kube-proxy:v1.29.12\", repo digest \"registry.k8s.io/kube-proxy@sha256:bc761494b78fa152a759457f42bc9b86ee9d18f5929bb127bd5f72f8e2112c39\", size \"28618977\" in 1.9107998s" Dec 13 01:27:57.994542 containerd[1572]: time="2024-12-13T01:27:57.994492493Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.29.12\" returns image reference \"sha256:d699d5830022f9e67c3271d1c2af58eaede81e3567df82728b7d2a8bf12ed153\"" Dec 13 01:27:57.994991 containerd[1572]: time="2024-12-13T01:27:57.994968673Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.1\"" Dec 13 01:27:57.996243 containerd[1572]: time="2024-12-13T01:27:57.996166344Z" level=info msg="CreateContainer within sandbox \"97232f960b170bf74dab418bd9420743cb749c77fddabf9ed0c9b819d18b0788\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Dec 13 01:27:58.010907 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1151476676.mount: Deactivated successfully. Dec 13 01:27:58.012622 containerd[1572]: time="2024-12-13T01:27:58.012586866Z" level=info msg="CreateContainer within sandbox \"97232f960b170bf74dab418bd9420743cb749c77fddabf9ed0c9b819d18b0788\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"33c75e38b7d8dd4e3dad9cebfdbcdb50c11b2cc2b2fcd74ed4e2e89bb33f1146\"" Dec 13 01:27:58.013123 containerd[1572]: time="2024-12-13T01:27:58.013086659Z" level=info msg="StartContainer for \"33c75e38b7d8dd4e3dad9cebfdbcdb50c11b2cc2b2fcd74ed4e2e89bb33f1146\"" Dec 13 01:27:58.072300 containerd[1572]: time="2024-12-13T01:27:58.072252122Z" level=info msg="StartContainer for \"33c75e38b7d8dd4e3dad9cebfdbcdb50c11b2cc2b2fcd74ed4e2e89bb33f1146\" returns successfully" Dec 13 01:27:58.290504 kubelet[1918]: E1213 01:27:58.290353 1918 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 01:27:58.782089 kubelet[1918]: E1213 01:27:58.781845 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:27:59.232154 kubelet[1918]: E1213 01:27:59.231969 1918 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-nmw4l" podUID="8e158755-5b09-4a6a-ba0b-890e0f7dd097" Dec 13 01:27:59.291369 kubelet[1918]: E1213 01:27:59.291316 1918 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 01:27:59.782965 kubelet[1918]: E1213 01:27:59.782888 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:00.783976 kubelet[1918]: E1213 01:28:00.783940 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:01.232643 kubelet[1918]: E1213 01:28:01.232516 1918 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-nmw4l" podUID="8e158755-5b09-4a6a-ba0b-890e0f7dd097" Dec 13 01:28:01.785195 kubelet[1918]: E1213 01:28:01.785108 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:02.759447 containerd[1572]: time="2024-12-13T01:28:02.759383912Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 01:28:02.760320 containerd[1572]: time="2024-12-13T01:28:02.760223605Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.29.1: active requests=0, bytes read=96154154" Dec 13 01:28:02.761554 containerd[1572]: time="2024-12-13T01:28:02.761513458Z" level=info msg="ImageCreate event name:\"sha256:7dd6ea186aba0d7a1791a79d426fe854527ca95192b26bbd19e8baf8373f7d0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 01:28:02.764475 containerd[1572]: time="2024-12-13T01:28:02.764436217Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:21e759d51c90dfb34fc1397dc180dd3a3fb564c2b0580d2f61ffe108f2a3c94b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 01:28:02.765199 containerd[1572]: time="2024-12-13T01:28:02.765168649Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.29.1\" with image id \"sha256:7dd6ea186aba0d7a1791a79d426fe854527ca95192b26bbd19e8baf8373f7d0e\", repo tag \"ghcr.io/flatcar/calico/cni:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:21e759d51c90dfb34fc1397dc180dd3a3fb564c2b0580d2f61ffe108f2a3c94b\", size \"97647238\" in 4.77007683s" Dec 13 01:28:02.765245 containerd[1572]: time="2024-12-13T01:28:02.765197101Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.1\" returns image reference \"sha256:7dd6ea186aba0d7a1791a79d426fe854527ca95192b26bbd19e8baf8373f7d0e\"" Dec 13 01:28:02.767046 containerd[1572]: time="2024-12-13T01:28:02.766999564Z" level=info msg="CreateContainer within sandbox \"56634de33987afcd6f9fcbc73b443bfb032f311c4af438bd0256fdd5ec9dc242\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Dec 13 01:28:02.783551 containerd[1572]: time="2024-12-13T01:28:02.783503697Z" level=info msg="CreateContainer within sandbox \"56634de33987afcd6f9fcbc73b443bfb032f311c4af438bd0256fdd5ec9dc242\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"dd5b17d0887cfeeee55e5fa92bc37aa63415899ff9b75f82d57e2b88ee5e213f\"" Dec 13 01:28:02.784136 containerd[1572]: time="2024-12-13T01:28:02.784089280Z" level=info msg="StartContainer for \"dd5b17d0887cfeeee55e5fa92bc37aa63415899ff9b75f82d57e2b88ee5e213f\"" Dec 13 01:28:02.785274 kubelet[1918]: E1213 01:28:02.785248 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:02.875520 containerd[1572]: time="2024-12-13T01:28:02.875467022Z" level=info msg="StartContainer for \"dd5b17d0887cfeeee55e5fa92bc37aa63415899ff9b75f82d57e2b88ee5e213f\" returns successfully" Dec 13 01:28:03.231589 kubelet[1918]: E1213 01:28:03.231485 1918 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-nmw4l" podUID="8e158755-5b09-4a6a-ba0b-890e0f7dd097" Dec 13 01:28:03.298821 kubelet[1918]: E1213 01:28:03.298774 1918 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 01:28:03.347095 kubelet[1918]: I1213 01:28:03.347006 1918 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-proxy-j2vgr" podStartSLOduration=8.437463889 podStartE2EDuration="12.346946501s" podCreationTimestamp="2024-12-13 01:27:51 +0000 UTC" firstStartedPulling="2024-12-13 01:27:54.085305746 +0000 UTC m=+3.662686021" lastFinishedPulling="2024-12-13 01:27:57.994788367 +0000 UTC m=+7.572168633" observedRunningTime="2024-12-13 01:27:58.305395176 +0000 UTC m=+7.882775451" watchObservedRunningTime="2024-12-13 01:28:03.346946501 +0000 UTC m=+12.924326766" Dec 13 01:28:03.786242 kubelet[1918]: E1213 01:28:03.786181 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:04.300451 kubelet[1918]: E1213 01:28:04.300411 1918 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 01:28:04.786864 kubelet[1918]: E1213 01:28:04.786804 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:05.232998 kubelet[1918]: E1213 01:28:05.232107 1918 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-nmw4l" podUID="8e158755-5b09-4a6a-ba0b-890e0f7dd097" Dec 13 01:28:05.787386 kubelet[1918]: E1213 01:28:05.787310 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:05.906210 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-dd5b17d0887cfeeee55e5fa92bc37aa63415899ff9b75f82d57e2b88ee5e213f-rootfs.mount: Deactivated successfully. Dec 13 01:28:05.911663 kubelet[1918]: I1213 01:28:05.911628 1918 kubelet_node_status.go:497] "Fast updating node status as it just became ready" Dec 13 01:28:06.654085 containerd[1572]: time="2024-12-13T01:28:06.653996493Z" level=info msg="shim disconnected" id=dd5b17d0887cfeeee55e5fa92bc37aa63415899ff9b75f82d57e2b88ee5e213f namespace=k8s.io Dec 13 01:28:06.654085 containerd[1572]: time="2024-12-13T01:28:06.654076176Z" level=warning msg="cleaning up after shim disconnected" id=dd5b17d0887cfeeee55e5fa92bc37aa63415899ff9b75f82d57e2b88ee5e213f namespace=k8s.io Dec 13 01:28:06.654085 containerd[1572]: time="2024-12-13T01:28:06.654085356Z" level=info msg="cleaning up dead shim" namespace=k8s.io Dec 13 01:28:06.787816 kubelet[1918]: E1213 01:28:06.787709 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:07.234824 containerd[1572]: time="2024-12-13T01:28:07.234761909Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-nmw4l,Uid:8e158755-5b09-4a6a-ba0b-890e0f7dd097,Namespace:calico-system,Attempt:0,}" Dec 13 01:28:07.300139 containerd[1572]: time="2024-12-13T01:28:07.300067848Z" level=error msg="Failed to destroy network for sandbox \"1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 13 01:28:07.302230 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30-shm.mount: Deactivated successfully. Dec 13 01:28:07.302900 containerd[1572]: time="2024-12-13T01:28:07.302826512Z" level=error msg="encountered an error cleaning up failed sandbox \"1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 13 01:28:07.302957 containerd[1572]: time="2024-12-13T01:28:07.302904745Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-nmw4l,Uid:8e158755-5b09-4a6a-ba0b-890e0f7dd097,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 13 01:28:07.303271 kubelet[1918]: E1213 01:28:07.303239 1918 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 13 01:28:07.303330 kubelet[1918]: E1213 01:28:07.303317 1918 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-nmw4l" Dec 13 01:28:07.303358 kubelet[1918]: E1213 01:28:07.303339 1918 kuberuntime_manager.go:1172] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-nmw4l" Dec 13 01:28:07.303462 kubelet[1918]: E1213 01:28:07.303423 1918 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-nmw4l_calico-system(8e158755-5b09-4a6a-ba0b-890e0f7dd097)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-nmw4l_calico-system(8e158755-5b09-4a6a-ba0b-890e0f7dd097)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-nmw4l" podUID="8e158755-5b09-4a6a-ba0b-890e0f7dd097" Dec 13 01:28:07.305524 kubelet[1918]: I1213 01:28:07.305503 1918 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30" Dec 13 01:28:07.306195 containerd[1572]: time="2024-12-13T01:28:07.306161521Z" level=info msg="StopPodSandbox for \"1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30\"" Dec 13 01:28:07.306390 containerd[1572]: time="2024-12-13T01:28:07.306359598Z" level=info msg="Ensure that sandbox 1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30 in task-service has been cleanup successfully" Dec 13 01:28:07.307510 kubelet[1918]: E1213 01:28:07.307489 1918 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 01:28:07.308492 containerd[1572]: time="2024-12-13T01:28:07.308468899Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.1\"" Dec 13 01:28:07.335215 containerd[1572]: time="2024-12-13T01:28:07.335156901Z" level=error msg="StopPodSandbox for \"1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30\" failed" error="failed to destroy network for sandbox \"1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 13 01:28:07.335459 kubelet[1918]: E1213 01:28:07.335422 1918 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30" Dec 13 01:28:07.335527 kubelet[1918]: E1213 01:28:07.335516 1918 kuberuntime_manager.go:1381] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30"} Dec 13 01:28:07.335567 kubelet[1918]: E1213 01:28:07.335557 1918 kuberuntime_manager.go:1081] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"8e158755-5b09-4a6a-ba0b-890e0f7dd097\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Dec 13 01:28:07.335634 kubelet[1918]: E1213 01:28:07.335591 1918 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"8e158755-5b09-4a6a-ba0b-890e0f7dd097\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-nmw4l" podUID="8e158755-5b09-4a6a-ba0b-890e0f7dd097" Dec 13 01:28:07.788378 kubelet[1918]: E1213 01:28:07.788297 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:08.455824 kubelet[1918]: I1213 01:28:08.455784 1918 topology_manager.go:215] "Topology Admit Handler" podUID="2767e440-abcd-45e1-95b3-0943b03016cf" podNamespace="default" podName="nginx-deployment-6d5f899847-785tc" Dec 13 01:28:08.624166 kubelet[1918]: I1213 01:28:08.624095 1918 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45wkg\" (UniqueName: \"kubernetes.io/projected/2767e440-abcd-45e1-95b3-0943b03016cf-kube-api-access-45wkg\") pod \"nginx-deployment-6d5f899847-785tc\" (UID: \"2767e440-abcd-45e1-95b3-0943b03016cf\") " pod="default/nginx-deployment-6d5f899847-785tc" Dec 13 01:28:08.789494 kubelet[1918]: E1213 01:28:08.789270 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:09.359782 containerd[1572]: time="2024-12-13T01:28:09.359696623Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-6d5f899847-785tc,Uid:2767e440-abcd-45e1-95b3-0943b03016cf,Namespace:default,Attempt:0,}" Dec 13 01:28:09.478693 containerd[1572]: time="2024-12-13T01:28:09.478591910Z" level=error msg="Failed to destroy network for sandbox \"eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 13 01:28:09.480778 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e-shm.mount: Deactivated successfully. Dec 13 01:28:09.481560 containerd[1572]: time="2024-12-13T01:28:09.481504407Z" level=error msg="encountered an error cleaning up failed sandbox \"eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 13 01:28:09.481632 containerd[1572]: time="2024-12-13T01:28:09.481560577Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-6d5f899847-785tc,Uid:2767e440-abcd-45e1-95b3-0943b03016cf,Namespace:default,Attempt:0,} failed, error" error="failed to setup network for sandbox \"eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 13 01:28:09.482325 kubelet[1918]: E1213 01:28:09.482277 1918 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 13 01:28:09.482617 kubelet[1918]: E1213 01:28:09.482347 1918 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="default/nginx-deployment-6d5f899847-785tc" Dec 13 01:28:09.482617 kubelet[1918]: E1213 01:28:09.482370 1918 kuberuntime_manager.go:1172] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="default/nginx-deployment-6d5f899847-785tc" Dec 13 01:28:09.482617 kubelet[1918]: E1213 01:28:09.482425 1918 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"nginx-deployment-6d5f899847-785tc_default(2767e440-abcd-45e1-95b3-0943b03016cf)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"nginx-deployment-6d5f899847-785tc_default(2767e440-abcd-45e1-95b3-0943b03016cf)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="default/nginx-deployment-6d5f899847-785tc" podUID="2767e440-abcd-45e1-95b3-0943b03016cf" Dec 13 01:28:09.790312 kubelet[1918]: E1213 01:28:09.790105 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:10.325630 kubelet[1918]: I1213 01:28:10.325559 1918 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e" Dec 13 01:28:10.327099 containerd[1572]: time="2024-12-13T01:28:10.326894205Z" level=info msg="StopPodSandbox for \"eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e\"" Dec 13 01:28:10.327272 containerd[1572]: time="2024-12-13T01:28:10.327226273Z" level=info msg="Ensure that sandbox eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e in task-service has been cleanup successfully" Dec 13 01:28:10.375108 containerd[1572]: time="2024-12-13T01:28:10.375014457Z" level=error msg="StopPodSandbox for \"eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e\" failed" error="failed to destroy network for sandbox \"eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 13 01:28:10.375707 kubelet[1918]: E1213 01:28:10.375404 1918 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e" Dec 13 01:28:10.375707 kubelet[1918]: E1213 01:28:10.375483 1918 kuberuntime_manager.go:1381] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e"} Dec 13 01:28:10.375707 kubelet[1918]: E1213 01:28:10.375540 1918 kuberuntime_manager.go:1081] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"2767e440-abcd-45e1-95b3-0943b03016cf\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Dec 13 01:28:10.375707 kubelet[1918]: E1213 01:28:10.375580 1918 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"2767e440-abcd-45e1-95b3-0943b03016cf\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="default/nginx-deployment-6d5f899847-785tc" podUID="2767e440-abcd-45e1-95b3-0943b03016cf" Dec 13 01:28:10.776420 kubelet[1918]: E1213 01:28:10.776167 1918 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:10.790708 kubelet[1918]: E1213 01:28:10.790630 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:11.791093 kubelet[1918]: E1213 01:28:11.790990 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:12.791446 kubelet[1918]: E1213 01:28:12.791376 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:13.792424 kubelet[1918]: E1213 01:28:13.792345 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:14.792561 kubelet[1918]: E1213 01:28:14.792481 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:15.133596 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3118808791.mount: Deactivated successfully. Dec 13 01:28:15.638495 containerd[1572]: time="2024-12-13T01:28:15.638377299Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 01:28:15.639871 containerd[1572]: time="2024-12-13T01:28:15.639814617Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.29.1: active requests=0, bytes read=142742010" Dec 13 01:28:15.641556 containerd[1572]: time="2024-12-13T01:28:15.641508957Z" level=info msg="ImageCreate event name:\"sha256:feb26d4585d68e875d9bd9bd6c27ea9f2d5c9ed9ef70f8b8cb0ebb0559a1d664\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 01:28:15.644380 containerd[1572]: time="2024-12-13T01:28:15.644302377Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:99c3917516efe1f807a0cfdf2d14b628b7c5cc6bd8a9ee5a253154f31756bea1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 01:28:15.645104 containerd[1572]: time="2024-12-13T01:28:15.645016741Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.29.1\" with image id \"sha256:feb26d4585d68e875d9bd9bd6c27ea9f2d5c9ed9ef70f8b8cb0ebb0559a1d664\", repo tag \"ghcr.io/flatcar/calico/node:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/node@sha256:99c3917516efe1f807a0cfdf2d14b628b7c5cc6bd8a9ee5a253154f31756bea1\", size \"142741872\" in 8.336512956s" Dec 13 01:28:15.645177 containerd[1572]: time="2024-12-13T01:28:15.645104333Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.1\" returns image reference \"sha256:feb26d4585d68e875d9bd9bd6c27ea9f2d5c9ed9ef70f8b8cb0ebb0559a1d664\"" Dec 13 01:28:15.660098 containerd[1572]: time="2024-12-13T01:28:15.660046747Z" level=info msg="CreateContainer within sandbox \"56634de33987afcd6f9fcbc73b443bfb032f311c4af438bd0256fdd5ec9dc242\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Dec 13 01:28:15.793587 kubelet[1918]: E1213 01:28:15.793469 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:15.808939 containerd[1572]: time="2024-12-13T01:28:15.808870093Z" level=info msg="CreateContainer within sandbox \"56634de33987afcd6f9fcbc73b443bfb032f311c4af438bd0256fdd5ec9dc242\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"c891fe248c381c3340d636039ce6a8f046809ab3a15f832945f3eab85a054716\"" Dec 13 01:28:15.809840 containerd[1572]: time="2024-12-13T01:28:15.809767479Z" level=info msg="StartContainer for \"c891fe248c381c3340d636039ce6a8f046809ab3a15f832945f3eab85a054716\"" Dec 13 01:28:16.060049 containerd[1572]: time="2024-12-13T01:28:16.059945985Z" level=info msg="StartContainer for \"c891fe248c381c3340d636039ce6a8f046809ab3a15f832945f3eab85a054716\" returns successfully" Dec 13 01:28:16.225380 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Dec 13 01:28:16.225578 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Dec 13 01:28:16.362691 kubelet[1918]: E1213 01:28:16.362433 1918 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 01:28:16.794787 kubelet[1918]: E1213 01:28:16.794575 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:17.367651 kubelet[1918]: E1213 01:28:17.364886 1918 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 01:28:17.795883 kubelet[1918]: E1213 01:28:17.795577 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:18.662094 kernel: bpftool[2749]: memfd_create() called without MFD_EXEC or MFD_NOEXEC_SEAL set Dec 13 01:28:18.796843 kubelet[1918]: E1213 01:28:18.796729 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:19.153483 systemd-networkd[1255]: vxlan.calico: Link UP Dec 13 01:28:19.153498 systemd-networkd[1255]: vxlan.calico: Gained carrier Dec 13 01:28:19.798015 kubelet[1918]: E1213 01:28:19.797849 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:20.779569 systemd-networkd[1255]: vxlan.calico: Gained IPv6LL Dec 13 01:28:20.798614 kubelet[1918]: E1213 01:28:20.798504 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:21.799246 kubelet[1918]: E1213 01:28:21.799100 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:22.233664 containerd[1572]: time="2024-12-13T01:28:22.233308993Z" level=info msg="StopPodSandbox for \"1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30\"" Dec 13 01:28:22.361463 kubelet[1918]: I1213 01:28:22.357872 1918 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="calico-system/calico-node-qpvfj" podStartSLOduration=9.795686282 podStartE2EDuration="31.35767948s" podCreationTimestamp="2024-12-13 01:27:51 +0000 UTC" firstStartedPulling="2024-12-13 01:27:54.083477397 +0000 UTC m=+3.660857663" lastFinishedPulling="2024-12-13 01:28:15.645470596 +0000 UTC m=+25.222850861" observedRunningTime="2024-12-13 01:28:16.431146244 +0000 UTC m=+26.008526509" watchObservedRunningTime="2024-12-13 01:28:22.35767948 +0000 UTC m=+31.935059745" Dec 13 01:28:22.426466 containerd[1572]: 2024-12-13 01:28:22.356 [INFO][2843] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30" Dec 13 01:28:22.426466 containerd[1572]: 2024-12-13 01:28:22.357 [INFO][2843] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30" iface="eth0" netns="/var/run/netns/cni-57a9bbea-c66b-1566-b67a-0845d5a4e897" Dec 13 01:28:22.426466 containerd[1572]: 2024-12-13 01:28:22.357 [INFO][2843] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30" iface="eth0" netns="/var/run/netns/cni-57a9bbea-c66b-1566-b67a-0845d5a4e897" Dec 13 01:28:22.426466 containerd[1572]: 2024-12-13 01:28:22.357 [INFO][2843] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30" iface="eth0" netns="/var/run/netns/cni-57a9bbea-c66b-1566-b67a-0845d5a4e897" Dec 13 01:28:22.426466 containerd[1572]: 2024-12-13 01:28:22.357 [INFO][2843] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30" Dec 13 01:28:22.426466 containerd[1572]: 2024-12-13 01:28:22.357 [INFO][2843] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30" Dec 13 01:28:22.426466 containerd[1572]: 2024-12-13 01:28:22.401 [INFO][2850] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30" HandleID="k8s-pod-network.1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30" Workload="10.0.0.50-k8s-csi--node--driver--nmw4l-eth0" Dec 13 01:28:22.426466 containerd[1572]: 2024-12-13 01:28:22.401 [INFO][2850] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Dec 13 01:28:22.426466 containerd[1572]: 2024-12-13 01:28:22.401 [INFO][2850] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Dec 13 01:28:22.426466 containerd[1572]: 2024-12-13 01:28:22.413 [WARNING][2850] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30" HandleID="k8s-pod-network.1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30" Workload="10.0.0.50-k8s-csi--node--driver--nmw4l-eth0" Dec 13 01:28:22.426466 containerd[1572]: 2024-12-13 01:28:22.413 [INFO][2850] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30" HandleID="k8s-pod-network.1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30" Workload="10.0.0.50-k8s-csi--node--driver--nmw4l-eth0" Dec 13 01:28:22.426466 containerd[1572]: 2024-12-13 01:28:22.417 [INFO][2850] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Dec 13 01:28:22.426466 containerd[1572]: 2024-12-13 01:28:22.422 [INFO][2843] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30" Dec 13 01:28:22.426943 containerd[1572]: time="2024-12-13T01:28:22.426704599Z" level=info msg="TearDown network for sandbox \"1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30\" successfully" Dec 13 01:28:22.426943 containerd[1572]: time="2024-12-13T01:28:22.426744616Z" level=info msg="StopPodSandbox for \"1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30\" returns successfully" Dec 13 01:28:22.430352 containerd[1572]: time="2024-12-13T01:28:22.428983315Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-nmw4l,Uid:8e158755-5b09-4a6a-ba0b-890e0f7dd097,Namespace:calico-system,Attempt:1,}" Dec 13 01:28:22.430910 systemd[1]: run-netns-cni\x2d57a9bbea\x2dc66b\x2d1566\x2db67a\x2d0845d5a4e897.mount: Deactivated successfully. Dec 13 01:28:22.682428 systemd-networkd[1255]: cali701ca78a3ee: Link UP Dec 13 01:28:22.682935 systemd-networkd[1255]: cali701ca78a3ee: Gained carrier Dec 13 01:28:22.711694 containerd[1572]: 2024-12-13 01:28:22.522 [INFO][2857] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {10.0.0.50-k8s-csi--node--driver--nmw4l-eth0 csi-node-driver- calico-system 8e158755-5b09-4a6a-ba0b-890e0f7dd097 1085 0 2024-12-13 01:27:51 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:55b695c467 k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s 10.0.0.50 csi-node-driver-nmw4l eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] cali701ca78a3ee [] []}} ContainerID="ce5b479e99d74027426bb31afccb63abafe070101864964785e2718542123558" Namespace="calico-system" Pod="csi-node-driver-nmw4l" WorkloadEndpoint="10.0.0.50-k8s-csi--node--driver--nmw4l-" Dec 13 01:28:22.711694 containerd[1572]: 2024-12-13 01:28:22.523 [INFO][2857] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="ce5b479e99d74027426bb31afccb63abafe070101864964785e2718542123558" Namespace="calico-system" Pod="csi-node-driver-nmw4l" WorkloadEndpoint="10.0.0.50-k8s-csi--node--driver--nmw4l-eth0" Dec 13 01:28:22.711694 containerd[1572]: 2024-12-13 01:28:22.583 [INFO][2871] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="ce5b479e99d74027426bb31afccb63abafe070101864964785e2718542123558" HandleID="k8s-pod-network.ce5b479e99d74027426bb31afccb63abafe070101864964785e2718542123558" Workload="10.0.0.50-k8s-csi--node--driver--nmw4l-eth0" Dec 13 01:28:22.711694 containerd[1572]: 2024-12-13 01:28:22.603 [INFO][2871] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="ce5b479e99d74027426bb31afccb63abafe070101864964785e2718542123558" HandleID="k8s-pod-network.ce5b479e99d74027426bb31afccb63abafe070101864964785e2718542123558" Workload="10.0.0.50-k8s-csi--node--driver--nmw4l-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000315260), Attrs:map[string]string{"namespace":"calico-system", "node":"10.0.0.50", "pod":"csi-node-driver-nmw4l", "timestamp":"2024-12-13 01:28:22.583596946 +0000 UTC"}, Hostname:"10.0.0.50", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Dec 13 01:28:22.711694 containerd[1572]: 2024-12-13 01:28:22.603 [INFO][2871] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Dec 13 01:28:22.711694 containerd[1572]: 2024-12-13 01:28:22.603 [INFO][2871] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Dec 13 01:28:22.711694 containerd[1572]: 2024-12-13 01:28:22.603 [INFO][2871] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host '10.0.0.50' Dec 13 01:28:22.711694 containerd[1572]: 2024-12-13 01:28:22.608 [INFO][2871] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.ce5b479e99d74027426bb31afccb63abafe070101864964785e2718542123558" host="10.0.0.50" Dec 13 01:28:22.711694 containerd[1572]: 2024-12-13 01:28:22.619 [INFO][2871] ipam/ipam.go 372: Looking up existing affinities for host host="10.0.0.50" Dec 13 01:28:22.711694 containerd[1572]: 2024-12-13 01:28:22.628 [INFO][2871] ipam/ipam.go 489: Trying affinity for 192.168.55.192/26 host="10.0.0.50" Dec 13 01:28:22.711694 containerd[1572]: 2024-12-13 01:28:22.636 [INFO][2871] ipam/ipam.go 155: Attempting to load block cidr=192.168.55.192/26 host="10.0.0.50" Dec 13 01:28:22.711694 containerd[1572]: 2024-12-13 01:28:22.643 [INFO][2871] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.55.192/26 host="10.0.0.50" Dec 13 01:28:22.711694 containerd[1572]: 2024-12-13 01:28:22.643 [INFO][2871] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.55.192/26 handle="k8s-pod-network.ce5b479e99d74027426bb31afccb63abafe070101864964785e2718542123558" host="10.0.0.50" Dec 13 01:28:22.711694 containerd[1572]: 2024-12-13 01:28:22.648 [INFO][2871] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.ce5b479e99d74027426bb31afccb63abafe070101864964785e2718542123558 Dec 13 01:28:22.711694 containerd[1572]: 2024-12-13 01:28:22.659 [INFO][2871] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.55.192/26 handle="k8s-pod-network.ce5b479e99d74027426bb31afccb63abafe070101864964785e2718542123558" host="10.0.0.50" Dec 13 01:28:22.711694 containerd[1572]: 2024-12-13 01:28:22.671 [INFO][2871] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.55.193/26] block=192.168.55.192/26 handle="k8s-pod-network.ce5b479e99d74027426bb31afccb63abafe070101864964785e2718542123558" host="10.0.0.50" Dec 13 01:28:22.711694 containerd[1572]: 2024-12-13 01:28:22.671 [INFO][2871] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.55.193/26] handle="k8s-pod-network.ce5b479e99d74027426bb31afccb63abafe070101864964785e2718542123558" host="10.0.0.50" Dec 13 01:28:22.711694 containerd[1572]: 2024-12-13 01:28:22.671 [INFO][2871] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Dec 13 01:28:22.711694 containerd[1572]: 2024-12-13 01:28:22.671 [INFO][2871] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.55.193/26] IPv6=[] ContainerID="ce5b479e99d74027426bb31afccb63abafe070101864964785e2718542123558" HandleID="k8s-pod-network.ce5b479e99d74027426bb31afccb63abafe070101864964785e2718542123558" Workload="10.0.0.50-k8s-csi--node--driver--nmw4l-eth0" Dec 13 01:28:22.712719 containerd[1572]: 2024-12-13 01:28:22.676 [INFO][2857] cni-plugin/k8s.go 386: Populated endpoint ContainerID="ce5b479e99d74027426bb31afccb63abafe070101864964785e2718542123558" Namespace="calico-system" Pod="csi-node-driver-nmw4l" WorkloadEndpoint="10.0.0.50-k8s-csi--node--driver--nmw4l-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"10.0.0.50-k8s-csi--node--driver--nmw4l-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"8e158755-5b09-4a6a-ba0b-890e0f7dd097", ResourceVersion:"1085", Generation:0, CreationTimestamp:time.Date(2024, time.December, 13, 1, 27, 51, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"55b695c467", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"10.0.0.50", ContainerID:"", Pod:"csi-node-driver-nmw4l", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.55.193/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali701ca78a3ee", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Dec 13 01:28:22.712719 containerd[1572]: 2024-12-13 01:28:22.676 [INFO][2857] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.55.193/32] ContainerID="ce5b479e99d74027426bb31afccb63abafe070101864964785e2718542123558" Namespace="calico-system" Pod="csi-node-driver-nmw4l" WorkloadEndpoint="10.0.0.50-k8s-csi--node--driver--nmw4l-eth0" Dec 13 01:28:22.712719 containerd[1572]: 2024-12-13 01:28:22.676 [INFO][2857] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali701ca78a3ee ContainerID="ce5b479e99d74027426bb31afccb63abafe070101864964785e2718542123558" Namespace="calico-system" Pod="csi-node-driver-nmw4l" WorkloadEndpoint="10.0.0.50-k8s-csi--node--driver--nmw4l-eth0" Dec 13 01:28:22.712719 containerd[1572]: 2024-12-13 01:28:22.683 [INFO][2857] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="ce5b479e99d74027426bb31afccb63abafe070101864964785e2718542123558" Namespace="calico-system" Pod="csi-node-driver-nmw4l" WorkloadEndpoint="10.0.0.50-k8s-csi--node--driver--nmw4l-eth0" Dec 13 01:28:22.712719 containerd[1572]: 2024-12-13 01:28:22.684 [INFO][2857] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="ce5b479e99d74027426bb31afccb63abafe070101864964785e2718542123558" Namespace="calico-system" Pod="csi-node-driver-nmw4l" WorkloadEndpoint="10.0.0.50-k8s-csi--node--driver--nmw4l-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"10.0.0.50-k8s-csi--node--driver--nmw4l-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"8e158755-5b09-4a6a-ba0b-890e0f7dd097", ResourceVersion:"1085", Generation:0, CreationTimestamp:time.Date(2024, time.December, 13, 1, 27, 51, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"55b695c467", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"10.0.0.50", ContainerID:"ce5b479e99d74027426bb31afccb63abafe070101864964785e2718542123558", Pod:"csi-node-driver-nmw4l", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.55.193/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali701ca78a3ee", MAC:"7a:0b:76:90:27:be", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Dec 13 01:28:22.712719 containerd[1572]: 2024-12-13 01:28:22.707 [INFO][2857] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="ce5b479e99d74027426bb31afccb63abafe070101864964785e2718542123558" Namespace="calico-system" Pod="csi-node-driver-nmw4l" WorkloadEndpoint="10.0.0.50-k8s-csi--node--driver--nmw4l-eth0" Dec 13 01:28:22.760456 containerd[1572]: time="2024-12-13T01:28:22.759872253Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 01:28:22.760456 containerd[1572]: time="2024-12-13T01:28:22.760052546Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 01:28:22.760456 containerd[1572]: time="2024-12-13T01:28:22.760080197Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 01:28:22.760456 containerd[1572]: time="2024-12-13T01:28:22.760232879Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 01:28:22.799994 kubelet[1918]: E1213 01:28:22.799880 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:22.803747 systemd-resolved[1462]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Dec 13 01:28:22.828811 containerd[1572]: time="2024-12-13T01:28:22.828722451Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-nmw4l,Uid:8e158755-5b09-4a6a-ba0b-890e0f7dd097,Namespace:calico-system,Attempt:1,} returns sandbox id \"ce5b479e99d74027426bb31afccb63abafe070101864964785e2718542123558\"" Dec 13 01:28:22.832231 containerd[1572]: time="2024-12-13T01:28:22.831718299Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.1\"" Dec 13 01:28:23.233989 containerd[1572]: time="2024-12-13T01:28:23.233901339Z" level=info msg="StopPodSandbox for \"eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e\"" Dec 13 01:28:23.688647 containerd[1572]: 2024-12-13 01:28:23.613 [INFO][2950] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e" Dec 13 01:28:23.688647 containerd[1572]: 2024-12-13 01:28:23.613 [INFO][2950] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e" iface="eth0" netns="/var/run/netns/cni-28eacd30-2289-4718-479f-9873c7c6fa96" Dec 13 01:28:23.688647 containerd[1572]: 2024-12-13 01:28:23.614 [INFO][2950] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e" iface="eth0" netns="/var/run/netns/cni-28eacd30-2289-4718-479f-9873c7c6fa96" Dec 13 01:28:23.688647 containerd[1572]: 2024-12-13 01:28:23.614 [INFO][2950] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e" iface="eth0" netns="/var/run/netns/cni-28eacd30-2289-4718-479f-9873c7c6fa96" Dec 13 01:28:23.688647 containerd[1572]: 2024-12-13 01:28:23.615 [INFO][2950] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e" Dec 13 01:28:23.688647 containerd[1572]: 2024-12-13 01:28:23.615 [INFO][2950] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e" Dec 13 01:28:23.688647 containerd[1572]: 2024-12-13 01:28:23.653 [INFO][2957] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e" HandleID="k8s-pod-network.eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e" Workload="10.0.0.50-k8s-nginx--deployment--6d5f899847--785tc-eth0" Dec 13 01:28:23.688647 containerd[1572]: 2024-12-13 01:28:23.653 [INFO][2957] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Dec 13 01:28:23.688647 containerd[1572]: 2024-12-13 01:28:23.653 [INFO][2957] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Dec 13 01:28:23.688647 containerd[1572]: 2024-12-13 01:28:23.669 [WARNING][2957] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e" HandleID="k8s-pod-network.eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e" Workload="10.0.0.50-k8s-nginx--deployment--6d5f899847--785tc-eth0" Dec 13 01:28:23.688647 containerd[1572]: 2024-12-13 01:28:23.670 [INFO][2957] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e" HandleID="k8s-pod-network.eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e" Workload="10.0.0.50-k8s-nginx--deployment--6d5f899847--785tc-eth0" Dec 13 01:28:23.688647 containerd[1572]: 2024-12-13 01:28:23.680 [INFO][2957] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Dec 13 01:28:23.688647 containerd[1572]: 2024-12-13 01:28:23.684 [INFO][2950] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e" Dec 13 01:28:23.689887 containerd[1572]: time="2024-12-13T01:28:23.689825131Z" level=info msg="TearDown network for sandbox \"eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e\" successfully" Dec 13 01:28:23.689887 containerd[1572]: time="2024-12-13T01:28:23.689874558Z" level=info msg="StopPodSandbox for \"eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e\" returns successfully" Dec 13 01:28:23.691434 containerd[1572]: time="2024-12-13T01:28:23.690907974Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-6d5f899847-785tc,Uid:2767e440-abcd-45e1-95b3-0943b03016cf,Namespace:default,Attempt:1,}" Dec 13 01:28:23.692985 systemd[1]: run-netns-cni\x2d28eacd30\x2d2289\x2d4718\x2d479f\x2d9873c7c6fa96.mount: Deactivated successfully. Dec 13 01:28:23.723976 systemd-networkd[1255]: cali701ca78a3ee: Gained IPv6LL Dec 13 01:28:23.800848 kubelet[1918]: E1213 01:28:23.800743 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:24.447506 systemd-networkd[1255]: califcfaaceb738: Link UP Dec 13 01:28:24.447919 systemd-networkd[1255]: califcfaaceb738: Gained carrier Dec 13 01:28:24.468514 containerd[1572]: 2024-12-13 01:28:24.267 [INFO][2966] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {10.0.0.50-k8s-nginx--deployment--6d5f899847--785tc-eth0 nginx-deployment-6d5f899847- default 2767e440-abcd-45e1-95b3-0943b03016cf 1092 0 2024-12-13 01:28:08 +0000 UTC map[app:nginx pod-template-hash:6d5f899847 projectcalico.org/namespace:default projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:default] map[] [] [] []} {k8s 10.0.0.50 nginx-deployment-6d5f899847-785tc eth0 default [] [] [kns.default ksa.default.default] califcfaaceb738 [] []}} ContainerID="a3614f4cfd7eba109151c53d576b36a7a51f0c23fde2eff9309e1756ce745727" Namespace="default" Pod="nginx-deployment-6d5f899847-785tc" WorkloadEndpoint="10.0.0.50-k8s-nginx--deployment--6d5f899847--785tc-" Dec 13 01:28:24.468514 containerd[1572]: 2024-12-13 01:28:24.267 [INFO][2966] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="a3614f4cfd7eba109151c53d576b36a7a51f0c23fde2eff9309e1756ce745727" Namespace="default" Pod="nginx-deployment-6d5f899847-785tc" WorkloadEndpoint="10.0.0.50-k8s-nginx--deployment--6d5f899847--785tc-eth0" Dec 13 01:28:24.468514 containerd[1572]: 2024-12-13 01:28:24.311 [INFO][2978] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="a3614f4cfd7eba109151c53d576b36a7a51f0c23fde2eff9309e1756ce745727" HandleID="k8s-pod-network.a3614f4cfd7eba109151c53d576b36a7a51f0c23fde2eff9309e1756ce745727" Workload="10.0.0.50-k8s-nginx--deployment--6d5f899847--785tc-eth0" Dec 13 01:28:24.468514 containerd[1572]: 2024-12-13 01:28:24.333 [INFO][2978] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="a3614f4cfd7eba109151c53d576b36a7a51f0c23fde2eff9309e1756ce745727" HandleID="k8s-pod-network.a3614f4cfd7eba109151c53d576b36a7a51f0c23fde2eff9309e1756ce745727" Workload="10.0.0.50-k8s-nginx--deployment--6d5f899847--785tc-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000377bc0), Attrs:map[string]string{"namespace":"default", "node":"10.0.0.50", "pod":"nginx-deployment-6d5f899847-785tc", "timestamp":"2024-12-13 01:28:24.311156141 +0000 UTC"}, Hostname:"10.0.0.50", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Dec 13 01:28:24.468514 containerd[1572]: 2024-12-13 01:28:24.334 [INFO][2978] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Dec 13 01:28:24.468514 containerd[1572]: 2024-12-13 01:28:24.334 [INFO][2978] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Dec 13 01:28:24.468514 containerd[1572]: 2024-12-13 01:28:24.334 [INFO][2978] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host '10.0.0.50' Dec 13 01:28:24.468514 containerd[1572]: 2024-12-13 01:28:24.337 [INFO][2978] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.a3614f4cfd7eba109151c53d576b36a7a51f0c23fde2eff9309e1756ce745727" host="10.0.0.50" Dec 13 01:28:24.468514 containerd[1572]: 2024-12-13 01:28:24.344 [INFO][2978] ipam/ipam.go 372: Looking up existing affinities for host host="10.0.0.50" Dec 13 01:28:24.468514 containerd[1572]: 2024-12-13 01:28:24.354 [INFO][2978] ipam/ipam.go 489: Trying affinity for 192.168.55.192/26 host="10.0.0.50" Dec 13 01:28:24.468514 containerd[1572]: 2024-12-13 01:28:24.398 [INFO][2978] ipam/ipam.go 155: Attempting to load block cidr=192.168.55.192/26 host="10.0.0.50" Dec 13 01:28:24.468514 containerd[1572]: 2024-12-13 01:28:24.401 [INFO][2978] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.55.192/26 host="10.0.0.50" Dec 13 01:28:24.468514 containerd[1572]: 2024-12-13 01:28:24.401 [INFO][2978] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.55.192/26 handle="k8s-pod-network.a3614f4cfd7eba109151c53d576b36a7a51f0c23fde2eff9309e1756ce745727" host="10.0.0.50" Dec 13 01:28:24.468514 containerd[1572]: 2024-12-13 01:28:24.405 [INFO][2978] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.a3614f4cfd7eba109151c53d576b36a7a51f0c23fde2eff9309e1756ce745727 Dec 13 01:28:24.468514 containerd[1572]: 2024-12-13 01:28:24.423 [INFO][2978] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.55.192/26 handle="k8s-pod-network.a3614f4cfd7eba109151c53d576b36a7a51f0c23fde2eff9309e1756ce745727" host="10.0.0.50" Dec 13 01:28:24.468514 containerd[1572]: 2024-12-13 01:28:24.434 [INFO][2978] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.55.194/26] block=192.168.55.192/26 handle="k8s-pod-network.a3614f4cfd7eba109151c53d576b36a7a51f0c23fde2eff9309e1756ce745727" host="10.0.0.50" Dec 13 01:28:24.468514 containerd[1572]: 2024-12-13 01:28:24.434 [INFO][2978] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.55.194/26] handle="k8s-pod-network.a3614f4cfd7eba109151c53d576b36a7a51f0c23fde2eff9309e1756ce745727" host="10.0.0.50" Dec 13 01:28:24.468514 containerd[1572]: 2024-12-13 01:28:24.434 [INFO][2978] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Dec 13 01:28:24.468514 containerd[1572]: 2024-12-13 01:28:24.434 [INFO][2978] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.55.194/26] IPv6=[] ContainerID="a3614f4cfd7eba109151c53d576b36a7a51f0c23fde2eff9309e1756ce745727" HandleID="k8s-pod-network.a3614f4cfd7eba109151c53d576b36a7a51f0c23fde2eff9309e1756ce745727" Workload="10.0.0.50-k8s-nginx--deployment--6d5f899847--785tc-eth0" Dec 13 01:28:24.469699 containerd[1572]: 2024-12-13 01:28:24.438 [INFO][2966] cni-plugin/k8s.go 386: Populated endpoint ContainerID="a3614f4cfd7eba109151c53d576b36a7a51f0c23fde2eff9309e1756ce745727" Namespace="default" Pod="nginx-deployment-6d5f899847-785tc" WorkloadEndpoint="10.0.0.50-k8s-nginx--deployment--6d5f899847--785tc-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"10.0.0.50-k8s-nginx--deployment--6d5f899847--785tc-eth0", GenerateName:"nginx-deployment-6d5f899847-", Namespace:"default", SelfLink:"", UID:"2767e440-abcd-45e1-95b3-0943b03016cf", ResourceVersion:"1092", Generation:0, CreationTimestamp:time.Date(2024, time.December, 13, 1, 28, 8, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app":"nginx", "pod-template-hash":"6d5f899847", "projectcalico.org/namespace":"default", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"default"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"10.0.0.50", ContainerID:"", Pod:"nginx-deployment-6d5f899847-785tc", Endpoint:"eth0", ServiceAccountName:"default", IPNetworks:[]string{"192.168.55.194/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.default", "ksa.default.default"}, InterfaceName:"califcfaaceb738", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Dec 13 01:28:24.469699 containerd[1572]: 2024-12-13 01:28:24.438 [INFO][2966] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.55.194/32] ContainerID="a3614f4cfd7eba109151c53d576b36a7a51f0c23fde2eff9309e1756ce745727" Namespace="default" Pod="nginx-deployment-6d5f899847-785tc" WorkloadEndpoint="10.0.0.50-k8s-nginx--deployment--6d5f899847--785tc-eth0" Dec 13 01:28:24.469699 containerd[1572]: 2024-12-13 01:28:24.438 [INFO][2966] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to califcfaaceb738 ContainerID="a3614f4cfd7eba109151c53d576b36a7a51f0c23fde2eff9309e1756ce745727" Namespace="default" Pod="nginx-deployment-6d5f899847-785tc" WorkloadEndpoint="10.0.0.50-k8s-nginx--deployment--6d5f899847--785tc-eth0" Dec 13 01:28:24.469699 containerd[1572]: 2024-12-13 01:28:24.447 [INFO][2966] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="a3614f4cfd7eba109151c53d576b36a7a51f0c23fde2eff9309e1756ce745727" Namespace="default" Pod="nginx-deployment-6d5f899847-785tc" WorkloadEndpoint="10.0.0.50-k8s-nginx--deployment--6d5f899847--785tc-eth0" Dec 13 01:28:24.469699 containerd[1572]: 2024-12-13 01:28:24.448 [INFO][2966] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="a3614f4cfd7eba109151c53d576b36a7a51f0c23fde2eff9309e1756ce745727" Namespace="default" Pod="nginx-deployment-6d5f899847-785tc" WorkloadEndpoint="10.0.0.50-k8s-nginx--deployment--6d5f899847--785tc-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"10.0.0.50-k8s-nginx--deployment--6d5f899847--785tc-eth0", GenerateName:"nginx-deployment-6d5f899847-", Namespace:"default", SelfLink:"", UID:"2767e440-abcd-45e1-95b3-0943b03016cf", ResourceVersion:"1092", Generation:0, CreationTimestamp:time.Date(2024, time.December, 13, 1, 28, 8, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app":"nginx", "pod-template-hash":"6d5f899847", "projectcalico.org/namespace":"default", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"default"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"10.0.0.50", ContainerID:"a3614f4cfd7eba109151c53d576b36a7a51f0c23fde2eff9309e1756ce745727", Pod:"nginx-deployment-6d5f899847-785tc", Endpoint:"eth0", ServiceAccountName:"default", IPNetworks:[]string{"192.168.55.194/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.default", "ksa.default.default"}, InterfaceName:"califcfaaceb738", MAC:"3e:cc:cc:f0:06:e0", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Dec 13 01:28:24.469699 containerd[1572]: 2024-12-13 01:28:24.464 [INFO][2966] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="a3614f4cfd7eba109151c53d576b36a7a51f0c23fde2eff9309e1756ce745727" Namespace="default" Pod="nginx-deployment-6d5f899847-785tc" WorkloadEndpoint="10.0.0.50-k8s-nginx--deployment--6d5f899847--785tc-eth0" Dec 13 01:28:24.502254 containerd[1572]: time="2024-12-13T01:28:24.501742427Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 01:28:24.502548 containerd[1572]: time="2024-12-13T01:28:24.501868006Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 01:28:24.502548 containerd[1572]: time="2024-12-13T01:28:24.502007255Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 01:28:24.503654 containerd[1572]: time="2024-12-13T01:28:24.503574391Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 01:28:24.569536 systemd-resolved[1462]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Dec 13 01:28:24.608285 containerd[1572]: time="2024-12-13T01:28:24.608233750Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-6d5f899847-785tc,Uid:2767e440-abcd-45e1-95b3-0943b03016cf,Namespace:default,Attempt:1,} returns sandbox id \"a3614f4cfd7eba109151c53d576b36a7a51f0c23fde2eff9309e1756ce745727\"" Dec 13 01:28:24.801661 kubelet[1918]: E1213 01:28:24.801175 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:25.058314 containerd[1572]: time="2024-12-13T01:28:25.058102185Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 01:28:25.059119 containerd[1572]: time="2024-12-13T01:28:25.059026396Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.29.1: active requests=0, bytes read=7902632" Dec 13 01:28:25.061062 containerd[1572]: time="2024-12-13T01:28:25.060937805Z" level=info msg="ImageCreate event name:\"sha256:bda8c42e04758c4f061339e213f50ccdc7502c4176fbf631aa12357e62b63540\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 01:28:25.064093 containerd[1572]: time="2024-12-13T01:28:25.063951124Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:eaa7e01fb16b603c155a67b81f16992281db7f831684c7b2081d3434587a7ff3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 01:28:25.065211 containerd[1572]: time="2024-12-13T01:28:25.065142014Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.29.1\" with image id \"sha256:bda8c42e04758c4f061339e213f50ccdc7502c4176fbf631aa12357e62b63540\", repo tag \"ghcr.io/flatcar/calico/csi:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:eaa7e01fb16b603c155a67b81f16992281db7f831684c7b2081d3434587a7ff3\", size \"9395716\" in 2.233359305s" Dec 13 01:28:25.065277 containerd[1572]: time="2024-12-13T01:28:25.065211533Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.1\" returns image reference \"sha256:bda8c42e04758c4f061339e213f50ccdc7502c4176fbf631aa12357e62b63540\"" Dec 13 01:28:25.066071 containerd[1572]: time="2024-12-13T01:28:25.065996959Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\"" Dec 13 01:28:25.067763 containerd[1572]: time="2024-12-13T01:28:25.067720085Z" level=info msg="CreateContainer within sandbox \"ce5b479e99d74027426bb31afccb63abafe070101864964785e2718542123558\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Dec 13 01:28:25.096132 containerd[1572]: time="2024-12-13T01:28:25.096051533Z" level=info msg="CreateContainer within sandbox \"ce5b479e99d74027426bb31afccb63abafe070101864964785e2718542123558\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"a768501046ce86f05ddba56ce6b8f424e347d31fd7fd926bd2729050f602d342\"" Dec 13 01:28:25.097017 containerd[1572]: time="2024-12-13T01:28:25.096950892Z" level=info msg="StartContainer for \"a768501046ce86f05ddba56ce6b8f424e347d31fd7fd926bd2729050f602d342\"" Dec 13 01:28:25.185777 containerd[1572]: time="2024-12-13T01:28:25.185702681Z" level=info msg="StartContainer for \"a768501046ce86f05ddba56ce6b8f424e347d31fd7fd926bd2729050f602d342\" returns successfully" Dec 13 01:28:25.801974 kubelet[1918]: E1213 01:28:25.801900 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:26.283577 systemd-networkd[1255]: califcfaaceb738: Gained IPv6LL Dec 13 01:28:26.802458 kubelet[1918]: E1213 01:28:26.802365 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:27.803051 kubelet[1918]: E1213 01:28:27.802977 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:28.306199 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1038059901.mount: Deactivated successfully. Dec 13 01:28:28.803638 kubelet[1918]: E1213 01:28:28.803573 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:29.290067 update_engine[1553]: I20241213 01:28:29.289819 1553 update_attempter.cc:509] Updating boot flags... Dec 13 01:28:29.361184 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 41 scanned by (udev-worker) (3105) Dec 13 01:28:29.417081 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 41 scanned by (udev-worker) (3104) Dec 13 01:28:29.472123 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 41 scanned by (udev-worker) (3104) Dec 13 01:28:29.804420 kubelet[1918]: E1213 01:28:29.804331 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:30.225103 containerd[1572]: time="2024-12-13T01:28:30.224899966Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/nginx:latest\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 01:28:30.226584 containerd[1572]: time="2024-12-13T01:28:30.226534627Z" level=info msg="stop pulling image ghcr.io/flatcar/nginx:latest: active requests=0, bytes read=71036027" Dec 13 01:28:30.227782 containerd[1572]: time="2024-12-13T01:28:30.227739066Z" level=info msg="ImageCreate event name:\"sha256:fa0a8cea5e76ad962111c39c85bb312edaf5b89eccd8f404eeea66c9759641e3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 01:28:30.233803 containerd[1572]: time="2024-12-13T01:28:30.233744596Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/nginx@sha256:e04edf30a4ea4c5a4107110797c72d3ee8a654415f00acd4019be17218afd9a1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 01:28:30.234976 containerd[1572]: time="2024-12-13T01:28:30.234915004Z" level=info msg="Pulled image \"ghcr.io/flatcar/nginx:latest\" with image id \"sha256:fa0a8cea5e76ad962111c39c85bb312edaf5b89eccd8f404eeea66c9759641e3\", repo tag \"ghcr.io/flatcar/nginx:latest\", repo digest \"ghcr.io/flatcar/nginx@sha256:e04edf30a4ea4c5a4107110797c72d3ee8a654415f00acd4019be17218afd9a1\", size \"71035905\" in 5.168834508s" Dec 13 01:28:30.235050 containerd[1572]: time="2024-12-13T01:28:30.234973054Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\" returns image reference \"sha256:fa0a8cea5e76ad962111c39c85bb312edaf5b89eccd8f404eeea66c9759641e3\"" Dec 13 01:28:30.235937 containerd[1572]: time="2024-12-13T01:28:30.235900639Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\"" Dec 13 01:28:30.236893 containerd[1572]: time="2024-12-13T01:28:30.236863486Z" level=info msg="CreateContainer within sandbox \"a3614f4cfd7eba109151c53d576b36a7a51f0c23fde2eff9309e1756ce745727\" for container &ContainerMetadata{Name:nginx,Attempt:0,}" Dec 13 01:28:30.262191 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount646081771.mount: Deactivated successfully. Dec 13 01:28:30.266956 containerd[1572]: time="2024-12-13T01:28:30.266896196Z" level=info msg="CreateContainer within sandbox \"a3614f4cfd7eba109151c53d576b36a7a51f0c23fde2eff9309e1756ce745727\" for &ContainerMetadata{Name:nginx,Attempt:0,} returns container id \"4673785661cd962615af750958437c686c5dc7e8f437cf5ae8a06cccdd7896a3\"" Dec 13 01:28:30.267634 containerd[1572]: time="2024-12-13T01:28:30.267601649Z" level=info msg="StartContainer for \"4673785661cd962615af750958437c686c5dc7e8f437cf5ae8a06cccdd7896a3\"" Dec 13 01:28:30.338528 systemd[1]: run-containerd-runc-k8s.io-4673785661cd962615af750958437c686c5dc7e8f437cf5ae8a06cccdd7896a3-runc.xBSakP.mount: Deactivated successfully. Dec 13 01:28:30.379673 containerd[1572]: time="2024-12-13T01:28:30.379599022Z" level=info msg="StartContainer for \"4673785661cd962615af750958437c686c5dc7e8f437cf5ae8a06cccdd7896a3\" returns successfully" Dec 13 01:28:30.425748 kubelet[1918]: I1213 01:28:30.425680 1918 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="default/nginx-deployment-6d5f899847-785tc" podStartSLOduration=16.800093963 podStartE2EDuration="22.425617677s" podCreationTimestamp="2024-12-13 01:28:08 +0000 UTC" firstStartedPulling="2024-12-13 01:28:24.609850673 +0000 UTC m=+34.187230938" lastFinishedPulling="2024-12-13 01:28:30.235374377 +0000 UTC m=+39.812754652" observedRunningTime="2024-12-13 01:28:30.425517429 +0000 UTC m=+40.002897694" watchObservedRunningTime="2024-12-13 01:28:30.425617677 +0000 UTC m=+40.002997952" Dec 13 01:28:30.776420 kubelet[1918]: E1213 01:28:30.776325 1918 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:30.805651 kubelet[1918]: E1213 01:28:30.805552 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:31.806320 kubelet[1918]: E1213 01:28:31.806211 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:31.840225 kubelet[1918]: E1213 01:28:31.840156 1918 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 01:28:32.807204 kubelet[1918]: E1213 01:28:32.807119 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:33.808319 kubelet[1918]: E1213 01:28:33.808196 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:34.337904 containerd[1572]: time="2024-12-13T01:28:34.337786040Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 01:28:34.339256 containerd[1572]: time="2024-12-13T01:28:34.339170744Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1: active requests=0, bytes read=10501081" Dec 13 01:28:34.341000 containerd[1572]: time="2024-12-13T01:28:34.340932377Z" level=info msg="ImageCreate event name:\"sha256:8b7d18f262d5cf6a6343578ad0db68a140c4c9989d9e02c58c27cb5d2c70320f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 01:28:34.346281 containerd[1572]: time="2024-12-13T01:28:34.346200110Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:a338da9488cbaa83c78457c3d7354d84149969c0480e88dd768e036632ff5b76\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 01:28:34.347131 containerd[1572]: time="2024-12-13T01:28:34.347058571Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\" with image id \"sha256:8b7d18f262d5cf6a6343578ad0db68a140c4c9989d9e02c58c27cb5d2c70320f\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:a338da9488cbaa83c78457c3d7354d84149969c0480e88dd768e036632ff5b76\", size \"11994117\" in 4.111111667s" Dec 13 01:28:34.347131 containerd[1572]: time="2024-12-13T01:28:34.347120417Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\" returns image reference \"sha256:8b7d18f262d5cf6a6343578ad0db68a140c4c9989d9e02c58c27cb5d2c70320f\"" Dec 13 01:28:34.349541 containerd[1572]: time="2024-12-13T01:28:34.349505222Z" level=info msg="CreateContainer within sandbox \"ce5b479e99d74027426bb31afccb63abafe070101864964785e2718542123558\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Dec 13 01:28:34.372961 containerd[1572]: time="2024-12-13T01:28:34.372891227Z" level=info msg="CreateContainer within sandbox \"ce5b479e99d74027426bb31afccb63abafe070101864964785e2718542123558\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"1ab0bddd6042aac0e85902f514c96816326f216d7d0826d05bc0a3774118587b\"" Dec 13 01:28:34.373674 containerd[1572]: time="2024-12-13T01:28:34.373567006Z" level=info msg="StartContainer for \"1ab0bddd6042aac0e85902f514c96816326f216d7d0826d05bc0a3774118587b\"" Dec 13 01:28:34.412967 systemd[1]: run-containerd-runc-k8s.io-1ab0bddd6042aac0e85902f514c96816326f216d7d0826d05bc0a3774118587b-runc.EKnIxL.mount: Deactivated successfully. Dec 13 01:28:34.464074 containerd[1572]: time="2024-12-13T01:28:34.463965167Z" level=info msg="StartContainer for \"1ab0bddd6042aac0e85902f514c96816326f216d7d0826d05bc0a3774118587b\" returns successfully" Dec 13 01:28:34.809643 kubelet[1918]: E1213 01:28:34.809529 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:35.345264 kubelet[1918]: I1213 01:28:35.345008 1918 csi_plugin.go:99] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Dec 13 01:28:35.345264 kubelet[1918]: I1213 01:28:35.345213 1918 csi_plugin.go:112] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Dec 13 01:28:35.810112 kubelet[1918]: E1213 01:28:35.810007 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:36.811305 kubelet[1918]: E1213 01:28:36.811235 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:37.379725 kubelet[1918]: I1213 01:28:37.379641 1918 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="calico-system/csi-node-driver-nmw4l" podStartSLOduration=34.86325399 podStartE2EDuration="46.3795702s" podCreationTimestamp="2024-12-13 01:27:51 +0000 UTC" firstStartedPulling="2024-12-13 01:28:22.831164192 +0000 UTC m=+32.408544457" lastFinishedPulling="2024-12-13 01:28:34.347480402 +0000 UTC m=+43.924860667" observedRunningTime="2024-12-13 01:28:35.446171613 +0000 UTC m=+45.023551878" watchObservedRunningTime="2024-12-13 01:28:37.3795702 +0000 UTC m=+46.956950465" Dec 13 01:28:37.379974 kubelet[1918]: I1213 01:28:37.379868 1918 topology_manager.go:215] "Topology Admit Handler" podUID="344508b1-e3a2-474d-84b2-6194aac12a19" podNamespace="default" podName="nfs-server-provisioner-0" Dec 13 01:28:37.402658 kubelet[1918]: I1213 01:28:37.402583 1918 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmrdc\" (UniqueName: \"kubernetes.io/projected/344508b1-e3a2-474d-84b2-6194aac12a19-kube-api-access-pmrdc\") pod \"nfs-server-provisioner-0\" (UID: \"344508b1-e3a2-474d-84b2-6194aac12a19\") " pod="default/nfs-server-provisioner-0" Dec 13 01:28:37.402658 kubelet[1918]: I1213 01:28:37.402657 1918 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data\" (UniqueName: \"kubernetes.io/empty-dir/344508b1-e3a2-474d-84b2-6194aac12a19-data\") pod \"nfs-server-provisioner-0\" (UID: \"344508b1-e3a2-474d-84b2-6194aac12a19\") " pod="default/nfs-server-provisioner-0" Dec 13 01:28:37.685232 containerd[1572]: time="2024-12-13T01:28:37.685089386Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nfs-server-provisioner-0,Uid:344508b1-e3a2-474d-84b2-6194aac12a19,Namespace:default,Attempt:0,}" Dec 13 01:28:37.805489 systemd-networkd[1255]: cali60e51b789ff: Link UP Dec 13 01:28:37.806646 systemd-networkd[1255]: cali60e51b789ff: Gained carrier Dec 13 01:28:37.812470 kubelet[1918]: E1213 01:28:37.812405 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:37.824659 containerd[1572]: 2024-12-13 01:28:37.732 [INFO][3271] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {10.0.0.50-k8s-nfs--server--provisioner--0-eth0 nfs-server-provisioner- default 344508b1-e3a2-474d-84b2-6194aac12a19 1170 0 2024-12-13 01:28:37 +0000 UTC map[app:nfs-server-provisioner apps.kubernetes.io/pod-index:0 chart:nfs-server-provisioner-1.8.0 controller-revision-hash:nfs-server-provisioner-d5cbb7f57 heritage:Helm projectcalico.org/namespace:default projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:nfs-server-provisioner release:nfs-server-provisioner statefulset.kubernetes.io/pod-name:nfs-server-provisioner-0] map[] [] [] []} {k8s 10.0.0.50 nfs-server-provisioner-0 eth0 nfs-server-provisioner [] [] [kns.default ksa.default.nfs-server-provisioner] cali60e51b789ff [{nfs TCP 2049 0 } {nfs-udp UDP 2049 0 } {nlockmgr TCP 32803 0 } {nlockmgr-udp UDP 32803 0 } {mountd TCP 20048 0 } {mountd-udp UDP 20048 0 } {rquotad TCP 875 0 } {rquotad-udp UDP 875 0 } {rpcbind TCP 111 0 } {rpcbind-udp UDP 111 0 } {statd TCP 662 0 } {statd-udp UDP 662 0 }] []}} ContainerID="1b0f2f5f77310cf8cd13d1699d9a80c57850e2c1ada2c656442e2a5d51c0e267" Namespace="default" Pod="nfs-server-provisioner-0" WorkloadEndpoint="10.0.0.50-k8s-nfs--server--provisioner--0-" Dec 13 01:28:37.824659 containerd[1572]: 2024-12-13 01:28:37.733 [INFO][3271] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="1b0f2f5f77310cf8cd13d1699d9a80c57850e2c1ada2c656442e2a5d51c0e267" Namespace="default" Pod="nfs-server-provisioner-0" WorkloadEndpoint="10.0.0.50-k8s-nfs--server--provisioner--0-eth0" Dec 13 01:28:37.824659 containerd[1572]: 2024-12-13 01:28:37.760 [INFO][3284] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="1b0f2f5f77310cf8cd13d1699d9a80c57850e2c1ada2c656442e2a5d51c0e267" HandleID="k8s-pod-network.1b0f2f5f77310cf8cd13d1699d9a80c57850e2c1ada2c656442e2a5d51c0e267" Workload="10.0.0.50-k8s-nfs--server--provisioner--0-eth0" Dec 13 01:28:37.824659 containerd[1572]: 2024-12-13 01:28:37.770 [INFO][3284] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="1b0f2f5f77310cf8cd13d1699d9a80c57850e2c1ada2c656442e2a5d51c0e267" HandleID="k8s-pod-network.1b0f2f5f77310cf8cd13d1699d9a80c57850e2c1ada2c656442e2a5d51c0e267" Workload="10.0.0.50-k8s-nfs--server--provisioner--0-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002ddf50), Attrs:map[string]string{"namespace":"default", "node":"10.0.0.50", "pod":"nfs-server-provisioner-0", "timestamp":"2024-12-13 01:28:37.76044468 +0000 UTC"}, Hostname:"10.0.0.50", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Dec 13 01:28:37.824659 containerd[1572]: 2024-12-13 01:28:37.771 [INFO][3284] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Dec 13 01:28:37.824659 containerd[1572]: 2024-12-13 01:28:37.771 [INFO][3284] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Dec 13 01:28:37.824659 containerd[1572]: 2024-12-13 01:28:37.771 [INFO][3284] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host '10.0.0.50' Dec 13 01:28:37.824659 containerd[1572]: 2024-12-13 01:28:37.773 [INFO][3284] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.1b0f2f5f77310cf8cd13d1699d9a80c57850e2c1ada2c656442e2a5d51c0e267" host="10.0.0.50" Dec 13 01:28:37.824659 containerd[1572]: 2024-12-13 01:28:37.777 [INFO][3284] ipam/ipam.go 372: Looking up existing affinities for host host="10.0.0.50" Dec 13 01:28:37.824659 containerd[1572]: 2024-12-13 01:28:37.781 [INFO][3284] ipam/ipam.go 489: Trying affinity for 192.168.55.192/26 host="10.0.0.50" Dec 13 01:28:37.824659 containerd[1572]: 2024-12-13 01:28:37.784 [INFO][3284] ipam/ipam.go 155: Attempting to load block cidr=192.168.55.192/26 host="10.0.0.50" Dec 13 01:28:37.824659 containerd[1572]: 2024-12-13 01:28:37.786 [INFO][3284] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.55.192/26 host="10.0.0.50" Dec 13 01:28:37.824659 containerd[1572]: 2024-12-13 01:28:37.786 [INFO][3284] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.55.192/26 handle="k8s-pod-network.1b0f2f5f77310cf8cd13d1699d9a80c57850e2c1ada2c656442e2a5d51c0e267" host="10.0.0.50" Dec 13 01:28:37.824659 containerd[1572]: 2024-12-13 01:28:37.789 [INFO][3284] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.1b0f2f5f77310cf8cd13d1699d9a80c57850e2c1ada2c656442e2a5d51c0e267 Dec 13 01:28:37.824659 containerd[1572]: 2024-12-13 01:28:37.794 [INFO][3284] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.55.192/26 handle="k8s-pod-network.1b0f2f5f77310cf8cd13d1699d9a80c57850e2c1ada2c656442e2a5d51c0e267" host="10.0.0.50" Dec 13 01:28:37.824659 containerd[1572]: 2024-12-13 01:28:37.800 [INFO][3284] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.55.195/26] block=192.168.55.192/26 handle="k8s-pod-network.1b0f2f5f77310cf8cd13d1699d9a80c57850e2c1ada2c656442e2a5d51c0e267" host="10.0.0.50" Dec 13 01:28:37.824659 containerd[1572]: 2024-12-13 01:28:37.800 [INFO][3284] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.55.195/26] handle="k8s-pod-network.1b0f2f5f77310cf8cd13d1699d9a80c57850e2c1ada2c656442e2a5d51c0e267" host="10.0.0.50" Dec 13 01:28:37.824659 containerd[1572]: 2024-12-13 01:28:37.800 [INFO][3284] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Dec 13 01:28:37.824659 containerd[1572]: 2024-12-13 01:28:37.800 [INFO][3284] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.55.195/26] IPv6=[] ContainerID="1b0f2f5f77310cf8cd13d1699d9a80c57850e2c1ada2c656442e2a5d51c0e267" HandleID="k8s-pod-network.1b0f2f5f77310cf8cd13d1699d9a80c57850e2c1ada2c656442e2a5d51c0e267" Workload="10.0.0.50-k8s-nfs--server--provisioner--0-eth0" Dec 13 01:28:37.825253 containerd[1572]: 2024-12-13 01:28:37.803 [INFO][3271] cni-plugin/k8s.go 386: Populated endpoint ContainerID="1b0f2f5f77310cf8cd13d1699d9a80c57850e2c1ada2c656442e2a5d51c0e267" Namespace="default" Pod="nfs-server-provisioner-0" WorkloadEndpoint="10.0.0.50-k8s-nfs--server--provisioner--0-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"10.0.0.50-k8s-nfs--server--provisioner--0-eth0", GenerateName:"nfs-server-provisioner-", Namespace:"default", SelfLink:"", UID:"344508b1-e3a2-474d-84b2-6194aac12a19", ResourceVersion:"1170", Generation:0, CreationTimestamp:time.Date(2024, time.December, 13, 1, 28, 37, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app":"nfs-server-provisioner", "apps.kubernetes.io/pod-index":"0", "chart":"nfs-server-provisioner-1.8.0", "controller-revision-hash":"nfs-server-provisioner-d5cbb7f57", "heritage":"Helm", "projectcalico.org/namespace":"default", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"nfs-server-provisioner", "release":"nfs-server-provisioner", "statefulset.kubernetes.io/pod-name":"nfs-server-provisioner-0"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"10.0.0.50", ContainerID:"", Pod:"nfs-server-provisioner-0", Endpoint:"eth0", ServiceAccountName:"nfs-server-provisioner", IPNetworks:[]string{"192.168.55.195/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.default", "ksa.default.nfs-server-provisioner"}, InterfaceName:"cali60e51b789ff", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"nfs", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x801, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"nfs-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x801, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"nlockmgr", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x8023, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"nlockmgr-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x8023, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"mountd", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x4e50, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"mountd-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x4e50, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"rquotad", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x36b, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"rquotad-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x36b, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"rpcbind", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x6f, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"rpcbind-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x6f, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"statd", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x296, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"statd-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x296, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Dec 13 01:28:37.825253 containerd[1572]: 2024-12-13 01:28:37.803 [INFO][3271] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.55.195/32] ContainerID="1b0f2f5f77310cf8cd13d1699d9a80c57850e2c1ada2c656442e2a5d51c0e267" Namespace="default" Pod="nfs-server-provisioner-0" WorkloadEndpoint="10.0.0.50-k8s-nfs--server--provisioner--0-eth0" Dec 13 01:28:37.825253 containerd[1572]: 2024-12-13 01:28:37.803 [INFO][3271] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali60e51b789ff ContainerID="1b0f2f5f77310cf8cd13d1699d9a80c57850e2c1ada2c656442e2a5d51c0e267" Namespace="default" Pod="nfs-server-provisioner-0" WorkloadEndpoint="10.0.0.50-k8s-nfs--server--provisioner--0-eth0" Dec 13 01:28:37.825253 containerd[1572]: 2024-12-13 01:28:37.806 [INFO][3271] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="1b0f2f5f77310cf8cd13d1699d9a80c57850e2c1ada2c656442e2a5d51c0e267" Namespace="default" Pod="nfs-server-provisioner-0" WorkloadEndpoint="10.0.0.50-k8s-nfs--server--provisioner--0-eth0" Dec 13 01:28:37.825485 containerd[1572]: 2024-12-13 01:28:37.807 [INFO][3271] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="1b0f2f5f77310cf8cd13d1699d9a80c57850e2c1ada2c656442e2a5d51c0e267" Namespace="default" Pod="nfs-server-provisioner-0" WorkloadEndpoint="10.0.0.50-k8s-nfs--server--provisioner--0-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"10.0.0.50-k8s-nfs--server--provisioner--0-eth0", GenerateName:"nfs-server-provisioner-", Namespace:"default", SelfLink:"", UID:"344508b1-e3a2-474d-84b2-6194aac12a19", ResourceVersion:"1170", Generation:0, CreationTimestamp:time.Date(2024, time.December, 13, 1, 28, 37, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app":"nfs-server-provisioner", "apps.kubernetes.io/pod-index":"0", "chart":"nfs-server-provisioner-1.8.0", "controller-revision-hash":"nfs-server-provisioner-d5cbb7f57", "heritage":"Helm", "projectcalico.org/namespace":"default", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"nfs-server-provisioner", "release":"nfs-server-provisioner", "statefulset.kubernetes.io/pod-name":"nfs-server-provisioner-0"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"10.0.0.50", ContainerID:"1b0f2f5f77310cf8cd13d1699d9a80c57850e2c1ada2c656442e2a5d51c0e267", Pod:"nfs-server-provisioner-0", Endpoint:"eth0", ServiceAccountName:"nfs-server-provisioner", IPNetworks:[]string{"192.168.55.195/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.default", "ksa.default.nfs-server-provisioner"}, InterfaceName:"cali60e51b789ff", MAC:"fa:21:26:e8:84:72", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"nfs", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x801, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"nfs-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x801, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"nlockmgr", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x8023, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"nlockmgr-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x8023, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"mountd", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x4e50, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"mountd-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x4e50, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"rquotad", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x36b, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"rquotad-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x36b, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"rpcbind", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x6f, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"rpcbind-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x6f, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"statd", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x296, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"statd-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x296, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Dec 13 01:28:37.825485 containerd[1572]: 2024-12-13 01:28:37.817 [INFO][3271] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="1b0f2f5f77310cf8cd13d1699d9a80c57850e2c1ada2c656442e2a5d51c0e267" Namespace="default" Pod="nfs-server-provisioner-0" WorkloadEndpoint="10.0.0.50-k8s-nfs--server--provisioner--0-eth0" Dec 13 01:28:37.847887 containerd[1572]: time="2024-12-13T01:28:37.847734386Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 01:28:37.847887 containerd[1572]: time="2024-12-13T01:28:37.847824999Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 01:28:37.848172 containerd[1572]: time="2024-12-13T01:28:37.847846753Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 01:28:37.848172 containerd[1572]: time="2024-12-13T01:28:37.847934921Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 01:28:37.876217 systemd-resolved[1462]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Dec 13 01:28:37.903364 containerd[1572]: time="2024-12-13T01:28:37.903320849Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nfs-server-provisioner-0,Uid:344508b1-e3a2-474d-84b2-6194aac12a19,Namespace:default,Attempt:0,} returns sandbox id \"1b0f2f5f77310cf8cd13d1699d9a80c57850e2c1ada2c656442e2a5d51c0e267\"" Dec 13 01:28:37.904659 containerd[1572]: time="2024-12-13T01:28:37.904612536Z" level=info msg="PullImage \"registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8\"" Dec 13 01:28:38.812735 kubelet[1918]: E1213 01:28:38.812662 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:39.083605 systemd-networkd[1255]: cali60e51b789ff: Gained IPv6LL Dec 13 01:28:39.813354 kubelet[1918]: E1213 01:28:39.813246 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:39.865085 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount153998661.mount: Deactivated successfully. Dec 13 01:28:40.814009 kubelet[1918]: E1213 01:28:40.813935 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:41.815001 kubelet[1918]: E1213 01:28:41.814908 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:42.815891 kubelet[1918]: E1213 01:28:42.815785 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:43.081537 containerd[1572]: time="2024-12-13T01:28:43.081302257Z" level=info msg="ImageCreate event name:\"registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 01:28:43.083601 containerd[1572]: time="2024-12-13T01:28:43.083463812Z" level=info msg="stop pulling image registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8: active requests=0, bytes read=91039406" Dec 13 01:28:43.085891 containerd[1572]: time="2024-12-13T01:28:43.085296804Z" level=info msg="ImageCreate event name:\"sha256:fd0b16f70b66b72bcb2f91d556fa33eba02729c44ffc5f2c16130e7f9fbed3c4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 01:28:43.091337 containerd[1572]: time="2024-12-13T01:28:43.091251094Z" level=info msg="ImageCreate event name:\"registry.k8s.io/sig-storage/nfs-provisioner@sha256:c825f3d5e28bde099bd7a3daace28772d412c9157ad47fa752a9ad0baafc118d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 01:28:43.092825 containerd[1572]: time="2024-12-13T01:28:43.092769882Z" level=info msg="Pulled image \"registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8\" with image id \"sha256:fd0b16f70b66b72bcb2f91d556fa33eba02729c44ffc5f2c16130e7f9fbed3c4\", repo tag \"registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8\", repo digest \"registry.k8s.io/sig-storage/nfs-provisioner@sha256:c825f3d5e28bde099bd7a3daace28772d412c9157ad47fa752a9ad0baafc118d\", size \"91036984\" in 5.188124961s" Dec 13 01:28:43.092825 containerd[1572]: time="2024-12-13T01:28:43.092822447Z" level=info msg="PullImage \"registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8\" returns image reference \"sha256:fd0b16f70b66b72bcb2f91d556fa33eba02729c44ffc5f2c16130e7f9fbed3c4\"" Dec 13 01:28:43.095355 containerd[1572]: time="2024-12-13T01:28:43.095284899Z" level=info msg="CreateContainer within sandbox \"1b0f2f5f77310cf8cd13d1699d9a80c57850e2c1ada2c656442e2a5d51c0e267\" for container &ContainerMetadata{Name:nfs-server-provisioner,Attempt:0,}" Dec 13 01:28:43.123593 containerd[1572]: time="2024-12-13T01:28:43.123460231Z" level=info msg="CreateContainer within sandbox \"1b0f2f5f77310cf8cd13d1699d9a80c57850e2c1ada2c656442e2a5d51c0e267\" for &ContainerMetadata{Name:nfs-server-provisioner,Attempt:0,} returns container id \"719c9514f55492a9236674a5a45aac9c923943f42390328d9958197567374d60\"" Dec 13 01:28:43.124503 containerd[1572]: time="2024-12-13T01:28:43.124434938Z" level=info msg="StartContainer for \"719c9514f55492a9236674a5a45aac9c923943f42390328d9958197567374d60\"" Dec 13 01:28:43.202589 containerd[1572]: time="2024-12-13T01:28:43.202436360Z" level=info msg="StartContainer for \"719c9514f55492a9236674a5a45aac9c923943f42390328d9958197567374d60\" returns successfully" Dec 13 01:28:43.466343 kubelet[1918]: I1213 01:28:43.466126 1918 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="default/nfs-server-provisioner-0" podStartSLOduration=1.2773321530000001 podStartE2EDuration="6.466064968s" podCreationTimestamp="2024-12-13 01:28:37 +0000 UTC" firstStartedPulling="2024-12-13 01:28:37.904433735 +0000 UTC m=+47.481813990" lastFinishedPulling="2024-12-13 01:28:43.09316654 +0000 UTC m=+52.670546805" observedRunningTime="2024-12-13 01:28:43.465509615 +0000 UTC m=+53.042889880" watchObservedRunningTime="2024-12-13 01:28:43.466064968 +0000 UTC m=+53.043445244" Dec 13 01:28:43.817171 kubelet[1918]: E1213 01:28:43.817074 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:44.818198 kubelet[1918]: E1213 01:28:44.818120 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:45.819284 kubelet[1918]: E1213 01:28:45.819220 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:46.819938 kubelet[1918]: E1213 01:28:46.819845 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:47.820713 kubelet[1918]: E1213 01:28:47.820638 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:48.821727 kubelet[1918]: E1213 01:28:48.821407 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:49.821937 kubelet[1918]: E1213 01:28:49.821847 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:50.776980 kubelet[1918]: E1213 01:28:50.776876 1918 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:50.794476 containerd[1572]: time="2024-12-13T01:28:50.794417633Z" level=info msg="StopPodSandbox for \"1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30\"" Dec 13 01:28:50.822460 kubelet[1918]: E1213 01:28:50.822365 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:50.915570 containerd[1572]: 2024-12-13 01:28:50.856 [WARNING][3475] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"10.0.0.50-k8s-csi--node--driver--nmw4l-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"8e158755-5b09-4a6a-ba0b-890e0f7dd097", ResourceVersion:"1147", Generation:0, CreationTimestamp:time.Date(2024, time.December, 13, 1, 27, 51, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"55b695c467", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"10.0.0.50", ContainerID:"ce5b479e99d74027426bb31afccb63abafe070101864964785e2718542123558", Pod:"csi-node-driver-nmw4l", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.55.193/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali701ca78a3ee", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Dec 13 01:28:50.915570 containerd[1572]: 2024-12-13 01:28:50.857 [INFO][3475] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30" Dec 13 01:28:50.915570 containerd[1572]: 2024-12-13 01:28:50.857 [INFO][3475] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30" iface="eth0" netns="" Dec 13 01:28:50.915570 containerd[1572]: 2024-12-13 01:28:50.857 [INFO][3475] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30" Dec 13 01:28:50.915570 containerd[1572]: 2024-12-13 01:28:50.857 [INFO][3475] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30" Dec 13 01:28:50.915570 containerd[1572]: 2024-12-13 01:28:50.888 [INFO][3483] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30" HandleID="k8s-pod-network.1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30" Workload="10.0.0.50-k8s-csi--node--driver--nmw4l-eth0" Dec 13 01:28:50.915570 containerd[1572]: 2024-12-13 01:28:50.888 [INFO][3483] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Dec 13 01:28:50.915570 containerd[1572]: 2024-12-13 01:28:50.888 [INFO][3483] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Dec 13 01:28:50.915570 containerd[1572]: 2024-12-13 01:28:50.900 [WARNING][3483] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30" HandleID="k8s-pod-network.1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30" Workload="10.0.0.50-k8s-csi--node--driver--nmw4l-eth0" Dec 13 01:28:50.915570 containerd[1572]: 2024-12-13 01:28:50.900 [INFO][3483] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30" HandleID="k8s-pod-network.1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30" Workload="10.0.0.50-k8s-csi--node--driver--nmw4l-eth0" Dec 13 01:28:50.915570 containerd[1572]: 2024-12-13 01:28:50.905 [INFO][3483] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Dec 13 01:28:50.915570 containerd[1572]: 2024-12-13 01:28:50.908 [INFO][3475] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30" Dec 13 01:28:50.915570 containerd[1572]: time="2024-12-13T01:28:50.915587905Z" level=info msg="TearDown network for sandbox \"1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30\" successfully" Dec 13 01:28:50.915570 containerd[1572]: time="2024-12-13T01:28:50.915628665Z" level=info msg="StopPodSandbox for \"1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30\" returns successfully" Dec 13 01:28:50.918093 containerd[1572]: time="2024-12-13T01:28:50.916773071Z" level=info msg="RemovePodSandbox for \"1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30\"" Dec 13 01:28:50.918093 containerd[1572]: time="2024-12-13T01:28:50.916816856Z" level=info msg="Forcibly stopping sandbox \"1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30\"" Dec 13 01:28:51.031217 containerd[1572]: 2024-12-13 01:28:50.980 [WARNING][3504] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"10.0.0.50-k8s-csi--node--driver--nmw4l-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"8e158755-5b09-4a6a-ba0b-890e0f7dd097", ResourceVersion:"1147", Generation:0, CreationTimestamp:time.Date(2024, time.December, 13, 1, 27, 51, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"55b695c467", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"10.0.0.50", ContainerID:"ce5b479e99d74027426bb31afccb63abafe070101864964785e2718542123558", Pod:"csi-node-driver-nmw4l", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.55.193/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali701ca78a3ee", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Dec 13 01:28:51.031217 containerd[1572]: 2024-12-13 01:28:50.981 [INFO][3504] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30" Dec 13 01:28:51.031217 containerd[1572]: 2024-12-13 01:28:50.981 [INFO][3504] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30" iface="eth0" netns="" Dec 13 01:28:51.031217 containerd[1572]: 2024-12-13 01:28:50.981 [INFO][3504] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30" Dec 13 01:28:51.031217 containerd[1572]: 2024-12-13 01:28:50.981 [INFO][3504] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30" Dec 13 01:28:51.031217 containerd[1572]: 2024-12-13 01:28:51.013 [INFO][3511] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30" HandleID="k8s-pod-network.1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30" Workload="10.0.0.50-k8s-csi--node--driver--nmw4l-eth0" Dec 13 01:28:51.031217 containerd[1572]: 2024-12-13 01:28:51.014 [INFO][3511] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Dec 13 01:28:51.031217 containerd[1572]: 2024-12-13 01:28:51.014 [INFO][3511] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Dec 13 01:28:51.031217 containerd[1572]: 2024-12-13 01:28:51.022 [WARNING][3511] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30" HandleID="k8s-pod-network.1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30" Workload="10.0.0.50-k8s-csi--node--driver--nmw4l-eth0" Dec 13 01:28:51.031217 containerd[1572]: 2024-12-13 01:28:51.022 [INFO][3511] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30" HandleID="k8s-pod-network.1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30" Workload="10.0.0.50-k8s-csi--node--driver--nmw4l-eth0" Dec 13 01:28:51.031217 containerd[1572]: 2024-12-13 01:28:51.025 [INFO][3511] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Dec 13 01:28:51.031217 containerd[1572]: 2024-12-13 01:28:51.028 [INFO][3504] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30" Dec 13 01:28:51.031217 containerd[1572]: time="2024-12-13T01:28:51.031157705Z" level=info msg="TearDown network for sandbox \"1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30\" successfully" Dec 13 01:28:51.062433 containerd[1572]: time="2024-12-13T01:28:51.062339751Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Dec 13 01:28:51.062433 containerd[1572]: time="2024-12-13T01:28:51.062462321Z" level=info msg="RemovePodSandbox \"1ea1154cc9108839002b9f99a1c8246022b85784fcd211f04c8fefc80d603c30\" returns successfully" Dec 13 01:28:51.063474 containerd[1572]: time="2024-12-13T01:28:51.063423986Z" level=info msg="StopPodSandbox for \"eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e\"" Dec 13 01:28:51.179014 containerd[1572]: 2024-12-13 01:28:51.128 [WARNING][3533] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"10.0.0.50-k8s-nginx--deployment--6d5f899847--785tc-eth0", GenerateName:"nginx-deployment-6d5f899847-", Namespace:"default", SelfLink:"", UID:"2767e440-abcd-45e1-95b3-0943b03016cf", ResourceVersion:"1125", Generation:0, CreationTimestamp:time.Date(2024, time.December, 13, 1, 28, 8, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app":"nginx", "pod-template-hash":"6d5f899847", "projectcalico.org/namespace":"default", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"default"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"10.0.0.50", ContainerID:"a3614f4cfd7eba109151c53d576b36a7a51f0c23fde2eff9309e1756ce745727", Pod:"nginx-deployment-6d5f899847-785tc", Endpoint:"eth0", ServiceAccountName:"default", IPNetworks:[]string{"192.168.55.194/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.default", "ksa.default.default"}, InterfaceName:"califcfaaceb738", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Dec 13 01:28:51.179014 containerd[1572]: 2024-12-13 01:28:51.129 [INFO][3533] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e" Dec 13 01:28:51.179014 containerd[1572]: 2024-12-13 01:28:51.129 [INFO][3533] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e" iface="eth0" netns="" Dec 13 01:28:51.179014 containerd[1572]: 2024-12-13 01:28:51.129 [INFO][3533] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e" Dec 13 01:28:51.179014 containerd[1572]: 2024-12-13 01:28:51.129 [INFO][3533] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e" Dec 13 01:28:51.179014 containerd[1572]: 2024-12-13 01:28:51.160 [INFO][3540] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e" HandleID="k8s-pod-network.eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e" Workload="10.0.0.50-k8s-nginx--deployment--6d5f899847--785tc-eth0" Dec 13 01:28:51.179014 containerd[1572]: 2024-12-13 01:28:51.161 [INFO][3540] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Dec 13 01:28:51.179014 containerd[1572]: 2024-12-13 01:28:51.161 [INFO][3540] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Dec 13 01:28:51.179014 containerd[1572]: 2024-12-13 01:28:51.170 [WARNING][3540] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e" HandleID="k8s-pod-network.eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e" Workload="10.0.0.50-k8s-nginx--deployment--6d5f899847--785tc-eth0" Dec 13 01:28:51.179014 containerd[1572]: 2024-12-13 01:28:51.170 [INFO][3540] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e" HandleID="k8s-pod-network.eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e" Workload="10.0.0.50-k8s-nginx--deployment--6d5f899847--785tc-eth0" Dec 13 01:28:51.179014 containerd[1572]: 2024-12-13 01:28:51.172 [INFO][3540] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Dec 13 01:28:51.179014 containerd[1572]: 2024-12-13 01:28:51.175 [INFO][3533] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e" Dec 13 01:28:51.179750 containerd[1572]: time="2024-12-13T01:28:51.179079938Z" level=info msg="TearDown network for sandbox \"eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e\" successfully" Dec 13 01:28:51.179750 containerd[1572]: time="2024-12-13T01:28:51.179116730Z" level=info msg="StopPodSandbox for \"eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e\" returns successfully" Dec 13 01:28:51.180646 containerd[1572]: time="2024-12-13T01:28:51.180591611Z" level=info msg="RemovePodSandbox for \"eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e\"" Dec 13 01:28:51.180646 containerd[1572]: time="2024-12-13T01:28:51.180635267Z" level=info msg="Forcibly stopping sandbox \"eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e\"" Dec 13 01:28:51.327379 containerd[1572]: 2024-12-13 01:28:51.253 [WARNING][3561] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"10.0.0.50-k8s-nginx--deployment--6d5f899847--785tc-eth0", GenerateName:"nginx-deployment-6d5f899847-", Namespace:"default", SelfLink:"", UID:"2767e440-abcd-45e1-95b3-0943b03016cf", ResourceVersion:"1125", Generation:0, CreationTimestamp:time.Date(2024, time.December, 13, 1, 28, 8, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app":"nginx", "pod-template-hash":"6d5f899847", "projectcalico.org/namespace":"default", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"default"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"10.0.0.50", ContainerID:"a3614f4cfd7eba109151c53d576b36a7a51f0c23fde2eff9309e1756ce745727", Pod:"nginx-deployment-6d5f899847-785tc", Endpoint:"eth0", ServiceAccountName:"default", IPNetworks:[]string{"192.168.55.194/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.default", "ksa.default.default"}, InterfaceName:"califcfaaceb738", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Dec 13 01:28:51.327379 containerd[1572]: 2024-12-13 01:28:51.253 [INFO][3561] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e" Dec 13 01:28:51.327379 containerd[1572]: 2024-12-13 01:28:51.253 [INFO][3561] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e" iface="eth0" netns="" Dec 13 01:28:51.327379 containerd[1572]: 2024-12-13 01:28:51.253 [INFO][3561] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e" Dec 13 01:28:51.327379 containerd[1572]: 2024-12-13 01:28:51.253 [INFO][3561] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e" Dec 13 01:28:51.327379 containerd[1572]: 2024-12-13 01:28:51.295 [INFO][3569] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e" HandleID="k8s-pod-network.eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e" Workload="10.0.0.50-k8s-nginx--deployment--6d5f899847--785tc-eth0" Dec 13 01:28:51.327379 containerd[1572]: 2024-12-13 01:28:51.295 [INFO][3569] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Dec 13 01:28:51.327379 containerd[1572]: 2024-12-13 01:28:51.295 [INFO][3569] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Dec 13 01:28:51.327379 containerd[1572]: 2024-12-13 01:28:51.311 [WARNING][3569] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e" HandleID="k8s-pod-network.eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e" Workload="10.0.0.50-k8s-nginx--deployment--6d5f899847--785tc-eth0" Dec 13 01:28:51.327379 containerd[1572]: 2024-12-13 01:28:51.311 [INFO][3569] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e" HandleID="k8s-pod-network.eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e" Workload="10.0.0.50-k8s-nginx--deployment--6d5f899847--785tc-eth0" Dec 13 01:28:51.327379 containerd[1572]: 2024-12-13 01:28:51.316 [INFO][3569] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Dec 13 01:28:51.327379 containerd[1572]: 2024-12-13 01:28:51.323 [INFO][3561] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e" Dec 13 01:28:51.328235 containerd[1572]: time="2024-12-13T01:28:51.327385713Z" level=info msg="TearDown network for sandbox \"eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e\" successfully" Dec 13 01:28:51.338013 containerd[1572]: time="2024-12-13T01:28:51.337682393Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Dec 13 01:28:51.338013 containerd[1572]: time="2024-12-13T01:28:51.337788150Z" level=info msg="RemovePodSandbox \"eb403d57d9d7fbd88bc0e6b1617ee39aa660ef21aefaf36e385809876615125e\" returns successfully" Dec 13 01:28:51.823286 kubelet[1918]: E1213 01:28:51.823097 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:52.824273 kubelet[1918]: E1213 01:28:52.824195 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:52.889213 kubelet[1918]: I1213 01:28:52.889052 1918 topology_manager.go:215] "Topology Admit Handler" podUID="52bdeae0-0744-4747-a19e-89a0d8532061" podNamespace="default" podName="test-pod-1" Dec 13 01:28:53.091133 kubelet[1918]: I1213 01:28:53.090738 1918 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lj9tz\" (UniqueName: \"kubernetes.io/projected/52bdeae0-0744-4747-a19e-89a0d8532061-kube-api-access-lj9tz\") pod \"test-pod-1\" (UID: \"52bdeae0-0744-4747-a19e-89a0d8532061\") " pod="default/test-pod-1" Dec 13 01:28:53.091133 kubelet[1918]: I1213 01:28:53.090820 1918 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-a0c8dee0-607a-43b1-9739-20536bd42fd1\" (UniqueName: \"kubernetes.io/nfs/52bdeae0-0744-4747-a19e-89a0d8532061-pvc-a0c8dee0-607a-43b1-9739-20536bd42fd1\") pod \"test-pod-1\" (UID: \"52bdeae0-0744-4747-a19e-89a0d8532061\") " pod="default/test-pod-1" Dec 13 01:28:53.289112 kernel: FS-Cache: Loaded Dec 13 01:28:53.404310 kernel: RPC: Registered named UNIX socket transport module. Dec 13 01:28:53.404466 kernel: RPC: Registered udp transport module. Dec 13 01:28:53.404496 kernel: RPC: Registered tcp transport module. Dec 13 01:28:53.406077 kernel: RPC: Registered tcp-with-tls transport module. Dec 13 01:28:53.406147 kernel: RPC: Registered tcp NFSv4.1 backchannel transport module. Dec 13 01:28:53.710719 kernel: NFS: Registering the id_resolver key type Dec 13 01:28:53.710866 kernel: Key type id_resolver registered Dec 13 01:28:53.710937 kernel: Key type id_legacy registered Dec 13 01:28:53.754488 nfsidmap[3594]: nss_getpwnam: name 'root@nfs-server-provisioner.default.svc.cluster.local' does not map into domain 'localdomain' Dec 13 01:28:53.761796 nfsidmap[3597]: nss_name_to_gid: name 'root@nfs-server-provisioner.default.svc.cluster.local' does not map into domain 'localdomain' Dec 13 01:28:53.797461 containerd[1572]: time="2024-12-13T01:28:53.795390418Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:test-pod-1,Uid:52bdeae0-0744-4747-a19e-89a0d8532061,Namespace:default,Attempt:0,}" Dec 13 01:28:53.825297 kubelet[1918]: E1213 01:28:53.825164 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:54.024692 systemd-networkd[1255]: cali5ec59c6bf6e: Link UP Dec 13 01:28:54.025513 systemd-networkd[1255]: cali5ec59c6bf6e: Gained carrier Dec 13 01:28:54.039655 containerd[1572]: 2024-12-13 01:28:53.902 [INFO][3601] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {10.0.0.50-k8s-test--pod--1-eth0 default 52bdeae0-0744-4747-a19e-89a0d8532061 1239 0 2024-12-13 01:28:37 +0000 UTC map[projectcalico.org/namespace:default projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:default] map[] [] [] []} {k8s 10.0.0.50 test-pod-1 eth0 default [] [] [kns.default ksa.default.default] cali5ec59c6bf6e [] []}} ContainerID="5fabae10d0d437424cfada7d21cd4f0adc3c24d090174ff15b9bfcaf715810d7" Namespace="default" Pod="test-pod-1" WorkloadEndpoint="10.0.0.50-k8s-test--pod--1-" Dec 13 01:28:54.039655 containerd[1572]: 2024-12-13 01:28:53.902 [INFO][3601] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="5fabae10d0d437424cfada7d21cd4f0adc3c24d090174ff15b9bfcaf715810d7" Namespace="default" Pod="test-pod-1" WorkloadEndpoint="10.0.0.50-k8s-test--pod--1-eth0" Dec 13 01:28:54.039655 containerd[1572]: 2024-12-13 01:28:53.941 [INFO][3614] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="5fabae10d0d437424cfada7d21cd4f0adc3c24d090174ff15b9bfcaf715810d7" HandleID="k8s-pod-network.5fabae10d0d437424cfada7d21cd4f0adc3c24d090174ff15b9bfcaf715810d7" Workload="10.0.0.50-k8s-test--pod--1-eth0" Dec 13 01:28:54.039655 containerd[1572]: 2024-12-13 01:28:53.954 [INFO][3614] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="5fabae10d0d437424cfada7d21cd4f0adc3c24d090174ff15b9bfcaf715810d7" HandleID="k8s-pod-network.5fabae10d0d437424cfada7d21cd4f0adc3c24d090174ff15b9bfcaf715810d7" Workload="10.0.0.50-k8s-test--pod--1-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000502380), Attrs:map[string]string{"namespace":"default", "node":"10.0.0.50", "pod":"test-pod-1", "timestamp":"2024-12-13 01:28:53.941595459 +0000 UTC"}, Hostname:"10.0.0.50", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Dec 13 01:28:54.039655 containerd[1572]: 2024-12-13 01:28:53.954 [INFO][3614] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Dec 13 01:28:54.039655 containerd[1572]: 2024-12-13 01:28:53.954 [INFO][3614] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Dec 13 01:28:54.039655 containerd[1572]: 2024-12-13 01:28:53.954 [INFO][3614] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host '10.0.0.50' Dec 13 01:28:54.039655 containerd[1572]: 2024-12-13 01:28:53.958 [INFO][3614] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.5fabae10d0d437424cfada7d21cd4f0adc3c24d090174ff15b9bfcaf715810d7" host="10.0.0.50" Dec 13 01:28:54.039655 containerd[1572]: 2024-12-13 01:28:53.968 [INFO][3614] ipam/ipam.go 372: Looking up existing affinities for host host="10.0.0.50" Dec 13 01:28:54.039655 containerd[1572]: 2024-12-13 01:28:53.980 [INFO][3614] ipam/ipam.go 489: Trying affinity for 192.168.55.192/26 host="10.0.0.50" Dec 13 01:28:54.039655 containerd[1572]: 2024-12-13 01:28:53.984 [INFO][3614] ipam/ipam.go 155: Attempting to load block cidr=192.168.55.192/26 host="10.0.0.50" Dec 13 01:28:54.039655 containerd[1572]: 2024-12-13 01:28:53.988 [INFO][3614] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.55.192/26 host="10.0.0.50" Dec 13 01:28:54.039655 containerd[1572]: 2024-12-13 01:28:53.989 [INFO][3614] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.55.192/26 handle="k8s-pod-network.5fabae10d0d437424cfada7d21cd4f0adc3c24d090174ff15b9bfcaf715810d7" host="10.0.0.50" Dec 13 01:28:54.039655 containerd[1572]: 2024-12-13 01:28:53.993 [INFO][3614] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.5fabae10d0d437424cfada7d21cd4f0adc3c24d090174ff15b9bfcaf715810d7 Dec 13 01:28:54.039655 containerd[1572]: 2024-12-13 01:28:54.001 [INFO][3614] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.55.192/26 handle="k8s-pod-network.5fabae10d0d437424cfada7d21cd4f0adc3c24d090174ff15b9bfcaf715810d7" host="10.0.0.50" Dec 13 01:28:54.039655 containerd[1572]: 2024-12-13 01:28:54.018 [INFO][3614] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.55.196/26] block=192.168.55.192/26 handle="k8s-pod-network.5fabae10d0d437424cfada7d21cd4f0adc3c24d090174ff15b9bfcaf715810d7" host="10.0.0.50" Dec 13 01:28:54.039655 containerd[1572]: 2024-12-13 01:28:54.018 [INFO][3614] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.55.196/26] handle="k8s-pod-network.5fabae10d0d437424cfada7d21cd4f0adc3c24d090174ff15b9bfcaf715810d7" host="10.0.0.50" Dec 13 01:28:54.039655 containerd[1572]: 2024-12-13 01:28:54.019 [INFO][3614] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Dec 13 01:28:54.039655 containerd[1572]: 2024-12-13 01:28:54.019 [INFO][3614] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.55.196/26] IPv6=[] ContainerID="5fabae10d0d437424cfada7d21cd4f0adc3c24d090174ff15b9bfcaf715810d7" HandleID="k8s-pod-network.5fabae10d0d437424cfada7d21cd4f0adc3c24d090174ff15b9bfcaf715810d7" Workload="10.0.0.50-k8s-test--pod--1-eth0" Dec 13 01:28:54.039655 containerd[1572]: 2024-12-13 01:28:54.022 [INFO][3601] cni-plugin/k8s.go 386: Populated endpoint ContainerID="5fabae10d0d437424cfada7d21cd4f0adc3c24d090174ff15b9bfcaf715810d7" Namespace="default" Pod="test-pod-1" WorkloadEndpoint="10.0.0.50-k8s-test--pod--1-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"10.0.0.50-k8s-test--pod--1-eth0", GenerateName:"", Namespace:"default", SelfLink:"", UID:"52bdeae0-0744-4747-a19e-89a0d8532061", ResourceVersion:"1239", Generation:0, CreationTimestamp:time.Date(2024, time.December, 13, 1, 28, 37, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"projectcalico.org/namespace":"default", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"default"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"10.0.0.50", ContainerID:"", Pod:"test-pod-1", Endpoint:"eth0", ServiceAccountName:"default", IPNetworks:[]string{"192.168.55.196/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.default", "ksa.default.default"}, InterfaceName:"cali5ec59c6bf6e", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Dec 13 01:28:54.040405 containerd[1572]: 2024-12-13 01:28:54.022 [INFO][3601] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.55.196/32] ContainerID="5fabae10d0d437424cfada7d21cd4f0adc3c24d090174ff15b9bfcaf715810d7" Namespace="default" Pod="test-pod-1" WorkloadEndpoint="10.0.0.50-k8s-test--pod--1-eth0" Dec 13 01:28:54.040405 containerd[1572]: 2024-12-13 01:28:54.022 [INFO][3601] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali5ec59c6bf6e ContainerID="5fabae10d0d437424cfada7d21cd4f0adc3c24d090174ff15b9bfcaf715810d7" Namespace="default" Pod="test-pod-1" WorkloadEndpoint="10.0.0.50-k8s-test--pod--1-eth0" Dec 13 01:28:54.040405 containerd[1572]: 2024-12-13 01:28:54.024 [INFO][3601] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="5fabae10d0d437424cfada7d21cd4f0adc3c24d090174ff15b9bfcaf715810d7" Namespace="default" Pod="test-pod-1" WorkloadEndpoint="10.0.0.50-k8s-test--pod--1-eth0" Dec 13 01:28:54.040405 containerd[1572]: 2024-12-13 01:28:54.025 [INFO][3601] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="5fabae10d0d437424cfada7d21cd4f0adc3c24d090174ff15b9bfcaf715810d7" Namespace="default" Pod="test-pod-1" WorkloadEndpoint="10.0.0.50-k8s-test--pod--1-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"10.0.0.50-k8s-test--pod--1-eth0", GenerateName:"", Namespace:"default", SelfLink:"", UID:"52bdeae0-0744-4747-a19e-89a0d8532061", ResourceVersion:"1239", Generation:0, CreationTimestamp:time.Date(2024, time.December, 13, 1, 28, 37, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"projectcalico.org/namespace":"default", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"default"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"10.0.0.50", ContainerID:"5fabae10d0d437424cfada7d21cd4f0adc3c24d090174ff15b9bfcaf715810d7", Pod:"test-pod-1", Endpoint:"eth0", ServiceAccountName:"default", IPNetworks:[]string{"192.168.55.196/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.default", "ksa.default.default"}, InterfaceName:"cali5ec59c6bf6e", MAC:"c2:f2:2b:66:fb:8e", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Dec 13 01:28:54.040405 containerd[1572]: 2024-12-13 01:28:54.036 [INFO][3601] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="5fabae10d0d437424cfada7d21cd4f0adc3c24d090174ff15b9bfcaf715810d7" Namespace="default" Pod="test-pod-1" WorkloadEndpoint="10.0.0.50-k8s-test--pod--1-eth0" Dec 13 01:28:54.065778 containerd[1572]: time="2024-12-13T01:28:54.065428038Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 01:28:54.065778 containerd[1572]: time="2024-12-13T01:28:54.065535218Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 01:28:54.065778 containerd[1572]: time="2024-12-13T01:28:54.065560928Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 01:28:54.065778 containerd[1572]: time="2024-12-13T01:28:54.065714688Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 01:28:54.097106 systemd-resolved[1462]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Dec 13 01:28:54.128372 containerd[1572]: time="2024-12-13T01:28:54.128312579Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:test-pod-1,Uid:52bdeae0-0744-4747-a19e-89a0d8532061,Namespace:default,Attempt:0,} returns sandbox id \"5fabae10d0d437424cfada7d21cd4f0adc3c24d090174ff15b9bfcaf715810d7\"" Dec 13 01:28:54.129995 containerd[1572]: time="2024-12-13T01:28:54.129971979Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\"" Dec 13 01:28:54.499685 containerd[1572]: time="2024-12-13T01:28:54.499605165Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/nginx:latest\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 01:28:54.501794 containerd[1572]: time="2024-12-13T01:28:54.501256068Z" level=info msg="stop pulling image ghcr.io/flatcar/nginx:latest: active requests=0, bytes read=61" Dec 13 01:28:54.505231 containerd[1572]: time="2024-12-13T01:28:54.505109885Z" level=info msg="Pulled image \"ghcr.io/flatcar/nginx:latest\" with image id \"sha256:fa0a8cea5e76ad962111c39c85bb312edaf5b89eccd8f404eeea66c9759641e3\", repo tag \"ghcr.io/flatcar/nginx:latest\", repo digest \"ghcr.io/flatcar/nginx@sha256:e04edf30a4ea4c5a4107110797c72d3ee8a654415f00acd4019be17218afd9a1\", size \"71035905\" in 375.106284ms" Dec 13 01:28:54.505231 containerd[1572]: time="2024-12-13T01:28:54.505161516Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\" returns image reference \"sha256:fa0a8cea5e76ad962111c39c85bb312edaf5b89eccd8f404eeea66c9759641e3\"" Dec 13 01:28:54.507092 containerd[1572]: time="2024-12-13T01:28:54.507048891Z" level=info msg="CreateContainer within sandbox \"5fabae10d0d437424cfada7d21cd4f0adc3c24d090174ff15b9bfcaf715810d7\" for container &ContainerMetadata{Name:test,Attempt:0,}" Dec 13 01:28:54.533191 containerd[1572]: time="2024-12-13T01:28:54.533121269Z" level=info msg="CreateContainer within sandbox \"5fabae10d0d437424cfada7d21cd4f0adc3c24d090174ff15b9bfcaf715810d7\" for &ContainerMetadata{Name:test,Attempt:0,} returns container id \"12ad99eba94310ed89634f6be0e8f054682a87515df6e38625528d7ccbe8a843\"" Dec 13 01:28:54.534305 containerd[1572]: time="2024-12-13T01:28:54.534222280Z" level=info msg="StartContainer for \"12ad99eba94310ed89634f6be0e8f054682a87515df6e38625528d7ccbe8a843\"" Dec 13 01:28:54.610935 containerd[1572]: time="2024-12-13T01:28:54.610758482Z" level=info msg="StartContainer for \"12ad99eba94310ed89634f6be0e8f054682a87515df6e38625528d7ccbe8a843\" returns successfully" Dec 13 01:28:54.826018 kubelet[1918]: E1213 01:28:54.825933 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:55.278372 systemd-networkd[1255]: cali5ec59c6bf6e: Gained IPv6LL Dec 13 01:28:55.826925 kubelet[1918]: E1213 01:28:55.826573 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:56.827179 kubelet[1918]: E1213 01:28:56.827000 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:57.828443 kubelet[1918]: E1213 01:28:57.828251 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:58.828909 kubelet[1918]: E1213 01:28:58.828728 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:28:59.829285 kubelet[1918]: E1213 01:28:59.829116 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 01:29:00.830512 kubelet[1918]: E1213 01:29:00.830366 1918 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests"