Mar 20 21:30:06.873819 kernel: Linux version 6.6.83-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.2.1_p20241221 p7) 14.2.1 20241221, GNU ld (Gentoo 2.44 p1) 2.44.0) #1 SMP PREEMPT_DYNAMIC Thu Mar 20 19:36:47 -00 2025
Mar 20 21:30:06.873841 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=619bfa043b53ac975036e415994a80721794ae8277072d0a93c174b4f7768019
Mar 20 21:30:06.873852 kernel: BIOS-provided physical RAM map:
Mar 20 21:30:06.873859 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable
Mar 20 21:30:06.873866 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved
Mar 20 21:30:06.873872 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved
Mar 20 21:30:06.873880 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000009cfdbfff] usable
Mar 20 21:30:06.873887 kernel: BIOS-e820: [mem 0x000000009cfdc000-0x000000009cffffff] reserved
Mar 20 21:30:06.873893 kernel: BIOS-e820: [mem 0x00000000b0000000-0x00000000bfffffff] reserved
Mar 20 21:30:06.873900 kernel: BIOS-e820: [mem 0x00000000fed1c000-0x00000000fed1ffff] reserved
Mar 20 21:30:06.873909 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved
Mar 20 21:30:06.873915 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved
Mar 20 21:30:06.873922 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved
Mar 20 21:30:06.873928 kernel: NX (Execute Disable) protection: active
Mar 20 21:30:06.873936 kernel: APIC: Static calls initialized
Mar 20 21:30:06.873946 kernel: SMBIOS 2.8 present.
Mar 20 21:30:06.873953 kernel: DMI: QEMU Standard PC (Q35 + ICH9, 2009), BIOS 1.16.2-debian-1.16.2-1 04/01/2014
Mar 20 21:30:06.873960 kernel: Hypervisor detected: KVM
Mar 20 21:30:06.873968 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00
Mar 20 21:30:06.873975 kernel: kvm-clock: using sched offset of 2341357465 cycles
Mar 20 21:30:06.873982 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns
Mar 20 21:30:06.873990 kernel: tsc: Detected 2794.750 MHz processor
Mar 20 21:30:06.873997 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved
Mar 20 21:30:06.874005 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable
Mar 20 21:30:06.874012 kernel: last_pfn = 0x9cfdc max_arch_pfn = 0x400000000
Mar 20 21:30:06.874022 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs
Mar 20 21:30:06.874029 kernel: x86/PAT: Configuration [0-7]: WB  WC  UC- UC  WB  WP  UC- WT  
Mar 20 21:30:06.874037 kernel: Using GB pages for direct mapping
Mar 20 21:30:06.874044 kernel: ACPI: Early table checksum verification disabled
Mar 20 21:30:06.874052 kernel: ACPI: RSDP 0x00000000000F59D0 000014 (v00 BOCHS )
Mar 20 21:30:06.874059 kernel: ACPI: RSDT 0x000000009CFE2408 000038 (v01 BOCHS  BXPC     00000001 BXPC 00000001)
Mar 20 21:30:06.874066 kernel: ACPI: FACP 0x000000009CFE21E8 0000F4 (v03 BOCHS  BXPC     00000001 BXPC 00000001)
Mar 20 21:30:06.874074 kernel: ACPI: DSDT 0x000000009CFE0040 0021A8 (v01 BOCHS  BXPC     00000001 BXPC 00000001)
Mar 20 21:30:06.874081 kernel: ACPI: FACS 0x000000009CFE0000 000040
Mar 20 21:30:06.874091 kernel: ACPI: APIC 0x000000009CFE22DC 000090 (v01 BOCHS  BXPC     00000001 BXPC 00000001)
Mar 20 21:30:06.874098 kernel: ACPI: HPET 0x000000009CFE236C 000038 (v01 BOCHS  BXPC     00000001 BXPC 00000001)
Mar 20 21:30:06.874106 kernel: ACPI: MCFG 0x000000009CFE23A4 00003C (v01 BOCHS  BXPC     00000001 BXPC 00000001)
Mar 20 21:30:06.874113 kernel: ACPI: WAET 0x000000009CFE23E0 000028 (v01 BOCHS  BXPC     00000001 BXPC 00000001)
Mar 20 21:30:06.874120 kernel: ACPI: Reserving FACP table memory at [mem 0x9cfe21e8-0x9cfe22db]
Mar 20 21:30:06.874128 kernel: ACPI: Reserving DSDT table memory at [mem 0x9cfe0040-0x9cfe21e7]
Mar 20 21:30:06.874139 kernel: ACPI: Reserving FACS table memory at [mem 0x9cfe0000-0x9cfe003f]
Mar 20 21:30:06.874148 kernel: ACPI: Reserving APIC table memory at [mem 0x9cfe22dc-0x9cfe236b]
Mar 20 21:30:06.874155 kernel: ACPI: Reserving HPET table memory at [mem 0x9cfe236c-0x9cfe23a3]
Mar 20 21:30:06.874163 kernel: ACPI: Reserving MCFG table memory at [mem 0x9cfe23a4-0x9cfe23df]
Mar 20 21:30:06.874171 kernel: ACPI: Reserving WAET table memory at [mem 0x9cfe23e0-0x9cfe2407]
Mar 20 21:30:06.874178 kernel: No NUMA configuration found
Mar 20 21:30:06.874185 kernel: Faking a node at [mem 0x0000000000000000-0x000000009cfdbfff]
Mar 20 21:30:06.874193 kernel: NODE_DATA(0) allocated [mem 0x9cfd6000-0x9cfdbfff]
Mar 20 21:30:06.874203 kernel: Zone ranges:
Mar 20 21:30:06.874210 kernel:   DMA      [mem 0x0000000000001000-0x0000000000ffffff]
Mar 20 21:30:06.874218 kernel:   DMA32    [mem 0x0000000001000000-0x000000009cfdbfff]
Mar 20 21:30:06.874225 kernel:   Normal   empty
Mar 20 21:30:06.874233 kernel: Movable zone start for each node
Mar 20 21:30:06.874240 kernel: Early memory node ranges
Mar 20 21:30:06.874247 kernel:   node   0: [mem 0x0000000000001000-0x000000000009efff]
Mar 20 21:30:06.874255 kernel:   node   0: [mem 0x0000000000100000-0x000000009cfdbfff]
Mar 20 21:30:06.874262 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000009cfdbfff]
Mar 20 21:30:06.874270 kernel: On node 0, zone DMA: 1 pages in unavailable ranges
Mar 20 21:30:06.874280 kernel: On node 0, zone DMA: 97 pages in unavailable ranges
Mar 20 21:30:06.874287 kernel: On node 0, zone DMA32: 12324 pages in unavailable ranges
Mar 20 21:30:06.874295 kernel: ACPI: PM-Timer IO Port: 0x608
Mar 20 21:30:06.874302 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1])
Mar 20 21:30:06.874310 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23
Mar 20 21:30:06.874317 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl)
Mar 20 21:30:06.874325 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level)
Mar 20 21:30:06.874332 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level)
Mar 20 21:30:06.874340 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level)
Mar 20 21:30:06.874349 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level)
Mar 20 21:30:06.874357 kernel: ACPI: Using ACPI (MADT) for SMP configuration information
Mar 20 21:30:06.874364 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000
Mar 20 21:30:06.874372 kernel: TSC deadline timer available
Mar 20 21:30:06.874379 kernel: smpboot: Allowing 4 CPUs, 0 hotplug CPUs
Mar 20 21:30:06.874387 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write()
Mar 20 21:30:06.874394 kernel: kvm-guest: KVM setup pv remote TLB flush
Mar 20 21:30:06.874402 kernel: kvm-guest: setup PV sched yield
Mar 20 21:30:06.874409 kernel: [mem 0xc0000000-0xfed1bfff] available for PCI devices
Mar 20 21:30:06.874419 kernel: Booting paravirtualized kernel on KVM
Mar 20 21:30:06.874427 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns
Mar 20 21:30:06.874434 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:4 nr_cpu_ids:4 nr_node_ids:1
Mar 20 21:30:06.874442 kernel: percpu: Embedded 58 pages/cpu s197032 r8192 d32344 u524288
Mar 20 21:30:06.874450 kernel: pcpu-alloc: s197032 r8192 d32344 u524288 alloc=1*2097152
Mar 20 21:30:06.874457 kernel: pcpu-alloc: [0] 0 1 2 3 
Mar 20 21:30:06.874464 kernel: kvm-guest: PV spinlocks enabled
Mar 20 21:30:06.874472 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear)
Mar 20 21:30:06.874480 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=619bfa043b53ac975036e415994a80721794ae8277072d0a93c174b4f7768019
Mar 20 21:30:06.874491 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space.
Mar 20 21:30:06.874498 kernel: random: crng init done
Mar 20 21:30:06.874506 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear)
Mar 20 21:30:06.874514 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear)
Mar 20 21:30:06.874521 kernel: Fallback order for Node 0: 0 
Mar 20 21:30:06.874529 kernel: Built 1 zonelists, mobility grouping on.  Total pages: 632732
Mar 20 21:30:06.874536 kernel: Policy zone: DMA32
Mar 20 21:30:06.874543 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off
Mar 20 21:30:06.874554 kernel: Memory: 2430496K/2571752K available (14336K kernel code, 2304K rwdata, 25060K rodata, 43592K init, 1472K bss, 140996K reserved, 0K cma-reserved)
Mar 20 21:30:06.874568 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1
Mar 20 21:30:06.874576 kernel: ftrace: allocating 37985 entries in 149 pages
Mar 20 21:30:06.874584 kernel: ftrace: allocated 149 pages with 4 groups
Mar 20 21:30:06.874591 kernel: Dynamic Preempt: voluntary
Mar 20 21:30:06.874599 kernel: rcu: Preemptible hierarchical RCU implementation.
Mar 20 21:30:06.874607 kernel: rcu:         RCU event tracing is enabled.
Mar 20 21:30:06.874625 kernel: rcu:         RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4.
Mar 20 21:30:06.874633 kernel:         Trampoline variant of Tasks RCU enabled.
Mar 20 21:30:06.874644 kernel:         Rude variant of Tasks RCU enabled.
Mar 20 21:30:06.874652 kernel:         Tracing variant of Tasks RCU enabled.
Mar 20 21:30:06.874659 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies.
Mar 20 21:30:06.874667 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4
Mar 20 21:30:06.874674 kernel: NR_IRQS: 33024, nr_irqs: 456, preallocated irqs: 16
Mar 20 21:30:06.874684 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention.
Mar 20 21:30:06.874694 kernel: Console: colour VGA+ 80x25
Mar 20 21:30:06.874705 kernel: printk: console [ttyS0] enabled
Mar 20 21:30:06.874715 kernel: ACPI: Core revision 20230628
Mar 20 21:30:06.874725 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns
Mar 20 21:30:06.874738 kernel: APIC: Switch to symmetric I/O mode setup
Mar 20 21:30:06.874748 kernel: x2apic enabled
Mar 20 21:30:06.874758 kernel: APIC: Switched APIC routing to: physical x2apic
Mar 20 21:30:06.874768 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask()
Mar 20 21:30:06.874779 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself()
Mar 20 21:30:06.874788 kernel: kvm-guest: setup PV IPIs
Mar 20 21:30:06.874805 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1
Mar 20 21:30:06.874813 kernel: tsc: Marking TSC unstable due to TSCs unsynchronized
Mar 20 21:30:06.874821 kernel: Calibrating delay loop (skipped) preset value.. 5589.50 BogoMIPS (lpj=2794750)
Mar 20 21:30:06.874829 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated
Mar 20 21:30:06.874837 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127
Mar 20 21:30:06.874847 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0
Mar 20 21:30:06.874855 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization
Mar 20 21:30:06.874862 kernel: Spectre V2 : Mitigation: Retpolines
Mar 20 21:30:06.874871 kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch
Mar 20 21:30:06.874878 kernel: Spectre V2 : Spectre v2 / SpectreRSB : Filling RSB on VMEXIT
Mar 20 21:30:06.874889 kernel: Spectre V2 : Enabling Speculation Barrier for firmware calls
Mar 20 21:30:06.874896 kernel: RETBleed: Mitigation: untrained return thunk
Mar 20 21:30:06.874904 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier
Mar 20 21:30:06.874913 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl
Mar 20 21:30:06.874920 kernel: Speculative Return Stack Overflow: IBPB-extending microcode not applied!
Mar 20 21:30:06.874929 kernel: Speculative Return Stack Overflow: WARNING: See https://kernel.org/doc/html/latest/admin-guide/hw-vuln/srso.html for mitigation options.
Mar 20 21:30:06.874937 kernel: Speculative Return Stack Overflow: Vulnerable: Safe RET, no microcode
Mar 20 21:30:06.874945 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers'
Mar 20 21:30:06.874955 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers'
Mar 20 21:30:06.874963 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers'
Mar 20 21:30:06.874971 kernel: x86/fpu: xstate_offset[2]:  576, xstate_sizes[2]:  256
Mar 20 21:30:06.874979 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format.
Mar 20 21:30:06.874986 kernel: Freeing SMP alternatives memory: 32K
Mar 20 21:30:06.874994 kernel: pid_max: default: 32768 minimum: 301
Mar 20 21:30:06.875002 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity
Mar 20 21:30:06.875010 kernel: landlock: Up and running.
Mar 20 21:30:06.875018 kernel: SELinux:  Initializing.
Mar 20 21:30:06.875028 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear)
Mar 20 21:30:06.875036 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear)
Mar 20 21:30:06.875044 kernel: smpboot: CPU0: AMD EPYC 7402P 24-Core Processor (family: 0x17, model: 0x31, stepping: 0x0)
Mar 20 21:30:06.875052 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4.
Mar 20 21:30:06.875060 kernel: RCU Tasks Rude: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4.
Mar 20 21:30:06.875067 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4.
Mar 20 21:30:06.875075 kernel: Performance Events: Fam17h+ core perfctr, AMD PMU driver.
Mar 20 21:30:06.875083 kernel: ... version:                0
Mar 20 21:30:06.875091 kernel: ... bit width:              48
Mar 20 21:30:06.875101 kernel: ... generic registers:      6
Mar 20 21:30:06.875109 kernel: ... value mask:             0000ffffffffffff
Mar 20 21:30:06.875116 kernel: ... max period:             00007fffffffffff
Mar 20 21:30:06.875124 kernel: ... fixed-purpose events:   0
Mar 20 21:30:06.875132 kernel: ... event mask:             000000000000003f
Mar 20 21:30:06.875140 kernel: signal: max sigframe size: 1776
Mar 20 21:30:06.875149 kernel: rcu: Hierarchical SRCU implementation.
Mar 20 21:30:06.875158 kernel: rcu:         Max phase no-delay instances is 400.
Mar 20 21:30:06.875167 kernel: smp: Bringing up secondary CPUs ...
Mar 20 21:30:06.875179 kernel: smpboot: x86: Booting SMP configuration:
Mar 20 21:30:06.875186 kernel: .... node  #0, CPUs:      #1 #2 #3
Mar 20 21:30:06.875194 kernel: smp: Brought up 1 node, 4 CPUs
Mar 20 21:30:06.875202 kernel: smpboot: Max logical packages: 1
Mar 20 21:30:06.875210 kernel: smpboot: Total of 4 processors activated (22358.00 BogoMIPS)
Mar 20 21:30:06.875217 kernel: devtmpfs: initialized
Mar 20 21:30:06.875225 kernel: x86/mm: Memory block size: 128MB
Mar 20 21:30:06.875233 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns
Mar 20 21:30:06.875241 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear)
Mar 20 21:30:06.875252 kernel: pinctrl core: initialized pinctrl subsystem
Mar 20 21:30:06.875259 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family
Mar 20 21:30:06.875267 kernel: audit: initializing netlink subsys (disabled)
Mar 20 21:30:06.875275 kernel: audit: type=2000 audit(1742506205.739:1): state=initialized audit_enabled=0 res=1
Mar 20 21:30:06.875283 kernel: thermal_sys: Registered thermal governor 'step_wise'
Mar 20 21:30:06.875291 kernel: thermal_sys: Registered thermal governor 'user_space'
Mar 20 21:30:06.875298 kernel: cpuidle: using governor menu
Mar 20 21:30:06.875306 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5
Mar 20 21:30:06.875314 kernel: dca service started, version 1.12.1
Mar 20 21:30:06.875324 kernel: PCI: MMCONFIG for domain 0000 [bus 00-ff] at [mem 0xb0000000-0xbfffffff] (base 0xb0000000)
Mar 20 21:30:06.875332 kernel: PCI: MMCONFIG at [mem 0xb0000000-0xbfffffff] reserved as E820 entry
Mar 20 21:30:06.875340 kernel: PCI: Using configuration type 1 for base access
Mar 20 21:30:06.875347 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible.
Mar 20 21:30:06.875355 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages
Mar 20 21:30:06.875363 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page
Mar 20 21:30:06.875371 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages
Mar 20 21:30:06.875379 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page
Mar 20 21:30:06.875387 kernel: ACPI: Added _OSI(Module Device)
Mar 20 21:30:06.875397 kernel: ACPI: Added _OSI(Processor Device)
Mar 20 21:30:06.875404 kernel: ACPI: Added _OSI(3.0 _SCP Extensions)
Mar 20 21:30:06.875412 kernel: ACPI: Added _OSI(Processor Aggregator Device)
Mar 20 21:30:06.875420 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded
Mar 20 21:30:06.875428 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC
Mar 20 21:30:06.875435 kernel: ACPI: Interpreter enabled
Mar 20 21:30:06.875443 kernel: ACPI: PM: (supports S0 S3 S5)
Mar 20 21:30:06.875451 kernel: ACPI: Using IOAPIC for interrupt routing
Mar 20 21:30:06.875459 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug
Mar 20 21:30:06.875469 kernel: PCI: Using E820 reservations for host bridge windows
Mar 20 21:30:06.875476 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F
Mar 20 21:30:06.875484 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff])
Mar 20 21:30:06.875698 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3]
Mar 20 21:30:06.875833 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR]
Mar 20 21:30:06.875972 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability]
Mar 20 21:30:06.875984 kernel: PCI host bridge to bus 0000:00
Mar 20 21:30:06.876112 kernel: pci_bus 0000:00: root bus resource [io  0x0000-0x0cf7 window]
Mar 20 21:30:06.876233 kernel: pci_bus 0000:00: root bus resource [io  0x0d00-0xffff window]
Mar 20 21:30:06.876349 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window]
Mar 20 21:30:06.876464 kernel: pci_bus 0000:00: root bus resource [mem 0x9d000000-0xafffffff window]
Mar 20 21:30:06.876604 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window]
Mar 20 21:30:06.876745 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x8ffffffff window]
Mar 20 21:30:06.876861 kernel: pci_bus 0000:00: root bus resource [bus 00-ff]
Mar 20 21:30:06.877015 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000
Mar 20 21:30:06.877163 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000
Mar 20 21:30:06.877295 kernel: pci 0000:00:01.0: reg 0x10: [mem 0xfd000000-0xfdffffff pref]
Mar 20 21:30:06.877431 kernel: pci 0000:00:01.0: reg 0x18: [mem 0xfebd0000-0xfebd0fff]
Mar 20 21:30:06.877569 kernel: pci 0000:00:01.0: reg 0x30: [mem 0xfebc0000-0xfebcffff pref]
Mar 20 21:30:06.877752 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff]
Mar 20 21:30:06.877891 kernel: pci 0000:00:02.0: [1af4:1005] type 00 class 0x00ff00
Mar 20 21:30:06.878037 kernel: pci 0000:00:02.0: reg 0x10: [io  0xc0c0-0xc0df]
Mar 20 21:30:06.878170 kernel: pci 0000:00:02.0: reg 0x14: [mem 0xfebd1000-0xfebd1fff]
Mar 20 21:30:06.878311 kernel: pci 0000:00:02.0: reg 0x20: [mem 0xfe000000-0xfe003fff 64bit pref]
Mar 20 21:30:06.878449 kernel: pci 0000:00:03.0: [1af4:1001] type 00 class 0x010000
Mar 20 21:30:06.878603 kernel: pci 0000:00:03.0: reg 0x10: [io  0xc000-0xc07f]
Mar 20 21:30:06.878749 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xfebd2000-0xfebd2fff]
Mar 20 21:30:06.878878 kernel: pci 0000:00:03.0: reg 0x20: [mem 0xfe004000-0xfe007fff 64bit pref]
Mar 20 21:30:06.879022 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000
Mar 20 21:30:06.879165 kernel: pci 0000:00:04.0: reg 0x10: [io  0xc0e0-0xc0ff]
Mar 20 21:30:06.879293 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xfebd3000-0xfebd3fff]
Mar 20 21:30:06.879418 kernel: pci 0000:00:04.0: reg 0x20: [mem 0xfe008000-0xfe00bfff 64bit pref]
Mar 20 21:30:06.879552 kernel: pci 0000:00:04.0: reg 0x30: [mem 0xfeb80000-0xfebbffff pref]
Mar 20 21:30:06.879733 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100
Mar 20 21:30:06.879866 kernel: pci 0000:00:1f.0: quirk: [io  0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO
Mar 20 21:30:06.879999 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601
Mar 20 21:30:06.880124 kernel: pci 0000:00:1f.2: reg 0x20: [io  0xc100-0xc11f]
Mar 20 21:30:06.880248 kernel: pci 0000:00:1f.2: reg 0x24: [mem 0xfebd4000-0xfebd4fff]
Mar 20 21:30:06.880380 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500
Mar 20 21:30:06.880505 kernel: pci 0000:00:1f.3: reg 0x20: [io  0x0700-0x073f]
Mar 20 21:30:06.880516 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10
Mar 20 21:30:06.880528 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10
Mar 20 21:30:06.880536 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11
Mar 20 21:30:06.880544 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11
Mar 20 21:30:06.880552 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10
Mar 20 21:30:06.880570 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10
Mar 20 21:30:06.880579 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11
Mar 20 21:30:06.880587 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11
Mar 20 21:30:06.880594 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16
Mar 20 21:30:06.880602 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17
Mar 20 21:30:06.880625 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18
Mar 20 21:30:06.880633 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19
Mar 20 21:30:06.880641 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20
Mar 20 21:30:06.880649 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21
Mar 20 21:30:06.880657 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22
Mar 20 21:30:06.880664 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23
Mar 20 21:30:06.880672 kernel: iommu: Default domain type: Translated
Mar 20 21:30:06.880680 kernel: iommu: DMA domain TLB invalidation policy: lazy mode
Mar 20 21:30:06.880688 kernel: PCI: Using ACPI for IRQ routing
Mar 20 21:30:06.880699 kernel: PCI: pci_cache_line_size set to 64 bytes
Mar 20 21:30:06.880706 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff]
Mar 20 21:30:06.880714 kernel: e820: reserve RAM buffer [mem 0x9cfdc000-0x9fffffff]
Mar 20 21:30:06.880842 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device
Mar 20 21:30:06.880967 kernel: pci 0000:00:01.0: vgaarb: bridge control possible
Mar 20 21:30:06.881108 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none
Mar 20 21:30:06.881120 kernel: vgaarb: loaded
Mar 20 21:30:06.881128 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0
Mar 20 21:30:06.881140 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter
Mar 20 21:30:06.881148 kernel: clocksource: Switched to clocksource kvm-clock
Mar 20 21:30:06.881156 kernel: VFS: Disk quotas dquot_6.6.0
Mar 20 21:30:06.881166 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes)
Mar 20 21:30:06.881174 kernel: pnp: PnP ACPI init
Mar 20 21:30:06.881315 kernel: system 00:05: [mem 0xb0000000-0xbfffffff window] has been reserved
Mar 20 21:30:06.881327 kernel: pnp: PnP ACPI: found 6 devices
Mar 20 21:30:06.881336 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns
Mar 20 21:30:06.881347 kernel: NET: Registered PF_INET protocol family
Mar 20 21:30:06.881355 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear)
Mar 20 21:30:06.881366 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear)
Mar 20 21:30:06.881377 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear)
Mar 20 21:30:06.881388 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear)
Mar 20 21:30:06.881399 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear)
Mar 20 21:30:06.881410 kernel: TCP: Hash tables configured (established 32768 bind 32768)
Mar 20 21:30:06.881420 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear)
Mar 20 21:30:06.881428 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear)
Mar 20 21:30:06.881439 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family
Mar 20 21:30:06.881447 kernel: NET: Registered PF_XDP protocol family
Mar 20 21:30:06.881578 kernel: pci_bus 0000:00: resource 4 [io  0x0000-0x0cf7 window]
Mar 20 21:30:06.881708 kernel: pci_bus 0000:00: resource 5 [io  0x0d00-0xffff window]
Mar 20 21:30:06.881825 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window]
Mar 20 21:30:06.881944 kernel: pci_bus 0000:00: resource 7 [mem 0x9d000000-0xafffffff window]
Mar 20 21:30:06.882073 kernel: pci_bus 0000:00: resource 8 [mem 0xc0000000-0xfebfffff window]
Mar 20 21:30:06.882190 kernel: pci_bus 0000:00: resource 9 [mem 0x100000000-0x8ffffffff window]
Mar 20 21:30:06.882205 kernel: PCI: CLS 0 bytes, default 64
Mar 20 21:30:06.882213 kernel: Initialise system trusted keyrings
Mar 20 21:30:06.882223 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0
Mar 20 21:30:06.882234 kernel: Key type asymmetric registered
Mar 20 21:30:06.882245 kernel: Asymmetric key parser 'x509' registered
Mar 20 21:30:06.882256 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251)
Mar 20 21:30:06.882265 kernel: io scheduler mq-deadline registered
Mar 20 21:30:06.882275 kernel: io scheduler kyber registered
Mar 20 21:30:06.882286 kernel: io scheduler bfq registered
Mar 20 21:30:06.882297 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00
Mar 20 21:30:06.882309 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22
Mar 20 21:30:06.882317 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23
Mar 20 21:30:06.882325 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20
Mar 20 21:30:06.882333 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled
Mar 20 21:30:06.882341 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A
Mar 20 21:30:06.882349 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12
Mar 20 21:30:06.882357 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1
Mar 20 21:30:06.882365 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12
Mar 20 21:30:06.882500 kernel: rtc_cmos 00:04: RTC can wake from S4
Mar 20 21:30:06.882516 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0
Mar 20 21:30:06.882692 kernel: rtc_cmos 00:04: registered as rtc0
Mar 20 21:30:06.882812 kernel: rtc_cmos 00:04: setting system clock to 2025-03-20T21:30:06 UTC (1742506206)
Mar 20 21:30:06.882927 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram, hpet irqs
Mar 20 21:30:06.882938 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled
Mar 20 21:30:06.882945 kernel: NET: Registered PF_INET6 protocol family
Mar 20 21:30:06.882953 kernel: Segment Routing with IPv6
Mar 20 21:30:06.882961 kernel: In-situ OAM (IOAM) with IPv6
Mar 20 21:30:06.882974 kernel: NET: Registered PF_PACKET protocol family
Mar 20 21:30:06.882982 kernel: Key type dns_resolver registered
Mar 20 21:30:06.882989 kernel: IPI shorthand broadcast: enabled
Mar 20 21:30:06.882997 kernel: sched_clock: Marking stable (543002525, 104648436)->(697670873, -50019912)
Mar 20 21:30:06.883005 kernel: registered taskstats version 1
Mar 20 21:30:06.883013 kernel: Loading compiled-in X.509 certificates
Mar 20 21:30:06.883021 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.83-flatcar: 9e7923b67df1c6f0613bc4380f7ea8de9ce851ac'
Mar 20 21:30:06.883029 kernel: Key type .fscrypt registered
Mar 20 21:30:06.883037 kernel: Key type fscrypt-provisioning registered
Mar 20 21:30:06.883048 kernel: ima: No TPM chip found, activating TPM-bypass!
Mar 20 21:30:06.883056 kernel: ima: Allocated hash algorithm: sha1
Mar 20 21:30:06.883064 kernel: ima: No architecture policies found
Mar 20 21:30:06.883071 kernel: clk: Disabling unused clocks
Mar 20 21:30:06.883079 kernel: Freeing unused kernel image (initmem) memory: 43592K
Mar 20 21:30:06.883087 kernel: Write protecting the kernel read-only data: 40960k
Mar 20 21:30:06.883095 kernel: Freeing unused kernel image (rodata/data gap) memory: 1564K
Mar 20 21:30:06.883103 kernel: Run /init as init process
Mar 20 21:30:06.883113 kernel:   with arguments:
Mar 20 21:30:06.883121 kernel:     /init
Mar 20 21:30:06.883129 kernel:   with environment:
Mar 20 21:30:06.883137 kernel:     HOME=/
Mar 20 21:30:06.883144 kernel:     TERM=linux
Mar 20 21:30:06.883152 kernel:     BOOT_IMAGE=/flatcar/vmlinuz-a
Mar 20 21:30:06.883161 systemd[1]: Successfully made /usr/ read-only.
Mar 20 21:30:06.883172 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE)
Mar 20 21:30:06.883184 systemd[1]: Detected virtualization kvm.
Mar 20 21:30:06.883193 systemd[1]: Detected architecture x86-64.
Mar 20 21:30:06.883201 systemd[1]: Running in initrd.
Mar 20 21:30:06.883209 systemd[1]: No hostname configured, using default hostname.
Mar 20 21:30:06.883219 systemd[1]: Hostname set to <localhost>.
Mar 20 21:30:06.883230 systemd[1]: Initializing machine ID from VM UUID.
Mar 20 21:30:06.883242 systemd[1]: Queued start job for default target initrd.target.
Mar 20 21:30:06.883254 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch.
Mar 20 21:30:06.883270 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch.
Mar 20 21:30:06.883291 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM...
Mar 20 21:30:06.883302 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM...
Mar 20 21:30:06.883311 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT...
Mar 20 21:30:06.883321 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A...
Mar 20 21:30:06.883333 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132...
Mar 20 21:30:06.883342 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr...
Mar 20 21:30:06.883355 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre).
Mar 20 21:30:06.883367 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes.
Mar 20 21:30:06.883377 systemd[1]: Reached target paths.target - Path Units.
Mar 20 21:30:06.883389 systemd[1]: Reached target slices.target - Slice Units.
Mar 20 21:30:06.883401 systemd[1]: Reached target swap.target - Swaps.
Mar 20 21:30:06.883412 systemd[1]: Reached target timers.target - Timer Units.
Mar 20 21:30:06.883426 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket.
Mar 20 21:30:06.883437 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket.
Mar 20 21:30:06.883448 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log).
Mar 20 21:30:06.883457 systemd[1]: Listening on systemd-journald.socket - Journal Sockets.
Mar 20 21:30:06.883466 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket.
Mar 20 21:30:06.883475 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket.
Mar 20 21:30:06.883484 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket.
Mar 20 21:30:06.883493 systemd[1]: Reached target sockets.target - Socket Units.
Mar 20 21:30:06.883501 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup...
Mar 20 21:30:06.883513 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes...
Mar 20 21:30:06.883521 systemd[1]: Finished network-cleanup.service - Network Cleanup.
Mar 20 21:30:06.883530 systemd[1]: Starting systemd-fsck-usr.service...
Mar 20 21:30:06.883539 systemd[1]: Starting systemd-journald.service - Journal Service...
Mar 20 21:30:06.883547 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules...
Mar 20 21:30:06.883556 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup...
Mar 20 21:30:06.883574 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup.
Mar 20 21:30:06.883583 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes.
Mar 20 21:30:06.883595 systemd[1]: Finished systemd-fsck-usr.service.
Mar 20 21:30:06.883604 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully...
Mar 20 21:30:06.883653 systemd-journald[193]: Collecting audit messages is disabled.
Mar 20 21:30:06.883674 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully.
Mar 20 21:30:06.883684 systemd-journald[193]: Journal started
Mar 20 21:30:06.883706 systemd-journald[193]: Runtime Journal (/run/log/journal/46d2dd9d9ab54cf4aaf8fd0b2963e989) is 6M, max 48.3M, 42.3M free.
Mar 20 21:30:06.868782 systemd-modules-load[195]: Inserted module 'overlay'
Mar 20 21:30:06.907923 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this.
Mar 20 21:30:06.907951 kernel: Bridge firewalling registered
Mar 20 21:30:06.907962 systemd[1]: Started systemd-journald.service - Journal Service.
Mar 20 21:30:06.895379 systemd-modules-load[195]: Inserted module 'br_netfilter'
Mar 20 21:30:06.908197 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules.
Mar 20 21:30:06.911777 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables...
Mar 20 21:30:06.912937 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev...
Mar 20 21:30:06.916485 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories...
Mar 20 21:30:06.930258 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup.
Mar 20 21:30:06.932476 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters...
Mar 20 21:30:06.940723 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables.
Mar 20 21:30:06.944277 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories.
Mar 20 21:30:06.945193 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev.
Mar 20 21:30:06.947821 systemd[1]: Starting systemd-resolved.service - Network Name Resolution...
Mar 20 21:30:06.964804 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters.
Mar 20 21:30:06.967011 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook...
Mar 20 21:30:06.997789 systemd-resolved[224]: Positive Trust Anchors:
Mar 20 21:30:06.997804 systemd-resolved[224]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d
Mar 20 21:30:06.997834 systemd-resolved[224]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test
Mar 20 21:30:07.000232 systemd-resolved[224]: Defaulting to hostname 'linux'.
Mar 20 21:30:07.010374 dracut-cmdline[231]: dracut-dracut-053
Mar 20 21:30:07.010374 dracut-cmdline[231]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=619bfa043b53ac975036e415994a80721794ae8277072d0a93c174b4f7768019
Mar 20 21:30:07.001283 systemd[1]: Started systemd-resolved.service - Network Name Resolution.
Mar 20 21:30:07.008405 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups.
Mar 20 21:30:07.088645 kernel: SCSI subsystem initialized
Mar 20 21:30:07.097639 kernel: Loading iSCSI transport class v2.0-870.
Mar 20 21:30:07.108649 kernel: iscsi: registered transport (tcp)
Mar 20 21:30:07.130635 kernel: iscsi: registered transport (qla4xxx)
Mar 20 21:30:07.130660 kernel: QLogic iSCSI HBA Driver
Mar 20 21:30:07.180369 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook.
Mar 20 21:30:07.182176 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook...
Mar 20 21:30:07.229531 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log.
Mar 20 21:30:07.229571 kernel: device-mapper: uevent: version 1.0.3
Mar 20 21:30:07.229591 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com
Mar 20 21:30:07.271638 kernel: raid6: avx2x4   gen() 27682 MB/s
Mar 20 21:30:07.288638 kernel: raid6: avx2x2   gen() 27748 MB/s
Mar 20 21:30:07.305776 kernel: raid6: avx2x1   gen() 23248 MB/s
Mar 20 21:30:07.305795 kernel: raid6: using algorithm avx2x2 gen() 27748 MB/s
Mar 20 21:30:07.323793 kernel: raid6: .... xor() 18081 MB/s, rmw enabled
Mar 20 21:30:07.323820 kernel: raid6: using avx2x2 recovery algorithm
Mar 20 21:30:07.345637 kernel: xor: automatically using best checksumming function   avx       
Mar 20 21:30:07.498646 kernel: Btrfs loaded, zoned=no, fsverity=no
Mar 20 21:30:07.512338 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook.
Mar 20 21:30:07.515184 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files...
Mar 20 21:30:07.543157 systemd-udevd[414]: Using default interface naming scheme 'v255'.
Mar 20 21:30:07.548709 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files.
Mar 20 21:30:07.553645 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook...
Mar 20 21:30:07.576138 dracut-pre-trigger[422]: rd.md=0: removing MD RAID activation
Mar 20 21:30:07.613194 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook.
Mar 20 21:30:07.616728 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices...
Mar 20 21:30:07.690393 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices.
Mar 20 21:30:07.694742 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook...
Mar 20 21:30:07.715465 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook.
Mar 20 21:30:07.719163 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems.
Mar 20 21:30:07.722138 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes.
Mar 20 21:30:07.723577 systemd[1]: Reached target remote-fs.target - Remote File Systems.
Mar 20 21:30:07.730669 kernel: virtio_blk virtio1: 4/0/0 default/read/poll queues
Mar 20 21:30:07.738806 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB)
Mar 20 21:30:07.738973 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk.
Mar 20 21:30:07.738988 kernel: GPT:9289727 != 19775487
Mar 20 21:30:07.739002 kernel: GPT:Alternate GPT header not at the end of the disk.
Mar 20 21:30:07.739016 kernel: GPT:9289727 != 19775487
Mar 20 21:30:07.739034 kernel: GPT: Use GNU Parted to correct GPT errors.
Mar 20 21:30:07.739048 kernel:  vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9
Mar 20 21:30:07.739062 kernel: cryptd: max_cpu_qlen set to 1000
Mar 20 21:30:07.731389 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook...
Mar 20 21:30:07.748639 kernel: libata version 3.00 loaded.
Mar 20 21:30:07.758677 kernel: ahci 0000:00:1f.2: version 3.0
Mar 20 21:30:07.792687 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16
Mar 20 21:30:07.792713 kernel: ahci 0000:00:1f.2: AHCI 0001.0000 32 slots 6 ports 1.5 Gbps 0x3f impl SATA mode
Mar 20 21:30:07.792873 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only 
Mar 20 21:30:07.793013 kernel: scsi host0: ahci
Mar 20 21:30:07.793164 kernel: scsi host1: ahci
Mar 20 21:30:07.793523 kernel: AVX2 version of gcm_enc/dec engaged.
Mar 20 21:30:07.793545 kernel: AES CTR mode by8 optimization enabled
Mar 20 21:30:07.793556 kernel: scsi host2: ahci
Mar 20 21:30:07.793761 kernel: scsi host3: ahci
Mar 20 21:30:07.793908 kernel: scsi host4: ahci
Mar 20 21:30:07.794054 kernel: BTRFS: device label OEM devid 1 transid 15 /dev/vda6 scanned by (udev-worker) (479)
Mar 20 21:30:07.794066 kernel: BTRFS: device fsid 48a514e8-9ecc-46c2-935b-caca347f921e devid 1 transid 39 /dev/vda3 scanned by (udev-worker) (460)
Mar 20 21:30:07.794077 kernel: scsi host5: ahci
Mar 20 21:30:07.794223 kernel: ata1: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4100 irq 34
Mar 20 21:30:07.794235 kernel: ata2: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4180 irq 34
Mar 20 21:30:07.794249 kernel: ata3: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4200 irq 34
Mar 20 21:30:07.794260 kernel: ata4: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4280 irq 34
Mar 20 21:30:07.794271 kernel: ata5: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4300 irq 34
Mar 20 21:30:07.794281 kernel: ata6: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4380 irq 34
Mar 20 21:30:07.764809 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook.
Mar 20 21:30:07.775496 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully.
Mar 20 21:30:07.775632 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters.
Mar 20 21:30:07.777173 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters...
Mar 20 21:30:07.778436 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully.
Mar 20 21:30:07.778570 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup.
Mar 20 21:30:07.779927 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup...
Mar 20 21:30:07.781974 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup...
Mar 20 21:30:07.811637 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM.
Mar 20 21:30:07.823769 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT.
Mar 20 21:30:07.853083 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup.
Mar 20 21:30:07.863966 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM.
Mar 20 21:30:07.879178 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A.
Mar 20 21:30:07.879656 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132.
Mar 20 21:30:07.880960 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary...
Mar 20 21:30:07.898887 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters...
Mar 20 21:30:07.909325 disk-uuid[558]: Primary Header is updated.
Mar 20 21:30:07.909325 disk-uuid[558]: Secondary Entries is updated.
Mar 20 21:30:07.909325 disk-uuid[558]: Secondary Header is updated.
Mar 20 21:30:07.913644 kernel:  vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9
Mar 20 21:30:07.918681 kernel:  vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9
Mar 20 21:30:07.928135 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters.
Mar 20 21:30:08.098643 kernel: ata2: SATA link down (SStatus 0 SControl 300)
Mar 20 21:30:08.098703 kernel: ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300)
Mar 20 21:30:08.099955 kernel: ata3.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100
Mar 20 21:30:08.101459 kernel: ata3.00: applying bridge limits
Mar 20 21:30:08.101471 kernel: ata4: SATA link down (SStatus 0 SControl 300)
Mar 20 21:30:08.101482 kernel: ata1: SATA link down (SStatus 0 SControl 300)
Mar 20 21:30:08.101492 kernel: ata3.00: configured for UDMA/100
Mar 20 21:30:08.102638 kernel: ata6: SATA link down (SStatus 0 SControl 300)
Mar 20 21:30:08.103654 kernel: scsi 2:0:0:0: CD-ROM            QEMU     QEMU DVD-ROM     2.5+ PQ: 0 ANSI: 5
Mar 20 21:30:08.107647 kernel: ata5: SATA link down (SStatus 0 SControl 300)
Mar 20 21:30:08.145650 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray
Mar 20 21:30:08.165333 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20
Mar 20 21:30:08.165351 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0
Mar 20 21:30:08.921603 disk-uuid[559]: The operation has completed successfully.
Mar 20 21:30:08.923090 kernel:  vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9
Mar 20 21:30:08.953724 systemd[1]: disk-uuid.service: Deactivated successfully.
Mar 20 21:30:08.953843 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary.
Mar 20 21:30:08.985682 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr...
Mar 20 21:30:09.007818 sh[594]: Success
Mar 20 21:30:09.019641 kernel: device-mapper: verity: sha256 using implementation "sha256-ni"
Mar 20 21:30:09.056558 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr.
Mar 20 21:30:09.060579 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr...
Mar 20 21:30:09.076978 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr.
Mar 20 21:30:09.084640 kernel: BTRFS info (device dm-0): first mount of filesystem 48a514e8-9ecc-46c2-935b-caca347f921e
Mar 20 21:30:09.084671 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm
Mar 20 21:30:09.084683 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead
Mar 20 21:30:09.086982 kernel: BTRFS info (device dm-0): disabling log replay at mount time
Mar 20 21:30:09.087001 kernel: BTRFS info (device dm-0): using free space tree
Mar 20 21:30:09.090829 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr.
Mar 20 21:30:09.091733 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met.
Mar 20 21:30:09.092648 systemd[1]: Starting ignition-setup.service - Ignition (setup)...
Mar 20 21:30:09.093712 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline...
Mar 20 21:30:09.126045 kernel: BTRFS info (device vda6): first mount of filesystem c415ef49-5595-4a0b-ba48-8f3e642f303e
Mar 20 21:30:09.126075 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm
Mar 20 21:30:09.126090 kernel: BTRFS info (device vda6): using free space tree
Mar 20 21:30:09.130645 kernel: BTRFS info (device vda6): auto enabling async discard
Mar 20 21:30:09.134639 kernel: BTRFS info (device vda6): last unmount of filesystem c415ef49-5595-4a0b-ba48-8f3e642f303e
Mar 20 21:30:09.140247 systemd[1]: Finished ignition-setup.service - Ignition (setup).
Mar 20 21:30:09.143390 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)...
Mar 20 21:30:09.210774 ignition[693]: Ignition 2.20.0
Mar 20 21:30:09.210787 ignition[693]: Stage: fetch-offline
Mar 20 21:30:09.210820 ignition[693]: no configs at "/usr/lib/ignition/base.d"
Mar 20 21:30:09.210829 ignition[693]: no config dir at "/usr/lib/ignition/base.platform.d/qemu"
Mar 20 21:30:09.213425 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline.
Mar 20 21:30:09.210925 ignition[693]: parsed url from cmdline: ""
Mar 20 21:30:09.217427 systemd[1]: Starting systemd-networkd.service - Network Configuration...
Mar 20 21:30:09.210929 ignition[693]: no config URL provided
Mar 20 21:30:09.210937 ignition[693]: reading system config file "/usr/lib/ignition/user.ign"
Mar 20 21:30:09.210946 ignition[693]: no config at "/usr/lib/ignition/user.ign"
Mar 20 21:30:09.210971 ignition[693]: op(1): [started]  loading QEMU firmware config module
Mar 20 21:30:09.210976 ignition[693]: op(1): executing: "modprobe" "qemu_fw_cfg"
Mar 20 21:30:09.225321 ignition[693]: op(1): [finished] loading QEMU firmware config module
Mar 20 21:30:09.259345 systemd-networkd[780]: lo: Link UP
Mar 20 21:30:09.259354 systemd-networkd[780]: lo: Gained carrier
Mar 20 21:30:09.261303 systemd-networkd[780]: Enumeration completed
Mar 20 21:30:09.261708 systemd-networkd[780]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name.
Mar 20 21:30:09.261713 systemd-networkd[780]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network.
Mar 20 21:30:09.262560 systemd[1]: Started systemd-networkd.service - Network Configuration.
Mar 20 21:30:09.262683 systemd-networkd[780]: eth0: Link UP
Mar 20 21:30:09.262687 systemd-networkd[780]: eth0: Gained carrier
Mar 20 21:30:09.262695 systemd-networkd[780]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name.
Mar 20 21:30:09.265776 systemd[1]: Reached target network.target - Network.
Mar 20 21:30:09.276819 ignition[693]: parsing config with SHA512: 1cea2f2c795a399aa57b6c35201017640e0291a9c1958d0aded08092295056b38b6de36eca5815d88c4e8b6828687499ab11841d1ecd44905f309ba2acdd3d5f
Mar 20 21:30:09.281421 unknown[693]: fetched base config from "system"
Mar 20 21:30:09.281470 unknown[693]: fetched user config from "qemu"
Mar 20 21:30:09.282604 ignition[693]: fetch-offline: fetch-offline passed
Mar 20 21:30:09.281664 systemd-networkd[780]: eth0: DHCPv4 address 10.0.0.132/16, gateway 10.0.0.1 acquired from 10.0.0.1
Mar 20 21:30:09.282748 ignition[693]: Ignition finished successfully
Mar 20 21:30:09.285505 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline).
Mar 20 21:30:09.286593 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json).
Mar 20 21:30:09.287541 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)...
Mar 20 21:30:09.312016 ignition[786]: Ignition 2.20.0
Mar 20 21:30:09.312029 ignition[786]: Stage: kargs
Mar 20 21:30:09.312168 ignition[786]: no configs at "/usr/lib/ignition/base.d"
Mar 20 21:30:09.312179 ignition[786]: no config dir at "/usr/lib/ignition/base.platform.d/qemu"
Mar 20 21:30:09.312980 ignition[786]: kargs: kargs passed
Mar 20 21:30:09.313022 ignition[786]: Ignition finished successfully
Mar 20 21:30:09.319313 systemd[1]: Finished ignition-kargs.service - Ignition (kargs).
Mar 20 21:30:09.321476 systemd[1]: Starting ignition-disks.service - Ignition (disks)...
Mar 20 21:30:09.344331 ignition[795]: Ignition 2.20.0
Mar 20 21:30:09.344342 ignition[795]: Stage: disks
Mar 20 21:30:09.344489 ignition[795]: no configs at "/usr/lib/ignition/base.d"
Mar 20 21:30:09.344509 ignition[795]: no config dir at "/usr/lib/ignition/base.platform.d/qemu"
Mar 20 21:30:09.347348 systemd[1]: Finished ignition-disks.service - Ignition (disks).
Mar 20 21:30:09.345289 ignition[795]: disks: disks passed
Mar 20 21:30:09.349117 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device.
Mar 20 21:30:09.345329 ignition[795]: Ignition finished successfully
Mar 20 21:30:09.351152 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems.
Mar 20 21:30:09.353182 systemd[1]: Reached target local-fs.target - Local File Systems.
Mar 20 21:30:09.355411 systemd[1]: Reached target sysinit.target - System Initialization.
Mar 20 21:30:09.356544 systemd[1]: Reached target basic.target - Basic System.
Mar 20 21:30:09.359381 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT...
Mar 20 21:30:09.385080 systemd-fsck[805]: ROOT: clean, 14/553520 files, 52654/553472 blocks
Mar 20 21:30:09.390748 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT.
Mar 20 21:30:09.394474 systemd[1]: Mounting sysroot.mount - /sysroot...
Mar 20 21:30:09.495634 kernel: EXT4-fs (vda9): mounted filesystem 79cdbe74-6884-4c57-b04d-c9a431509f16 r/w with ordered data mode. Quota mode: none.
Mar 20 21:30:09.495888 systemd[1]: Mounted sysroot.mount - /sysroot.
Mar 20 21:30:09.496830 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System.
Mar 20 21:30:09.499671 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem...
Mar 20 21:30:09.501747 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr...
Mar 20 21:30:09.502799 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met.
Mar 20 21:30:09.502838 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot).
Mar 20 21:30:09.502860 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup.
Mar 20 21:30:09.520280 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr.
Mar 20 21:30:09.524257 kernel: BTRFS: device label OEM devid 1 transid 16 /dev/vda6 scanned by mount (813)
Mar 20 21:30:09.523242 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup...
Mar 20 21:30:09.528187 kernel: BTRFS info (device vda6): first mount of filesystem c415ef49-5595-4a0b-ba48-8f3e642f303e
Mar 20 21:30:09.528213 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm
Mar 20 21:30:09.528224 kernel: BTRFS info (device vda6): using free space tree
Mar 20 21:30:09.530656 kernel: BTRFS info (device vda6): auto enabling async discard
Mar 20 21:30:09.531584 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem.
Mar 20 21:30:09.560970 initrd-setup-root[837]: cut: /sysroot/etc/passwd: No such file or directory
Mar 20 21:30:09.565681 initrd-setup-root[844]: cut: /sysroot/etc/group: No such file or directory
Mar 20 21:30:09.570391 initrd-setup-root[851]: cut: /sysroot/etc/shadow: No such file or directory
Mar 20 21:30:09.574895 initrd-setup-root[858]: cut: /sysroot/etc/gshadow: No such file or directory
Mar 20 21:30:09.656631 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup.
Mar 20 21:30:09.658064 systemd[1]: Starting ignition-mount.service - Ignition (mount)...
Mar 20 21:30:09.660567 systemd[1]: Starting sysroot-boot.service - /sysroot/boot...
Mar 20 21:30:09.679645 kernel: BTRFS info (device vda6): last unmount of filesystem c415ef49-5595-4a0b-ba48-8f3e642f303e
Mar 20 21:30:09.698852 systemd[1]: Finished sysroot-boot.service - /sysroot/boot.
Mar 20 21:30:09.711945 ignition[927]: INFO     : Ignition 2.20.0
Mar 20 21:30:09.711945 ignition[927]: INFO     : Stage: mount
Mar 20 21:30:09.713845 ignition[927]: INFO     : no configs at "/usr/lib/ignition/base.d"
Mar 20 21:30:09.713845 ignition[927]: INFO     : no config dir at "/usr/lib/ignition/base.platform.d/qemu"
Mar 20 21:30:09.713845 ignition[927]: INFO     : mount: mount passed
Mar 20 21:30:09.713845 ignition[927]: INFO     : Ignition finished successfully
Mar 20 21:30:09.718138 systemd[1]: Finished ignition-mount.service - Ignition (mount).
Mar 20 21:30:09.721600 systemd[1]: Starting ignition-files.service - Ignition (files)...
Mar 20 21:30:10.084118 systemd[1]: sysroot-oem.mount: Deactivated successfully.
Mar 20 21:30:10.086594 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem...
Mar 20 21:30:10.107636 kernel: BTRFS: device label OEM devid 1 transid 17 /dev/vda6 scanned by mount (939)
Mar 20 21:30:10.107664 kernel: BTRFS info (device vda6): first mount of filesystem c415ef49-5595-4a0b-ba48-8f3e642f303e
Mar 20 21:30:10.110154 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm
Mar 20 21:30:10.110166 kernel: BTRFS info (device vda6): using free space tree
Mar 20 21:30:10.112649 kernel: BTRFS info (device vda6): auto enabling async discard
Mar 20 21:30:10.114536 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem.
Mar 20 21:30:10.144296 ignition[956]: INFO     : Ignition 2.20.0
Mar 20 21:30:10.144296 ignition[956]: INFO     : Stage: files
Mar 20 21:30:10.146208 ignition[956]: INFO     : no configs at "/usr/lib/ignition/base.d"
Mar 20 21:30:10.146208 ignition[956]: INFO     : no config dir at "/usr/lib/ignition/base.platform.d/qemu"
Mar 20 21:30:10.146208 ignition[956]: DEBUG    : files: compiled without relabeling support, skipping
Mar 20 21:30:10.149734 ignition[956]: INFO     : files: ensureUsers: op(1): [started]  creating or modifying user "core"
Mar 20 21:30:10.149734 ignition[956]: DEBUG    : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core"
Mar 20 21:30:10.152645 ignition[956]: INFO     : files: ensureUsers: op(1): [finished] creating or modifying user "core"
Mar 20 21:30:10.152645 ignition[956]: INFO     : files: ensureUsers: op(2): [started]  adding ssh keys to user "core"
Mar 20 21:30:10.152645 ignition[956]: INFO     : files: ensureUsers: op(2): [finished] adding ssh keys to user "core"
Mar 20 21:30:10.151897 unknown[956]: wrote ssh authorized keys file for user: core
Mar 20 21:30:10.157734 ignition[956]: INFO     : files: createFilesystemsFiles: createFiles: op(3): [started]  writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz"
Mar 20 21:30:10.157734 ignition[956]: INFO     : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-amd64.tar.gz: attempt #1
Mar 20 21:30:10.199973 ignition[956]: INFO     : files: createFilesystemsFiles: createFiles: op(3): GET result: OK
Mar 20 21:30:10.332483 ignition[956]: INFO     : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz"
Mar 20 21:30:10.332483 ignition[956]: INFO     : files: createFilesystemsFiles: createFiles: op(4): [started]  writing file "/sysroot/opt/bin/cilium.tar.gz"
Mar 20 21:30:10.336481 ignition[956]: INFO     : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-amd64.tar.gz: attempt #1
Mar 20 21:30:10.801275 ignition[956]: INFO     : files: createFilesystemsFiles: createFiles: op(4): GET result: OK
Mar 20 21:30:10.905459 ignition[956]: INFO     : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz"
Mar 20 21:30:10.907606 ignition[956]: INFO     : files: createFilesystemsFiles: createFiles: op(5): [started]  writing file "/sysroot/home/core/install.sh"
Mar 20 21:30:10.907606 ignition[956]: INFO     : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh"
Mar 20 21:30:10.907606 ignition[956]: INFO     : files: createFilesystemsFiles: createFiles: op(6): [started]  writing file "/sysroot/home/core/nginx.yaml"
Mar 20 21:30:10.907606 ignition[956]: INFO     : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml"
Mar 20 21:30:10.907606 ignition[956]: INFO     : files: createFilesystemsFiles: createFiles: op(7): [started]  writing file "/sysroot/home/core/nfs-pod.yaml"
Mar 20 21:30:10.907606 ignition[956]: INFO     : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml"
Mar 20 21:30:10.907606 ignition[956]: INFO     : files: createFilesystemsFiles: createFiles: op(8): [started]  writing file "/sysroot/home/core/nfs-pvc.yaml"
Mar 20 21:30:10.907606 ignition[956]: INFO     : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml"
Mar 20 21:30:10.907606 ignition[956]: INFO     : files: createFilesystemsFiles: createFiles: op(9): [started]  writing file "/sysroot/etc/flatcar/update.conf"
Mar 20 21:30:10.907606 ignition[956]: INFO     : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf"
Mar 20 21:30:10.907606 ignition[956]: INFO     : files: createFilesystemsFiles: createFiles: op(a): [started]  writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.0-x86-64.raw"
Mar 20 21:30:10.907606 ignition[956]: INFO     : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.0-x86-64.raw"
Mar 20 21:30:10.907606 ignition[956]: INFO     : files: createFilesystemsFiles: createFiles: op(b): [started]  writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.0-x86-64.raw"
Mar 20 21:30:10.907606 ignition[956]: INFO     : files: createFilesystemsFiles: createFiles: op(b): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.31.0-x86-64.raw: attempt #1
Mar 20 21:30:11.155783 systemd-networkd[780]: eth0: Gained IPv6LL
Mar 20 21:30:11.341796 ignition[956]: INFO     : files: createFilesystemsFiles: createFiles: op(b): GET result: OK
Mar 20 21:30:11.692591 ignition[956]: INFO     : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.0-x86-64.raw"
Mar 20 21:30:11.692591 ignition[956]: INFO     : files: op(c): [started]  processing unit "prepare-helm.service"
Mar 20 21:30:11.696220 ignition[956]: INFO     : files: op(c): op(d): [started]  writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service"
Mar 20 21:30:11.698401 ignition[956]: INFO     : files: op(c): op(d): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service"
Mar 20 21:30:11.698401 ignition[956]: INFO     : files: op(c): [finished] processing unit "prepare-helm.service"
Mar 20 21:30:11.698401 ignition[956]: INFO     : files: op(e): [started]  processing unit "coreos-metadata.service"
Mar 20 21:30:11.702721 ignition[956]: INFO     : files: op(e): op(f): [started]  writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service"
Mar 20 21:30:11.702721 ignition[956]: INFO     : files: op(e): op(f): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service"
Mar 20 21:30:11.702721 ignition[956]: INFO     : files: op(e): [finished] processing unit "coreos-metadata.service"
Mar 20 21:30:11.702721 ignition[956]: INFO     : files: op(10): [started]  setting preset to disabled for "coreos-metadata.service"
Mar 20 21:30:11.721482 ignition[956]: INFO     : files: op(10): op(11): [started]  removing enablement symlink(s) for "coreos-metadata.service"
Mar 20 21:30:11.727572 ignition[956]: INFO     : files: op(10): op(11): [finished] removing enablement symlink(s) for "coreos-metadata.service"
Mar 20 21:30:11.729255 ignition[956]: INFO     : files: op(10): [finished] setting preset to disabled for "coreos-metadata.service"
Mar 20 21:30:11.729255 ignition[956]: INFO     : files: op(12): [started]  setting preset to enabled for "prepare-helm.service"
Mar 20 21:30:11.732019 ignition[956]: INFO     : files: op(12): [finished] setting preset to enabled for "prepare-helm.service"
Mar 20 21:30:11.733466 ignition[956]: INFO     : files: createResultFile: createFiles: op(13): [started]  writing file "/sysroot/etc/.ignition-result.json"
Mar 20 21:30:11.735204 ignition[956]: INFO     : files: createResultFile: createFiles: op(13): [finished] writing file "/sysroot/etc/.ignition-result.json"
Mar 20 21:30:11.736860 ignition[956]: INFO     : files: files passed
Mar 20 21:30:11.737600 ignition[956]: INFO     : Ignition finished successfully
Mar 20 21:30:11.740751 systemd[1]: Finished ignition-files.service - Ignition (files).
Mar 20 21:30:11.742960 systemd[1]: Starting ignition-quench.service - Ignition (record completion)...
Mar 20 21:30:11.744028 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion...
Mar 20 21:30:11.769586 systemd[1]: ignition-quench.service: Deactivated successfully.
Mar 20 21:30:11.769760 systemd[1]: Finished ignition-quench.service - Ignition (record completion).
Mar 20 21:30:11.774072 initrd-setup-root-after-ignition[986]: grep: /sysroot/oem/oem-release: No such file or directory
Mar 20 21:30:11.778001 initrd-setup-root-after-ignition[988]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory
Mar 20 21:30:11.778001 initrd-setup-root-after-ignition[988]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory
Mar 20 21:30:11.781444 initrd-setup-root-after-ignition[992]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory
Mar 20 21:30:11.784694 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion.
Mar 20 21:30:11.785204 systemd[1]: Reached target ignition-complete.target - Ignition Complete.
Mar 20 21:30:11.788141 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root...
Mar 20 21:30:11.855132 systemd[1]: initrd-parse-etc.service: Deactivated successfully.
Mar 20 21:30:11.855310 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root.
Mar 20 21:30:11.856266 systemd[1]: Reached target initrd-fs.target - Initrd File Systems.
Mar 20 21:30:11.859015 systemd[1]: Reached target initrd.target - Initrd Default Target.
Mar 20 21:30:11.861263 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met.
Mar 20 21:30:11.862376 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook...
Mar 20 21:30:11.894044 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook.
Mar 20 21:30:11.896153 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons...
Mar 20 21:30:11.921442 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups.
Mar 20 21:30:11.922060 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes.
Mar 20 21:30:11.922434 systemd[1]: Stopped target timers.target - Timer Units.
Mar 20 21:30:11.923027 systemd[1]: dracut-pre-pivot.service: Deactivated successfully.
Mar 20 21:30:11.923161 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook.
Mar 20 21:30:11.929856 systemd[1]: Stopped target initrd.target - Initrd Default Target.
Mar 20 21:30:11.930439 systemd[1]: Stopped target basic.target - Basic System.
Mar 20 21:30:11.931035 systemd[1]: Stopped target ignition-complete.target - Ignition Complete.
Mar 20 21:30:11.931369 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup.
Mar 20 21:30:11.931891 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device.
Mar 20 21:30:11.932224 systemd[1]: Stopped target remote-fs.target - Remote File Systems.
Mar 20 21:30:11.932553 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems.
Mar 20 21:30:11.933073 systemd[1]: Stopped target sysinit.target - System Initialization.
Mar 20 21:30:11.933375 systemd[1]: Stopped target local-fs.target - Local File Systems.
Mar 20 21:30:11.933870 systemd[1]: Stopped target swap.target - Swaps.
Mar 20 21:30:11.934197 systemd[1]: dracut-pre-mount.service: Deactivated successfully.
Mar 20 21:30:11.934311 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook.
Mar 20 21:30:11.952073 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes.
Mar 20 21:30:11.953199 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre).
Mar 20 21:30:11.955507 systemd[1]: clevis-luks-askpass.path: Deactivated successfully.
Mar 20 21:30:11.955651 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch.
Mar 20 21:30:11.957875 systemd[1]: dracut-initqueue.service: Deactivated successfully.
Mar 20 21:30:11.958011 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook.
Mar 20 21:30:11.960221 systemd[1]: ignition-fetch-offline.service: Deactivated successfully.
Mar 20 21:30:11.960329 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline).
Mar 20 21:30:11.963070 systemd[1]: Stopped target paths.target - Path Units.
Mar 20 21:30:11.965175 systemd[1]: systemd-ask-password-console.path: Deactivated successfully.
Mar 20 21:30:11.968696 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch.
Mar 20 21:30:11.969521 systemd[1]: Stopped target slices.target - Slice Units.
Mar 20 21:30:11.969945 systemd[1]: Stopped target sockets.target - Socket Units.
Mar 20 21:30:11.973533 systemd[1]: iscsid.socket: Deactivated successfully.
Mar 20 21:30:11.973644 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket.
Mar 20 21:30:11.975254 systemd[1]: iscsiuio.socket: Deactivated successfully.
Mar 20 21:30:11.975331 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket.
Mar 20 21:30:11.976989 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully.
Mar 20 21:30:11.977136 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion.
Mar 20 21:30:11.978745 systemd[1]: ignition-files.service: Deactivated successfully.
Mar 20 21:30:11.978853 systemd[1]: Stopped ignition-files.service - Ignition (files).
Mar 20 21:30:11.981506 systemd[1]: Stopping ignition-mount.service - Ignition (mount)...
Mar 20 21:30:11.982349 systemd[1]: kmod-static-nodes.service: Deactivated successfully.
Mar 20 21:30:11.982472 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes.
Mar 20 21:30:11.985655 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot...
Mar 20 21:30:11.988747 systemd[1]: systemd-udev-trigger.service: Deactivated successfully.
Mar 20 21:30:11.990672 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices.
Mar 20 21:30:11.993966 systemd[1]: dracut-pre-trigger.service: Deactivated successfully.
Mar 20 21:30:11.995111 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook.
Mar 20 21:30:12.001009 ignition[1012]: INFO     : Ignition 2.20.0
Mar 20 21:30:12.001009 ignition[1012]: INFO     : Stage: umount
Mar 20 21:30:12.002632 ignition[1012]: INFO     : no configs at "/usr/lib/ignition/base.d"
Mar 20 21:30:12.002632 ignition[1012]: INFO     : no config dir at "/usr/lib/ignition/base.platform.d/qemu"
Mar 20 21:30:12.002632 ignition[1012]: INFO     : umount: umount passed
Mar 20 21:30:12.002632 ignition[1012]: INFO     : Ignition finished successfully
Mar 20 21:30:12.006669 systemd[1]: ignition-mount.service: Deactivated successfully.
Mar 20 21:30:12.007637 systemd[1]: Stopped ignition-mount.service - Ignition (mount).
Mar 20 21:30:12.011709 systemd[1]: sysroot-boot.mount: Deactivated successfully.
Mar 20 21:30:12.013235 systemd[1]: initrd-cleanup.service: Deactivated successfully.
Mar 20 21:30:12.014234 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons.
Mar 20 21:30:12.018495 systemd[1]: Stopped target network.target - Network.
Mar 20 21:30:12.020437 systemd[1]: ignition-disks.service: Deactivated successfully.
Mar 20 21:30:12.021392 systemd[1]: Stopped ignition-disks.service - Ignition (disks).
Mar 20 21:30:12.023303 systemd[1]: ignition-kargs.service: Deactivated successfully.
Mar 20 21:30:12.024203 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs).
Mar 20 21:30:12.026156 systemd[1]: ignition-setup.service: Deactivated successfully.
Mar 20 21:30:12.027122 systemd[1]: Stopped ignition-setup.service - Ignition (setup).
Mar 20 21:30:12.029133 systemd[1]: ignition-setup-pre.service: Deactivated successfully.
Mar 20 21:30:12.029183 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup.
Mar 20 21:30:12.032326 systemd[1]: Stopping systemd-networkd.service - Network Configuration...
Mar 20 21:30:12.034435 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution...
Mar 20 21:30:12.042901 systemd[1]: systemd-resolved.service: Deactivated successfully.
Mar 20 21:30:12.043941 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution.
Mar 20 21:30:12.048290 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully.
Mar 20 21:30:12.049833 systemd[1]: systemd-networkd.service: Deactivated successfully.
Mar 20 21:30:12.050857 systemd[1]: Stopped systemd-networkd.service - Network Configuration.
Mar 20 21:30:12.054047 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully.
Mar 20 21:30:12.055932 systemd[1]: systemd-networkd.socket: Deactivated successfully.
Mar 20 21:30:12.056895 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket.
Mar 20 21:30:12.059884 systemd[1]: Stopping network-cleanup.service - Network Cleanup...
Mar 20 21:30:12.061762 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully.
Mar 20 21:30:12.062732 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline.
Mar 20 21:30:12.065261 systemd[1]: systemd-sysctl.service: Deactivated successfully.
Mar 20 21:30:12.066191 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables.
Mar 20 21:30:12.068271 systemd[1]: systemd-modules-load.service: Deactivated successfully.
Mar 20 21:30:12.069246 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules.
Mar 20 21:30:12.071381 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully.
Mar 20 21:30:12.071444 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories.
Mar 20 21:30:12.075049 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files...
Mar 20 21:30:12.078366 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully.
Mar 20 21:30:12.079578 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully.
Mar 20 21:30:12.084293 systemd[1]: systemd-udevd.service: Deactivated successfully.
Mar 20 21:30:12.085337 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files.
Mar 20 21:30:12.092791 systemd[1]: systemd-udevd-control.socket: Deactivated successfully.
Mar 20 21:30:12.092867 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket.
Mar 20 21:30:12.093379 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully.
Mar 20 21:30:12.093437 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket.
Mar 20 21:30:12.093849 systemd[1]: dracut-pre-udev.service: Deactivated successfully.
Mar 20 21:30:12.093912 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook.
Mar 20 21:30:12.094822 systemd[1]: dracut-cmdline.service: Deactivated successfully.
Mar 20 21:30:12.094885 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook.
Mar 20 21:30:12.095511 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully.
Mar 20 21:30:12.095560 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters.
Mar 20 21:30:12.097157 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database...
Mar 20 21:30:12.105227 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully.
Mar 20 21:30:12.105286 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev.
Mar 20 21:30:12.110385 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully.
Mar 20 21:30:12.110448 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup.
Mar 20 21:30:12.114572 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully.
Mar 20 21:30:12.114667 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully.
Mar 20 21:30:12.115074 systemd[1]: network-cleanup.service: Deactivated successfully.
Mar 20 21:30:12.115182 systemd[1]: Stopped network-cleanup.service - Network Cleanup.
Mar 20 21:30:12.116368 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully.
Mar 20 21:30:12.116476 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database.
Mar 20 21:30:12.268006 systemd[1]: sysroot-boot.service: Deactivated successfully.
Mar 20 21:30:12.269019 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot.
Mar 20 21:30:12.271009 systemd[1]: Reached target initrd-switch-root.target - Switch Root.
Mar 20 21:30:12.273258 systemd[1]: initrd-setup-root.service: Deactivated successfully.
Mar 20 21:30:12.274347 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup.
Mar 20 21:30:12.277354 systemd[1]: Starting initrd-switch-root.service - Switch Root...
Mar 20 21:30:12.297245 systemd[1]: Switching root.
Mar 20 21:30:12.330784 systemd-journald[193]: Journal stopped
Mar 20 21:30:13.709193 systemd-journald[193]: Received SIGTERM from PID 1 (systemd).
Mar 20 21:30:13.709268 kernel: SELinux:  policy capability network_peer_controls=1
Mar 20 21:30:13.709282 kernel: SELinux:  policy capability open_perms=1
Mar 20 21:30:13.709305 kernel: SELinux:  policy capability extended_socket_class=1
Mar 20 21:30:13.709321 kernel: SELinux:  policy capability always_check_network=0
Mar 20 21:30:13.709333 kernel: SELinux:  policy capability cgroup_seclabel=1
Mar 20 21:30:13.709350 kernel: SELinux:  policy capability nnp_nosuid_transition=1
Mar 20 21:30:13.709363 kernel: SELinux:  policy capability genfs_seclabel_symlinks=0
Mar 20 21:30:13.709374 kernel: SELinux:  policy capability ioctl_skip_cloexec=0
Mar 20 21:30:13.709400 kernel: audit: type=1403 audit(1742506212.860:2): auid=4294967295 ses=4294967295 lsm=selinux res=1
Mar 20 21:30:13.709412 systemd[1]: Successfully loaded SELinux policy in 40.940ms.
Mar 20 21:30:13.709428 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 13.132ms.
Mar 20 21:30:13.709442 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE)
Mar 20 21:30:13.709456 systemd[1]: Detected virtualization kvm.
Mar 20 21:30:13.709472 systemd[1]: Detected architecture x86-64.
Mar 20 21:30:13.709485 systemd[1]: Detected first boot.
Mar 20 21:30:13.709499 systemd[1]: Initializing machine ID from VM UUID.
Mar 20 21:30:13.709515 zram_generator::config[1058]: No configuration found.
Mar 20 21:30:13.709537 kernel: Guest personality initialized and is inactive
Mar 20 21:30:13.709550 kernel: VMCI host device registered (name=vmci, major=10, minor=125)
Mar 20 21:30:13.709562 kernel: Initialized host personality
Mar 20 21:30:13.709573 kernel: NET: Registered PF_VSOCK protocol family
Mar 20 21:30:13.709590 systemd[1]: Populated /etc with preset unit settings.
Mar 20 21:30:13.709603 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully.
Mar 20 21:30:13.709629 systemd[1]: initrd-switch-root.service: Deactivated successfully.
Mar 20 21:30:13.709642 systemd[1]: Stopped initrd-switch-root.service - Switch Root.
Mar 20 21:30:13.709657 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1.
Mar 20 21:30:13.709670 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config.
Mar 20 21:30:13.709684 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run.
Mar 20 21:30:13.709696 systemd[1]: Created slice system-getty.slice - Slice /system/getty.
Mar 20 21:30:13.709708 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe.
Mar 20 21:30:13.709720 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty.
Mar 20 21:30:13.709733 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit.
Mar 20 21:30:13.709746 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck.
Mar 20 21:30:13.709760 systemd[1]: Created slice user.slice - User and Session Slice.
Mar 20 21:30:13.709773 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch.
Mar 20 21:30:13.709785 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch.
Mar 20 21:30:13.709798 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch.
Mar 20 21:30:13.709810 systemd[1]: Set up automount boot.automount - Boot partition Automount Point.
Mar 20 21:30:13.709823 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point.
Mar 20 21:30:13.709836 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM...
Mar 20 21:30:13.709849 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0...
Mar 20 21:30:13.709863 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre).
Mar 20 21:30:13.709875 systemd[1]: Stopped target initrd-switch-root.target - Switch Root.
Mar 20 21:30:13.709888 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems.
Mar 20 21:30:13.709900 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System.
Mar 20 21:30:13.709912 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes.
Mar 20 21:30:13.709925 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes.
Mar 20 21:30:13.709937 systemd[1]: Reached target remote-fs.target - Remote File Systems.
Mar 20 21:30:13.709951 systemd[1]: Reached target slices.target - Slice Units.
Mar 20 21:30:13.709963 systemd[1]: Reached target swap.target - Swaps.
Mar 20 21:30:13.709980 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes.
Mar 20 21:30:13.709993 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket.
Mar 20 21:30:13.710007 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption.
Mar 20 21:30:13.710020 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket.
Mar 20 21:30:13.710033 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket.
Mar 20 21:30:13.710046 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket.
Mar 20 21:30:13.710058 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket.
Mar 20 21:30:13.710070 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System...
Mar 20 21:30:13.710083 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System...
Mar 20 21:30:13.710097 systemd[1]: Mounting media.mount - External Media Directory...
Mar 20 21:30:13.710110 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen).
Mar 20 21:30:13.710122 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System...
Mar 20 21:30:13.710135 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System...
Mar 20 21:30:13.710148 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp...
Mar 20 21:30:13.710161 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw).
Mar 20 21:30:13.710173 systemd[1]: Reached target machines.target - Containers.
Mar 20 21:30:13.710185 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files...
Mar 20 21:30:13.710200 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met.
Mar 20 21:30:13.710213 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes...
Mar 20 21:30:13.710227 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs...
Mar 20 21:30:13.710240 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod...
Mar 20 21:30:13.710253 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm...
Mar 20 21:30:13.710265 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore...
Mar 20 21:30:13.710279 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse...
Mar 20 21:30:13.710291 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop...
Mar 20 21:30:13.710304 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf).
Mar 20 21:30:13.710319 systemd[1]: systemd-fsck-root.service: Deactivated successfully.
Mar 20 21:30:13.710331 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device.
Mar 20 21:30:13.710346 systemd[1]: systemd-fsck-usr.service: Deactivated successfully.
Mar 20 21:30:13.710359 systemd[1]: Stopped systemd-fsck-usr.service.
Mar 20 21:30:13.710372 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67).
Mar 20 21:30:13.710396 systemd[1]: Starting systemd-journald.service - Journal Service...
Mar 20 21:30:13.710409 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules...
Mar 20 21:30:13.710422 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line...
Mar 20 21:30:13.710437 kernel: fuse: init (API version 7.39)
Mar 20 21:30:13.710449 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems...
Mar 20 21:30:13.710481 systemd-journald[1123]: Collecting audit messages is disabled.
Mar 20 21:30:13.710507 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials...
Mar 20 21:30:13.710520 kernel: loop: module loaded
Mar 20 21:30:13.710532 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices...
Mar 20 21:30:13.710546 systemd-journald[1123]: Journal started
Mar 20 21:30:13.710571 systemd-journald[1123]: Runtime Journal (/run/log/journal/46d2dd9d9ab54cf4aaf8fd0b2963e989) is 6M, max 48.3M, 42.3M free.
Mar 20 21:30:13.427120 systemd[1]: Queued start job for default target multi-user.target.
Mar 20 21:30:13.439507 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6.
Mar 20 21:30:13.440009 systemd[1]: systemd-journald.service: Deactivated successfully.
Mar 20 21:30:13.713014 systemd[1]: verity-setup.service: Deactivated successfully.
Mar 20 21:30:13.715718 systemd[1]: Stopped verity-setup.service.
Mar 20 21:30:13.718644 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen).
Mar 20 21:30:13.740651 systemd[1]: Started systemd-journald.service - Journal Service.
Mar 20 21:30:13.741509 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System.
Mar 20 21:30:13.742733 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System.
Mar 20 21:30:13.743980 systemd[1]: Mounted media.mount - External Media Directory.
Mar 20 21:30:13.745106 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System.
Mar 20 21:30:13.746399 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System.
Mar 20 21:30:13.747667 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp.
Mar 20 21:30:13.748973 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes.
Mar 20 21:30:13.750589 systemd[1]: modprobe@configfs.service: Deactivated successfully.
Mar 20 21:30:13.750849 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs.
Mar 20 21:30:13.752385 systemd[1]: modprobe@dm_mod.service: Deactivated successfully.
Mar 20 21:30:13.752603 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod.
Mar 20 21:30:13.754117 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully.
Mar 20 21:30:13.754332 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore.
Mar 20 21:30:13.755855 systemd[1]: modprobe@fuse.service: Deactivated successfully.
Mar 20 21:30:13.756070 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse.
Mar 20 21:30:13.757800 systemd[1]: modprobe@loop.service: Deactivated successfully.
Mar 20 21:30:13.758001 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop.
Mar 20 21:30:13.766128 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line.
Mar 20 21:30:13.769684 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules.
Mar 20 21:30:13.771401 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems.
Mar 20 21:30:13.776856 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials.
Mar 20 21:30:13.784888 systemd[1]: Reached target network-pre.target - Preparation for Network.
Mar 20 21:30:13.788849 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System...
Mar 20 21:30:13.791478 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System...
Mar 20 21:30:13.792647 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/).
Mar 20 21:30:13.792687 systemd[1]: Reached target local-fs.target - Local File Systems.
Mar 20 21:30:13.795004 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management.
Mar 20 21:30:13.803640 kernel: ACPI: bus type drm_connector registered
Mar 20 21:30:13.810362 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown...
Mar 20 21:30:13.815140 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache...
Mar 20 21:30:13.817789 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met.
Mar 20 21:30:13.819070 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database...
Mar 20 21:30:13.822678 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage...
Mar 20 21:30:13.824821 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore).
Mar 20 21:30:13.826058 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed...
Mar 20 21:30:13.827260 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met.
Mar 20 21:30:13.829746 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables...
Mar 20 21:30:13.837771 systemd-journald[1123]: Time spent on flushing to /var/log/journal/46d2dd9d9ab54cf4aaf8fd0b2963e989 is 15.272ms for 963 entries.
Mar 20 21:30:13.837771 systemd-journald[1123]: System Journal (/var/log/journal/46d2dd9d9ab54cf4aaf8fd0b2963e989) is 8M, max 195.6M, 187.6M free.
Mar 20 21:30:14.032901 systemd-journald[1123]: Received client request to flush runtime journal.
Mar 20 21:30:14.032959 kernel: loop0: detected capacity change from 0 to 151640
Mar 20 21:30:14.032991 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher
Mar 20 21:30:13.839863 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/...
Mar 20 21:30:13.845042 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files.
Mar 20 21:30:13.846749 systemd[1]: modprobe@drm.service: Deactivated successfully.
Mar 20 21:30:13.846974 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm.
Mar 20 21:30:13.848488 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System.
Mar 20 21:30:13.849868 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System.
Mar 20 21:30:13.851445 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown.
Mar 20 21:30:13.868774 systemd[1]: Starting systemd-sysusers.service - Create System Users...
Mar 20 21:30:13.894876 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices.
Mar 20 21:30:13.904571 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization...
Mar 20 21:30:13.925790 udevadm[1188]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in.
Mar 20 21:30:13.930907 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables.
Mar 20 21:30:13.963964 systemd[1]: Finished systemd-sysusers.service - Create System Users.
Mar 20 21:30:13.966496 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev...
Mar 20 21:30:14.033000 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed.
Mar 20 21:30:14.034765 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage.
Mar 20 21:30:14.039340 systemd[1]: Reached target first-boot-complete.target - First Boot Complete.
Mar 20 21:30:14.042727 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk...
Mar 20 21:30:14.046120 systemd-tmpfiles[1193]: ACLs are not supported, ignoring.
Mar 20 21:30:14.046137 systemd-tmpfiles[1193]: ACLs are not supported, ignoring.
Mar 20 21:30:14.052878 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev.
Mar 20 21:30:14.057639 kernel: loop1: detected capacity change from 0 to 109808
Mar 20 21:30:14.126434 kernel: loop2: detected capacity change from 0 to 205544
Mar 20 21:30:14.158649 kernel: loop3: detected capacity change from 0 to 151640
Mar 20 21:30:14.170648 kernel: loop4: detected capacity change from 0 to 109808
Mar 20 21:30:14.184644 kernel: loop5: detected capacity change from 0 to 205544
Mar 20 21:30:14.187749 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk.
Mar 20 21:30:14.197655 (sd-merge)[1203]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'.
Mar 20 21:30:14.198289 (sd-merge)[1203]: Merged extensions into '/usr'.
Mar 20 21:30:14.202660 systemd[1]: Reload requested from client PID 1177 ('systemd-sysext') (unit systemd-sysext.service)...
Mar 20 21:30:14.202679 systemd[1]: Reloading...
Mar 20 21:30:14.268655 zram_generator::config[1232]: No configuration found.
Mar 20 21:30:14.335776 ldconfig[1172]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start.
Mar 20 21:30:14.409137 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly.
Mar 20 21:30:14.477982 systemd[1]: etc-machine\x2did.mount: Deactivated successfully.
Mar 20 21:30:14.478595 systemd[1]: Reloading finished in 275 ms.
Mar 20 21:30:14.503494 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache.
Mar 20 21:30:14.505018 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/.
Mar 20 21:30:14.519324 systemd[1]: Starting ensure-sysext.service...
Mar 20 21:30:14.521548 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories...
Mar 20 21:30:14.534049 systemd[1]: Reload requested from client PID 1269 ('systemctl') (unit ensure-sysext.service)...
Mar 20 21:30:14.534068 systemd[1]: Reloading...
Mar 20 21:30:14.585817 zram_generator::config[1299]: No configuration found.
Mar 20 21:30:14.650889 systemd-tmpfiles[1270]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring.
Mar 20 21:30:14.651164 systemd-tmpfiles[1270]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring.
Mar 20 21:30:14.652116 systemd-tmpfiles[1270]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring.
Mar 20 21:30:14.652391 systemd-tmpfiles[1270]: ACLs are not supported, ignoring.
Mar 20 21:30:14.652466 systemd-tmpfiles[1270]: ACLs are not supported, ignoring.
Mar 20 21:30:14.656556 systemd-tmpfiles[1270]: Detected autofs mount point /boot during canonicalization of boot.
Mar 20 21:30:14.656568 systemd-tmpfiles[1270]: Skipping /boot
Mar 20 21:30:14.669266 systemd-tmpfiles[1270]: Detected autofs mount point /boot during canonicalization of boot.
Mar 20 21:30:14.669278 systemd-tmpfiles[1270]: Skipping /boot
Mar 20 21:30:14.741999 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly.
Mar 20 21:30:14.811677 systemd[1]: Reloading finished in 277 ms.
Mar 20 21:30:14.824360 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database.
Mar 20 21:30:14.854273 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories.
Mar 20 21:30:14.863386 systemd[1]: Starting audit-rules.service - Load Audit Rules...
Mar 20 21:30:14.865865 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs...
Mar 20 21:30:14.868266 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog...
Mar 20 21:30:14.879570 systemd[1]: Starting systemd-resolved.service - Network Name Resolution...
Mar 20 21:30:14.884946 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files...
Mar 20 21:30:14.889850 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP...
Mar 20 21:30:14.896756 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen).
Mar 20 21:30:14.897169 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met.
Mar 20 21:30:14.899429 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod...
Mar 20 21:30:14.901950 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore...
Mar 20 21:30:14.905759 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop...
Mar 20 21:30:14.907033 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met.
Mar 20 21:30:14.907174 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67).
Mar 20 21:30:14.912560 systemd[1]: Starting systemd-userdbd.service - User Database Manager...
Mar 20 21:30:14.913666 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen).
Mar 20 21:30:14.915424 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog.
Mar 20 21:30:14.917708 systemd[1]: modprobe@dm_mod.service: Deactivated successfully.
Mar 20 21:30:14.917992 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod.
Mar 20 21:30:14.919945 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully.
Mar 20 21:30:14.920162 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore.
Mar 20 21:30:14.922069 systemd[1]: modprobe@loop.service: Deactivated successfully.
Mar 20 21:30:14.922280 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop.
Mar 20 21:30:14.931226 systemd-udevd[1342]: Using default interface naming scheme 'v255'.
Mar 20 21:30:14.946916 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP.
Mar 20 21:30:14.951308 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen).
Mar 20 21:30:14.951704 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met.
Mar 20 21:30:14.954925 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod...
Mar 20 21:30:14.961176 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore...
Mar 20 21:30:14.963806 augenrules[1373]: No rules
Mar 20 21:30:14.967004 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop...
Mar 20 21:30:14.968526 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met.
Mar 20 21:30:14.968700 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67).
Mar 20 21:30:14.972264 systemd[1]: Starting systemd-update-done.service - Update is Completed...
Mar 20 21:30:14.973669 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen).
Mar 20 21:30:14.975051 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files.
Mar 20 21:30:14.977922 systemd[1]: audit-rules.service: Deactivated successfully.
Mar 20 21:30:14.978555 systemd[1]: Finished audit-rules.service - Load Audit Rules.
Mar 20 21:30:14.984582 systemd[1]: modprobe@dm_mod.service: Deactivated successfully.
Mar 20 21:30:14.985287 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod.
Mar 20 21:30:14.995632 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs.
Mar 20 21:30:14.997307 systemd[1]: Started systemd-userdbd.service - User Database Manager.
Mar 20 21:30:15.000014 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully.
Mar 20 21:30:15.000274 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore.
Mar 20 21:30:15.002083 systemd[1]: modprobe@loop.service: Deactivated successfully.
Mar 20 21:30:15.002661 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop.
Mar 20 21:30:15.024786 systemd[1]: Finished systemd-update-done.service - Update is Completed.
Mar 20 21:30:15.032882 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 39 scanned by (udev-worker) (1392)
Mar 20 21:30:15.039958 systemd[1]: Finished ensure-sysext.service.
Mar 20 21:30:15.046834 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen).
Mar 20 21:30:15.050794 systemd[1]: Starting audit-rules.service - Load Audit Rules...
Mar 20 21:30:15.052432 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met.
Mar 20 21:30:15.056007 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod...
Mar 20 21:30:15.060837 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm...
Mar 20 21:30:15.068543 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore...
Mar 20 21:30:15.072846 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop...
Mar 20 21:30:15.074110 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met.
Mar 20 21:30:15.074153 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67).
Mar 20 21:30:15.077808 systemd[1]: Starting systemd-networkd.service - Network Configuration...
Mar 20 21:30:15.080942 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization...
Mar 20 21:30:15.082692 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt).
Mar 20 21:30:15.082726 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen).
Mar 20 21:30:15.083421 systemd[1]: modprobe@dm_mod.service: Deactivated successfully.
Mar 20 21:30:15.083723 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod.
Mar 20 21:30:15.085219 systemd[1]: modprobe@drm.service: Deactivated successfully.
Mar 20 21:30:15.085449 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm.
Mar 20 21:30:15.086795 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped.
Mar 20 21:30:15.093583 augenrules[1413]: /sbin/augenrules: No change
Mar 20 21:30:15.097038 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully.
Mar 20 21:30:15.097282 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore.
Mar 20 21:30:15.101226 systemd[1]: modprobe@loop.service: Deactivated successfully.
Mar 20 21:30:15.103036 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop.
Mar 20 21:30:15.109007 augenrules[1444]: No rules
Mar 20 21:30:15.113517 systemd-resolved[1341]: Positive Trust Anchors:
Mar 20 21:30:15.113541 systemd-resolved[1341]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d
Mar 20 21:30:15.113573 systemd-resolved[1341]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test
Mar 20 21:30:15.114663 systemd[1]: audit-rules.service: Deactivated successfully.
Mar 20 21:30:15.116031 systemd[1]: Finished audit-rules.service - Load Audit Rules.
Mar 20 21:30:15.120910 systemd-resolved[1341]: Defaulting to hostname 'linux'.
Mar 20 21:30:15.124884 systemd[1]: Started systemd-resolved.service - Network Name Resolution.
Mar 20 21:30:15.134828 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM.
Mar 20 21:30:15.138178 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups.
Mar 20 21:30:15.142817 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM...
Mar 20 21:30:15.144416 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore).
Mar 20 21:30:15.144517 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met.
Mar 20 21:30:15.149674 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input2
Mar 20 21:30:15.156819 kernel: ACPI: button: Power Button [PWRF]
Mar 20 21:30:15.162510 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM.
Mar 20 21:30:15.185456 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt
Mar 20 21:30:15.185875 kernel: i2c i2c-0: 1/1 memory slots populated (from DMI)
Mar 20 21:30:15.189396 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD
Mar 20 21:30:15.193195 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input3
Mar 20 21:30:15.207711 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization.
Mar 20 21:30:15.209204 systemd[1]: Reached target time-set.target - System Time Set.
Mar 20 21:30:15.210372 systemd-networkd[1424]: lo: Link UP
Mar 20 21:30:15.210377 systemd-networkd[1424]: lo: Gained carrier
Mar 20 21:30:15.212785 systemd-networkd[1424]: Enumeration completed
Mar 20 21:30:15.212879 systemd[1]: Started systemd-networkd.service - Network Configuration.
Mar 20 21:30:15.214174 systemd-networkd[1424]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name.
Mar 20 21:30:15.214180 systemd-networkd[1424]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network.
Mar 20 21:30:15.214345 systemd[1]: Reached target network.target - Network.
Mar 20 21:30:15.216034 systemd-networkd[1424]: eth0: Link UP
Mar 20 21:30:15.216039 systemd-networkd[1424]: eth0: Gained carrier
Mar 20 21:30:15.216055 systemd-networkd[1424]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name.
Mar 20 21:30:15.219805 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd...
Mar 20 21:30:15.224132 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured...
Mar 20 21:30:15.237721 systemd-networkd[1424]: eth0: DHCPv4 address 10.0.0.132/16, gateway 10.0.0.1 acquired from 10.0.0.1
Mar 20 21:30:15.238510 systemd-timesyncd[1427]: Network configuration changed, trying to establish connection.
Mar 20 21:30:15.239574 systemd-timesyncd[1427]: Contacted time server 10.0.0.1:123 (10.0.0.1).
Mar 20 21:30:15.239639 systemd-timesyncd[1427]: Initial clock synchronization to Thu 2025-03-20 21:30:15.056645 UTC.
Mar 20 21:30:15.246584 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup...
Mar 20 21:30:15.254842 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd.
Mar 20 21:30:15.285743 kernel: mousedev: PS/2 mouse device common for all mice
Mar 20 21:30:15.297703 kernel: kvm_amd: TSC scaling supported
Mar 20 21:30:15.297741 kernel: kvm_amd: Nested Virtualization enabled
Mar 20 21:30:15.297755 kernel: kvm_amd: Nested Paging enabled
Mar 20 21:30:15.298813 kernel: kvm_amd: LBR virtualization supported
Mar 20 21:30:15.298838 kernel: kvm_amd: Virtual VMLOAD VMSAVE supported
Mar 20 21:30:15.300033 kernel: kvm_amd: Virtual GIF supported
Mar 20 21:30:15.320658 kernel: EDAC MC: Ver: 3.0.0
Mar 20 21:30:15.356490 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization.
Mar 20 21:30:15.360797 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup.
Mar 20 21:30:15.364231 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes...
Mar 20 21:30:15.388834 lvm[1472]:   WARNING: Failed to connect to lvmetad. Falling back to device scanning.
Mar 20 21:30:15.424420 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes.
Mar 20 21:30:15.426168 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes.
Mar 20 21:30:15.427434 systemd[1]: Reached target sysinit.target - System Initialization.
Mar 20 21:30:15.428779 systemd[1]: Started motdgen.path - Watch for update engine configuration changes.
Mar 20 21:30:15.430164 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data.
Mar 20 21:30:15.431816 systemd[1]: Started logrotate.timer - Daily rotation of log files.
Mar 20 21:30:15.433137 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information..
Mar 20 21:30:15.434470 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories.
Mar 20 21:30:15.435789 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate).
Mar 20 21:30:15.435829 systemd[1]: Reached target paths.target - Path Units.
Mar 20 21:30:15.436829 systemd[1]: Reached target timers.target - Timer Units.
Mar 20 21:30:15.439126 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket.
Mar 20 21:30:15.442066 systemd[1]: Starting docker.socket - Docker Socket for the API...
Mar 20 21:30:15.445784 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local).
Mar 20 21:30:15.447226 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK).
Mar 20 21:30:15.448527 systemd[1]: Reached target ssh-access.target - SSH Access Available.
Mar 20 21:30:15.453207 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket.
Mar 20 21:30:15.454635 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket.
Mar 20 21:30:15.457002 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes...
Mar 20 21:30:15.458646 systemd[1]: Listening on docker.socket - Docker Socket for the API.
Mar 20 21:30:15.459799 systemd[1]: Reached target sockets.target - Socket Units.
Mar 20 21:30:15.460777 systemd[1]: Reached target basic.target - Basic System.
Mar 20 21:30:15.461759 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met.
Mar 20 21:30:15.461789 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met.
Mar 20 21:30:15.462792 systemd[1]: Starting containerd.service - containerd container runtime...
Mar 20 21:30:15.464849 systemd[1]: Starting dbus.service - D-Bus System Message Bus...
Mar 20 21:30:15.469223 lvm[1476]:   WARNING: Failed to connect to lvmetad. Falling back to device scanning.
Mar 20 21:30:15.469315 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit...
Mar 20 21:30:15.478175 systemd[1]: Starting extend-filesystems.service - Extend Filesystems...
Mar 20 21:30:15.480761 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment).
Mar 20 21:30:15.481866 jq[1479]: false
Mar 20 21:30:15.482063 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd...
Mar 20 21:30:15.485704 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin...
Mar 20 21:30:15.488870 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline...
Mar 20 21:30:15.493740 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys...
Mar 20 21:30:15.498236 systemd[1]: Starting systemd-logind.service - User Login Management...
Mar 20 21:30:15.500232 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0).
Mar 20 21:30:15.500775 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details.
Mar 20 21:30:15.502349 systemd[1]: Starting update-engine.service - Update Engine...
Mar 20 21:30:15.503796 dbus-daemon[1478]: [system] SELinux support is enabled
Mar 20 21:30:15.505658 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition...
Mar 20 21:30:15.508637 extend-filesystems[1480]: Found loop3
Mar 20 21:30:15.520209 extend-filesystems[1480]: Found loop4
Mar 20 21:30:15.520209 extend-filesystems[1480]: Found loop5
Mar 20 21:30:15.520209 extend-filesystems[1480]: Found sr0
Mar 20 21:30:15.520209 extend-filesystems[1480]: Found vda
Mar 20 21:30:15.520209 extend-filesystems[1480]: Found vda1
Mar 20 21:30:15.520209 extend-filesystems[1480]: Found vda2
Mar 20 21:30:15.520209 extend-filesystems[1480]: Found vda3
Mar 20 21:30:15.520209 extend-filesystems[1480]: Found usr
Mar 20 21:30:15.520209 extend-filesystems[1480]: Found vda4
Mar 20 21:30:15.520209 extend-filesystems[1480]: Found vda6
Mar 20 21:30:15.520209 extend-filesystems[1480]: Found vda7
Mar 20 21:30:15.520209 extend-filesystems[1480]: Found vda9
Mar 20 21:30:15.520209 extend-filesystems[1480]: Checking size of /dev/vda9
Mar 20 21:30:15.520209 extend-filesystems[1480]: Resized partition /dev/vda9
Mar 20 21:30:15.556877 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 39 scanned by (udev-worker) (1398)
Mar 20 21:30:15.556907 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks
Mar 20 21:30:15.511947 systemd[1]: Started dbus.service - D-Bus System Message Bus.
Mar 20 21:30:15.557079 extend-filesystems[1502]: resize2fs 1.47.2 (1-Jan-2025)
Mar 20 21:30:15.517357 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes.
Mar 20 21:30:15.558479 update_engine[1489]: I20250320 21:30:15.525638  1489 main.cc:92] Flatcar Update Engine starting
Mar 20 21:30:15.558479 update_engine[1489]: I20250320 21:30:15.538874  1489 update_check_scheduler.cc:74] Next update check in 11m30s
Mar 20 21:30:15.521850 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'.
Mar 20 21:30:15.567111 jq[1491]: true
Mar 20 21:30:15.522283 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped.
Mar 20 21:30:15.523000 systemd[1]: motdgen.service: Deactivated successfully.
Mar 20 21:30:15.523274 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd.
Mar 20 21:30:15.531737 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully.
Mar 20 21:30:15.532459 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline.
Mar 20 21:30:15.572637 kernel: EXT4-fs (vda9): resized filesystem to 1864699
Mar 20 21:30:15.571033 (ntainerd)[1505]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR
Mar 20 21:30:15.592338 tar[1503]: linux-amd64/helm
Mar 20 21:30:15.606972 jq[1504]: true
Mar 20 21:30:15.607208 extend-filesystems[1502]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required
Mar 20 21:30:15.607208 extend-filesystems[1502]: old_desc_blocks = 1, new_desc_blocks = 1
Mar 20 21:30:15.607208 extend-filesystems[1502]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long.
Mar 20 21:30:15.616488 extend-filesystems[1480]: Resized filesystem in /dev/vda9
Mar 20 21:30:15.607731 systemd[1]: extend-filesystems.service: Deactivated successfully.
Mar 20 21:30:15.608106 systemd[1]: Finished extend-filesystems.service - Extend Filesystems.
Mar 20 21:30:15.613804 systemd-logind[1487]: Watching system buttons on /dev/input/event1 (Power Button)
Mar 20 21:30:15.613825 systemd-logind[1487]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard)
Mar 20 21:30:15.616978 systemd-logind[1487]: New seat seat0.
Mar 20 21:30:15.626408 systemd[1]: Started systemd-logind.service - User Login Management.
Mar 20 21:30:15.627795 systemd[1]: Started update-engine.service - Update Engine.
Mar 20 21:30:15.631169 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml).
Mar 20 21:30:15.631355 systemd[1]: Reached target system-config.target - Load system-provided cloud configs.
Mar 20 21:30:15.635779 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url).
Mar 20 21:30:15.635904 systemd[1]: Reached target user-config.target - Load user-provided cloud configs.
Mar 20 21:30:15.639350 systemd[1]: Started locksmithd.service - Cluster reboot manager.
Mar 20 21:30:15.659431 bash[1534]: Updated "/home/core/.ssh/authorized_keys"
Mar 20 21:30:15.663445 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition.
Mar 20 21:30:15.665490 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met.
Mar 20 21:30:15.675973 sshd_keygen[1498]: ssh-keygen: generating new host keys: RSA ECDSA ED25519
Mar 20 21:30:15.683152 locksmithd[1533]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot"
Mar 20 21:30:15.700058 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys.
Mar 20 21:30:15.703866 systemd[1]: Starting issuegen.service - Generate /run/issue...
Mar 20 21:30:15.724478 systemd[1]: issuegen.service: Deactivated successfully.
Mar 20 21:30:15.724800 systemd[1]: Finished issuegen.service - Generate /run/issue.
Mar 20 21:30:15.727866 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions...
Mar 20 21:30:15.747636 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions.
Mar 20 21:30:15.750824 systemd[1]: Started getty@tty1.service - Getty on tty1.
Mar 20 21:30:15.755825 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0.
Mar 20 21:30:15.757131 systemd[1]: Reached target getty.target - Login Prompts.
Mar 20 21:30:15.772936 containerd[1505]: time="2025-03-20T21:30:15Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8
Mar 20 21:30:15.774709 containerd[1505]: time="2025-03-20T21:30:15.774662860Z" level=info msg="starting containerd" revision=88aa2f531d6c2922003cc7929e51daf1c14caa0a version=v2.0.1
Mar 20 21:30:15.783929 containerd[1505]: time="2025-03-20T21:30:15.783894329Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="6.884µs"
Mar 20 21:30:15.783929 containerd[1505]: time="2025-03-20T21:30:15.783923173Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1
Mar 20 21:30:15.784002 containerd[1505]: time="2025-03-20T21:30:15.783941397Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1
Mar 20 21:30:15.784126 containerd[1505]: time="2025-03-20T21:30:15.784106437Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1
Mar 20 21:30:15.784147 containerd[1505]: time="2025-03-20T21:30:15.784125503Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1
Mar 20 21:30:15.784181 containerd[1505]: time="2025-03-20T21:30:15.784169465Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1
Mar 20 21:30:15.784265 containerd[1505]: time="2025-03-20T21:30:15.784245878Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1
Mar 20 21:30:15.784265 containerd[1505]: time="2025-03-20T21:30:15.784259734Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1
Mar 20 21:30:15.784566 containerd[1505]: time="2025-03-20T21:30:15.784542114Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1
Mar 20 21:30:15.784566 containerd[1505]: time="2025-03-20T21:30:15.784560348Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1
Mar 20 21:30:15.784604 containerd[1505]: time="2025-03-20T21:30:15.784570727Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1
Mar 20 21:30:15.784604 containerd[1505]: time="2025-03-20T21:30:15.784588621Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1
Mar 20 21:30:15.784716 containerd[1505]: time="2025-03-20T21:30:15.784692325Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1
Mar 20 21:30:15.784958 containerd[1505]: time="2025-03-20T21:30:15.784930893Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1
Mar 20 21:30:15.784981 containerd[1505]: time="2025-03-20T21:30:15.784969716Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1
Mar 20 21:30:15.784981 containerd[1505]: time="2025-03-20T21:30:15.784978813Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1
Mar 20 21:30:15.785025 containerd[1505]: time="2025-03-20T21:30:15.785000523Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1
Mar 20 21:30:15.785297 containerd[1505]: time="2025-03-20T21:30:15.785239081Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1
Mar 20 21:30:15.785428 containerd[1505]: time="2025-03-20T21:30:15.785399031Z" level=info msg="metadata content store policy set" policy=shared
Mar 20 21:30:15.791097 containerd[1505]: time="2025-03-20T21:30:15.791054314Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1
Mar 20 21:30:15.791138 containerd[1505]: time="2025-03-20T21:30:15.791097355Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1
Mar 20 21:30:15.791138 containerd[1505]: time="2025-03-20T21:30:15.791112784Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1
Mar 20 21:30:15.791138 containerd[1505]: time="2025-03-20T21:30:15.791125578Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1
Mar 20 21:30:15.791208 containerd[1505]: time="2025-03-20T21:30:15.791138232Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1
Mar 20 21:30:15.791208 containerd[1505]: time="2025-03-20T21:30:15.791163439Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1
Mar 20 21:30:15.791208 containerd[1505]: time="2025-03-20T21:30:15.791176704Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1
Mar 20 21:30:15.791208 containerd[1505]: time="2025-03-20T21:30:15.791188907Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1
Mar 20 21:30:15.791208 containerd[1505]: time="2025-03-20T21:30:15.791200027Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1
Mar 20 21:30:15.791300 containerd[1505]: time="2025-03-20T21:30:15.791210898Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1
Mar 20 21:30:15.791300 containerd[1505]: time="2025-03-20T21:30:15.791221758Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1
Mar 20 21:30:15.791300 containerd[1505]: time="2025-03-20T21:30:15.791234161Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2
Mar 20 21:30:15.791400 containerd[1505]: time="2025-03-20T21:30:15.791368694Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1
Mar 20 21:30:15.791400 containerd[1505]: time="2025-03-20T21:30:15.791396446Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1
Mar 20 21:30:15.791440 containerd[1505]: time="2025-03-20T21:30:15.791414279Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1
Mar 20 21:30:15.791440 containerd[1505]: time="2025-03-20T21:30:15.791425230Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1
Mar 20 21:30:15.791440 containerd[1505]: time="2025-03-20T21:30:15.791435679Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1
Mar 20 21:30:15.791504 containerd[1505]: time="2025-03-20T21:30:15.791446169Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1
Mar 20 21:30:15.791504 containerd[1505]: time="2025-03-20T21:30:15.791458141Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1
Mar 20 21:30:15.791504 containerd[1505]: time="2025-03-20T21:30:15.791468801Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1
Mar 20 21:30:15.791504 containerd[1505]: time="2025-03-20T21:30:15.791481245Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1
Mar 20 21:30:15.791504 containerd[1505]: time="2025-03-20T21:30:15.791496052Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1
Mar 20 21:30:15.791602 containerd[1505]: time="2025-03-20T21:30:15.791506672Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1
Mar 20 21:30:15.791602 containerd[1505]: time="2025-03-20T21:30:15.791569280Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\""
Mar 20 21:30:15.791602 containerd[1505]: time="2025-03-20T21:30:15.791585280Z" level=info msg="Start snapshots syncer"
Mar 20 21:30:15.791697 containerd[1505]: time="2025-03-20T21:30:15.791675880Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1
Mar 20 21:30:15.792076 containerd[1505]: time="2025-03-20T21:30:15.792028621Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}"
Mar 20 21:30:15.792192 containerd[1505]: time="2025-03-20T21:30:15.792082292Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1
Mar 20 21:30:15.792192 containerd[1505]: time="2025-03-20T21:30:15.792152062Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1
Mar 20 21:30:15.792354 containerd[1505]: time="2025-03-20T21:30:15.792316501Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1
Mar 20 21:30:15.792354 containerd[1505]: time="2025-03-20T21:30:15.792350845Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1
Mar 20 21:30:15.792401 containerd[1505]: time="2025-03-20T21:30:15.792363689Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1
Mar 20 21:30:15.792401 containerd[1505]: time="2025-03-20T21:30:15.792374600Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1
Mar 20 21:30:15.792401 containerd[1505]: time="2025-03-20T21:30:15.792385961Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1
Mar 20 21:30:15.792401 containerd[1505]: time="2025-03-20T21:30:15.792397292Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1
Mar 20 21:30:15.792479 containerd[1505]: time="2025-03-20T21:30:15.792408563Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1
Mar 20 21:30:15.792479 containerd[1505]: time="2025-03-20T21:30:15.792429603Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1
Mar 20 21:30:15.792479 containerd[1505]: time="2025-03-20T21:30:15.792441134Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1
Mar 20 21:30:15.792479 containerd[1505]: time="2025-03-20T21:30:15.792451083Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1
Mar 20 21:30:15.792550 containerd[1505]: time="2025-03-20T21:30:15.792488052Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1
Mar 20 21:30:15.792550 containerd[1505]: time="2025-03-20T21:30:15.792501898Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1
Mar 20 21:30:15.792550 containerd[1505]: time="2025-03-20T21:30:15.792511456Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1
Mar 20 21:30:15.792550 containerd[1505]: time="2025-03-20T21:30:15.792542004Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1
Mar 20 21:30:15.792642 containerd[1505]: time="2025-03-20T21:30:15.792551632Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1
Mar 20 21:30:15.793503 containerd[1505]: time="2025-03-20T21:30:15.792870500Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1
Mar 20 21:30:15.793503 containerd[1505]: time="2025-03-20T21:30:15.792907890Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1
Mar 20 21:30:15.793503 containerd[1505]: time="2025-03-20T21:30:15.792934119Z" level=info msg="runtime interface created"
Mar 20 21:30:15.793503 containerd[1505]: time="2025-03-20T21:30:15.792941964Z" level=info msg="created NRI interface"
Mar 20 21:30:15.793503 containerd[1505]: time="2025-03-20T21:30:15.792959266Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1
Mar 20 21:30:15.793503 containerd[1505]: time="2025-03-20T21:30:15.792982059Z" level=info msg="Connect containerd service"
Mar 20 21:30:15.793503 containerd[1505]: time="2025-03-20T21:30:15.793019228Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this"
Mar 20 21:30:15.793988 containerd[1505]: time="2025-03-20T21:30:15.793953801Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config"
Mar 20 21:30:15.874697 containerd[1505]: time="2025-03-20T21:30:15.874657871Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc
Mar 20 21:30:15.874967 containerd[1505]: time="2025-03-20T21:30:15.874832879Z" level=info msg="Start subscribing containerd event"
Mar 20 21:30:15.875019 containerd[1505]: time="2025-03-20T21:30:15.874983081Z" level=info msg=serving... address=/run/containerd/containerd.sock
Mar 20 21:30:15.875019 containerd[1505]: time="2025-03-20T21:30:15.874989463Z" level=info msg="Start recovering state"
Mar 20 21:30:15.875137 containerd[1505]: time="2025-03-20T21:30:15.875106382Z" level=info msg="Start event monitor"
Mar 20 21:30:15.875287 containerd[1505]: time="2025-03-20T21:30:15.875275018Z" level=info msg="Start cni network conf syncer for default"
Mar 20 21:30:15.875364 containerd[1505]: time="2025-03-20T21:30:15.875348957Z" level=info msg="Start streaming server"
Mar 20 21:30:15.875410 containerd[1505]: time="2025-03-20T21:30:15.875400082Z" level=info msg="Registered namespace \"k8s.io\" with NRI"
Mar 20 21:30:15.875457 containerd[1505]: time="2025-03-20T21:30:15.875441941Z" level=info msg="runtime interface starting up..."
Mar 20 21:30:15.875498 containerd[1505]: time="2025-03-20T21:30:15.875487987Z" level=info msg="starting plugins..."
Mar 20 21:30:15.875558 containerd[1505]: time="2025-03-20T21:30:15.875546948Z" level=info msg="Synchronizing NRI (plugin) with current runtime state"
Mar 20 21:30:15.875804 containerd[1505]: time="2025-03-20T21:30:15.875790024Z" level=info msg="containerd successfully booted in 0.103428s"
Mar 20 21:30:15.875892 systemd[1]: Started containerd.service - containerd container runtime.
Mar 20 21:30:15.973744 tar[1503]: linux-amd64/LICENSE
Mar 20 21:30:15.973847 tar[1503]: linux-amd64/README.md
Mar 20 21:30:15.999717 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin.
Mar 20 21:30:16.851802 systemd-networkd[1424]: eth0: Gained IPv6LL
Mar 20 21:30:16.855026 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured.
Mar 20 21:30:16.857855 systemd[1]: Reached target network-online.target - Network is Online.
Mar 20 21:30:16.860818 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent...
Mar 20 21:30:16.863122 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent...
Mar 20 21:30:16.882056 systemd[1]: Starting nvidia.service - NVIDIA Configure Service...
Mar 20 21:30:16.902878 systemd[1]: coreos-metadata.service: Deactivated successfully.
Mar 20 21:30:16.903135 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent.
Mar 20 21:30:16.904842 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met.
Mar 20 21:30:16.905295 systemd[1]: Finished nvidia.service - NVIDIA Configure Service.
Mar 20 21:30:17.473754 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent.
Mar 20 21:30:17.475341 systemd[1]: Reached target multi-user.target - Multi-User System.
Mar 20 21:30:17.476834 systemd[1]: Startup finished in 674ms (kernel) + 6.168s (initrd) + 4.655s (userspace) = 11.498s.
Mar 20 21:30:17.505936 (kubelet)[1605]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS
Mar 20 21:30:17.886215 kubelet[1605]: E0320 21:30:17.886038    1605 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory"
Mar 20 21:30:17.890221 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE
Mar 20 21:30:17.890404 systemd[1]: kubelet.service: Failed with result 'exit-code'.
Mar 20 21:30:17.890804 systemd[1]: kubelet.service: Consumed 901ms CPU time, 237.5M memory peak.
Mar 20 21:30:19.570513 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd.
Mar 20 21:30:19.571929 systemd[1]: Started sshd@0-10.0.0.132:22-10.0.0.1:47596.service - OpenSSH per-connection server daemon (10.0.0.1:47596).
Mar 20 21:30:19.639226 sshd[1618]: Accepted publickey for core from 10.0.0.1 port 47596 ssh2: RSA SHA256:KJ7ck8imsv1/sWVS7eR1M7V7NSskkAYjKibngyOtAC0
Mar 20 21:30:19.641360 sshd-session[1618]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Mar 20 21:30:19.647829 systemd[1]: Created slice user-500.slice - User Slice of UID 500.
Mar 20 21:30:19.648976 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500...
Mar 20 21:30:19.654967 systemd-logind[1487]: New session 1 of user core.
Mar 20 21:30:19.677494 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500.
Mar 20 21:30:19.680521 systemd[1]: Starting user@500.service - User Manager for UID 500...
Mar 20 21:30:19.693963 (systemd)[1622]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0)
Mar 20 21:30:19.696333 systemd-logind[1487]: New session c1 of user core.
Mar 20 21:30:19.845461 systemd[1622]: Queued start job for default target default.target.
Mar 20 21:30:19.854904 systemd[1622]: Created slice app.slice - User Application Slice.
Mar 20 21:30:19.854929 systemd[1622]: Reached target paths.target - Paths.
Mar 20 21:30:19.854971 systemd[1622]: Reached target timers.target - Timers.
Mar 20 21:30:19.856563 systemd[1622]: Starting dbus.socket - D-Bus User Message Bus Socket...
Mar 20 21:30:19.867410 systemd[1622]: Listening on dbus.socket - D-Bus User Message Bus Socket.
Mar 20 21:30:19.867550 systemd[1622]: Reached target sockets.target - Sockets.
Mar 20 21:30:19.867592 systemd[1622]: Reached target basic.target - Basic System.
Mar 20 21:30:19.867659 systemd[1622]: Reached target default.target - Main User Target.
Mar 20 21:30:19.867694 systemd[1622]: Startup finished in 163ms.
Mar 20 21:30:19.867992 systemd[1]: Started user@500.service - User Manager for UID 500.
Mar 20 21:30:19.869710 systemd[1]: Started session-1.scope - Session 1 of User core.
Mar 20 21:30:19.935986 systemd[1]: Started sshd@1-10.0.0.132:22-10.0.0.1:47598.service - OpenSSH per-connection server daemon (10.0.0.1:47598).
Mar 20 21:30:19.983224 sshd[1633]: Accepted publickey for core from 10.0.0.1 port 47598 ssh2: RSA SHA256:KJ7ck8imsv1/sWVS7eR1M7V7NSskkAYjKibngyOtAC0
Mar 20 21:30:19.984538 sshd-session[1633]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Mar 20 21:30:19.988430 systemd-logind[1487]: New session 2 of user core.
Mar 20 21:30:19.997734 systemd[1]: Started session-2.scope - Session 2 of User core.
Mar 20 21:30:20.051112 sshd[1635]: Connection closed by 10.0.0.1 port 47598
Mar 20 21:30:20.051479 sshd-session[1633]: pam_unix(sshd:session): session closed for user core
Mar 20 21:30:20.066551 systemd[1]: sshd@1-10.0.0.132:22-10.0.0.1:47598.service: Deactivated successfully.
Mar 20 21:30:20.068442 systemd[1]: session-2.scope: Deactivated successfully.
Mar 20 21:30:20.069858 systemd-logind[1487]: Session 2 logged out. Waiting for processes to exit.
Mar 20 21:30:20.071153 systemd[1]: Started sshd@2-10.0.0.132:22-10.0.0.1:47610.service - OpenSSH per-connection server daemon (10.0.0.1:47610).
Mar 20 21:30:20.071833 systemd-logind[1487]: Removed session 2.
Mar 20 21:30:20.123506 sshd[1640]: Accepted publickey for core from 10.0.0.1 port 47610 ssh2: RSA SHA256:KJ7ck8imsv1/sWVS7eR1M7V7NSskkAYjKibngyOtAC0
Mar 20 21:30:20.124925 sshd-session[1640]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Mar 20 21:30:20.129562 systemd-logind[1487]: New session 3 of user core.
Mar 20 21:30:20.138751 systemd[1]: Started session-3.scope - Session 3 of User core.
Mar 20 21:30:20.189081 sshd[1643]: Connection closed by 10.0.0.1 port 47610
Mar 20 21:30:20.189480 sshd-session[1640]: pam_unix(sshd:session): session closed for user core
Mar 20 21:30:20.197149 systemd[1]: sshd@2-10.0.0.132:22-10.0.0.1:47610.service: Deactivated successfully.
Mar 20 21:30:20.198921 systemd[1]: session-3.scope: Deactivated successfully.
Mar 20 21:30:20.200566 systemd-logind[1487]: Session 3 logged out. Waiting for processes to exit.
Mar 20 21:30:20.201904 systemd[1]: Started sshd@3-10.0.0.132:22-10.0.0.1:47618.service - OpenSSH per-connection server daemon (10.0.0.1:47618).
Mar 20 21:30:20.202790 systemd-logind[1487]: Removed session 3.
Mar 20 21:30:20.249990 sshd[1648]: Accepted publickey for core from 10.0.0.1 port 47618 ssh2: RSA SHA256:KJ7ck8imsv1/sWVS7eR1M7V7NSskkAYjKibngyOtAC0
Mar 20 21:30:20.251350 sshd-session[1648]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Mar 20 21:30:20.255982 systemd-logind[1487]: New session 4 of user core.
Mar 20 21:30:20.271832 systemd[1]: Started session-4.scope - Session 4 of User core.
Mar 20 21:30:20.325223 sshd[1651]: Connection closed by 10.0.0.1 port 47618
Mar 20 21:30:20.325590 sshd-session[1648]: pam_unix(sshd:session): session closed for user core
Mar 20 21:30:20.340648 systemd[1]: sshd@3-10.0.0.132:22-10.0.0.1:47618.service: Deactivated successfully.
Mar 20 21:30:20.342591 systemd[1]: session-4.scope: Deactivated successfully.
Mar 20 21:30:20.344281 systemd-logind[1487]: Session 4 logged out. Waiting for processes to exit.
Mar 20 21:30:20.345417 systemd[1]: Started sshd@4-10.0.0.132:22-10.0.0.1:47626.service - OpenSSH per-connection server daemon (10.0.0.1:47626).
Mar 20 21:30:20.346682 systemd-logind[1487]: Removed session 4.
Mar 20 21:30:20.391569 sshd[1656]: Accepted publickey for core from 10.0.0.1 port 47626 ssh2: RSA SHA256:KJ7ck8imsv1/sWVS7eR1M7V7NSskkAYjKibngyOtAC0
Mar 20 21:30:20.393203 sshd-session[1656]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Mar 20 21:30:20.397558 systemd-logind[1487]: New session 5 of user core.
Mar 20 21:30:20.411811 systemd[1]: Started session-5.scope - Session 5 of User core.
Mar 20 21:30:20.471543 sudo[1660]:     core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1
Mar 20 21:30:20.472022 sudo[1660]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500)
Mar 20 21:30:20.492201 sudo[1660]: pam_unix(sudo:session): session closed for user root
Mar 20 21:30:20.493750 sshd[1659]: Connection closed by 10.0.0.1 port 47626
Mar 20 21:30:20.494074 sshd-session[1656]: pam_unix(sshd:session): session closed for user core
Mar 20 21:30:20.504531 systemd[1]: sshd@4-10.0.0.132:22-10.0.0.1:47626.service: Deactivated successfully.
Mar 20 21:30:20.506509 systemd[1]: session-5.scope: Deactivated successfully.
Mar 20 21:30:20.508200 systemd-logind[1487]: Session 5 logged out. Waiting for processes to exit.
Mar 20 21:30:20.509854 systemd[1]: Started sshd@5-10.0.0.132:22-10.0.0.1:47634.service - OpenSSH per-connection server daemon (10.0.0.1:47634).
Mar 20 21:30:20.510715 systemd-logind[1487]: Removed session 5.
Mar 20 21:30:20.562956 sshd[1665]: Accepted publickey for core from 10.0.0.1 port 47634 ssh2: RSA SHA256:KJ7ck8imsv1/sWVS7eR1M7V7NSskkAYjKibngyOtAC0
Mar 20 21:30:20.564265 sshd-session[1665]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Mar 20 21:30:20.568408 systemd-logind[1487]: New session 6 of user core.
Mar 20 21:30:20.577741 systemd[1]: Started session-6.scope - Session 6 of User core.
Mar 20 21:30:20.631191 sudo[1670]:     core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules
Mar 20 21:30:20.631553 sudo[1670]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500)
Mar 20 21:30:20.636123 sudo[1670]: pam_unix(sudo:session): session closed for user root
Mar 20 21:30:20.642819 sudo[1669]:     core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules
Mar 20 21:30:20.643156 sudo[1669]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500)
Mar 20 21:30:20.654363 systemd[1]: Starting audit-rules.service - Load Audit Rules...
Mar 20 21:30:20.695167 augenrules[1692]: No rules
Mar 20 21:30:20.696883 systemd[1]: audit-rules.service: Deactivated successfully.
Mar 20 21:30:20.697168 systemd[1]: Finished audit-rules.service - Load Audit Rules.
Mar 20 21:30:20.698397 sudo[1669]: pam_unix(sudo:session): session closed for user root
Mar 20 21:30:20.699816 sshd[1668]: Connection closed by 10.0.0.1 port 47634
Mar 20 21:30:20.700106 sshd-session[1665]: pam_unix(sshd:session): session closed for user core
Mar 20 21:30:20.715034 systemd[1]: sshd@5-10.0.0.132:22-10.0.0.1:47634.service: Deactivated successfully.
Mar 20 21:30:20.716782 systemd[1]: session-6.scope: Deactivated successfully.
Mar 20 21:30:20.718107 systemd-logind[1487]: Session 6 logged out. Waiting for processes to exit.
Mar 20 21:30:20.719380 systemd[1]: Started sshd@6-10.0.0.132:22-10.0.0.1:47644.service - OpenSSH per-connection server daemon (10.0.0.1:47644).
Mar 20 21:30:20.720189 systemd-logind[1487]: Removed session 6.
Mar 20 21:30:20.764772 sshd[1700]: Accepted publickey for core from 10.0.0.1 port 47644 ssh2: RSA SHA256:KJ7ck8imsv1/sWVS7eR1M7V7NSskkAYjKibngyOtAC0
Mar 20 21:30:20.765994 sshd-session[1700]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Mar 20 21:30:20.770030 systemd-logind[1487]: New session 7 of user core.
Mar 20 21:30:20.780728 systemd[1]: Started session-7.scope - Session 7 of User core.
Mar 20 21:30:20.833718 sudo[1704]:     core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh
Mar 20 21:30:20.834039 sudo[1704]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500)
Mar 20 21:30:21.276209 systemd[1]: Starting docker.service - Docker Application Container Engine...
Mar 20 21:30:21.289981 (dockerd)[1725]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU
Mar 20 21:30:21.529059 dockerd[1725]: time="2025-03-20T21:30:21.528934032Z" level=info msg="Starting up"
Mar 20 21:30:21.530219 dockerd[1725]: time="2025-03-20T21:30:21.530196601Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider"
Mar 20 21:30:21.905986 dockerd[1725]: time="2025-03-20T21:30:21.905887963Z" level=info msg="Loading containers: start."
Mar 20 21:30:22.077638 kernel: Initializing XFRM netlink socket
Mar 20 21:30:22.167913 systemd-networkd[1424]: docker0: Link UP
Mar 20 21:30:22.327125 dockerd[1725]: time="2025-03-20T21:30:22.327062632Z" level=info msg="Loading containers: done."
Mar 20 21:30:22.340693 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck2761680489-merged.mount: Deactivated successfully.
Mar 20 21:30:22.341894 dockerd[1725]: time="2025-03-20T21:30:22.341856011Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2
Mar 20 21:30:22.341963 dockerd[1725]: time="2025-03-20T21:30:22.341933067Z" level=info msg="Docker daemon" commit=c710b88579fcb5e0d53f96dcae976d79323b9166 containerd-snapshotter=false storage-driver=overlay2 version=27.4.1
Mar 20 21:30:22.342063 dockerd[1725]: time="2025-03-20T21:30:22.342043611Z" level=info msg="Daemon has completed initialization"
Mar 20 21:30:22.378244 systemd[1]: Started docker.service - Docker Application Container Engine.
Mar 20 21:30:22.378399 dockerd[1725]: time="2025-03-20T21:30:22.378292750Z" level=info msg="API listen on /run/docker.sock"
Mar 20 21:30:23.076360 containerd[1505]: time="2025-03-20T21:30:23.076320783Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.7\""
Mar 20 21:30:23.640149 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2119154195.mount: Deactivated successfully.
Mar 20 21:30:24.460412 containerd[1505]: time="2025-03-20T21:30:24.460330009Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.31.7\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Mar 20 21:30:24.478312 containerd[1505]: time="2025-03-20T21:30:24.478201101Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.31.7: active requests=0, bytes read=27959268"
Mar 20 21:30:24.506696 containerd[1505]: time="2025-03-20T21:30:24.506632173Z" level=info msg="ImageCreate event name:\"sha256:f084bc047a8cf7c8484d47c51e70e646dde3977d916f282feb99207b7b9241af\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Mar 20 21:30:24.526496 containerd[1505]: time="2025-03-20T21:30:24.526441824Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:22c19cc70fe5806d0a2cb28a6b6b33fd34e6f9e50616bdf6d53649bcfafbc277\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Mar 20 21:30:24.527368 containerd[1505]: time="2025-03-20T21:30:24.527335488Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.31.7\" with image id \"sha256:f084bc047a8cf7c8484d47c51e70e646dde3977d916f282feb99207b7b9241af\", repo tag \"registry.k8s.io/kube-apiserver:v1.31.7\", repo digest \"registry.k8s.io/kube-apiserver@sha256:22c19cc70fe5806d0a2cb28a6b6b33fd34e6f9e50616bdf6d53649bcfafbc277\", size \"27956068\" in 1.450980453s"
Mar 20 21:30:24.527410 containerd[1505]: time="2025-03-20T21:30:24.527372019Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.7\" returns image reference \"sha256:f084bc047a8cf7c8484d47c51e70e646dde3977d916f282feb99207b7b9241af\""
Mar 20 21:30:24.528742 containerd[1505]: time="2025-03-20T21:30:24.528688396Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.7\""
Mar 20 21:30:25.691532 containerd[1505]: time="2025-03-20T21:30:25.691485637Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.31.7\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Mar 20 21:30:25.692559 containerd[1505]: time="2025-03-20T21:30:25.692283582Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.31.7: active requests=0, bytes read=24713776"
Mar 20 21:30:25.693919 containerd[1505]: time="2025-03-20T21:30:25.693877681Z" level=info msg="ImageCreate event name:\"sha256:652dcad615a9a0c252c253860d5b5b7bfebd3efe159dc033a8555bc15a6d1985\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Mar 20 21:30:25.695986 containerd[1505]: time="2025-03-20T21:30:25.695954950Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:6abe7a0accecf29db6ebab18a10f844678ffed693d79e2e51a18a6f2b4530cbb\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Mar 20 21:30:25.696848 containerd[1505]: time="2025-03-20T21:30:25.696819410Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.31.7\" with image id \"sha256:652dcad615a9a0c252c253860d5b5b7bfebd3efe159dc033a8555bc15a6d1985\", repo tag \"registry.k8s.io/kube-controller-manager:v1.31.7\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:6abe7a0accecf29db6ebab18a10f844678ffed693d79e2e51a18a6f2b4530cbb\", size \"26201384\" in 1.168088304s"
Mar 20 21:30:25.696885 containerd[1505]: time="2025-03-20T21:30:25.696849648Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.7\" returns image reference \"sha256:652dcad615a9a0c252c253860d5b5b7bfebd3efe159dc033a8555bc15a6d1985\""
Mar 20 21:30:25.697571 containerd[1505]: time="2025-03-20T21:30:25.697526265Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.7\""
Mar 20 21:30:26.976812 containerd[1505]: time="2025-03-20T21:30:26.976747395Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.31.7\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Mar 20 21:30:26.977513 containerd[1505]: time="2025-03-20T21:30:26.977445470Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.31.7: active requests=0, bytes read=18780368"
Mar 20 21:30:26.978825 containerd[1505]: time="2025-03-20T21:30:26.978765853Z" level=info msg="ImageCreate event name:\"sha256:7f1f6a63d8aa14cf61d0045e912ad312b4ade24637cecccc933b163582eae68c\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Mar 20 21:30:26.981185 containerd[1505]: time="2025-03-20T21:30:26.981148127Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:fb80249bcb77ee72b1c9fa5b70bc28a83ed107c9ca71957841ad91db379963bf\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Mar 20 21:30:26.982086 containerd[1505]: time="2025-03-20T21:30:26.982054155Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.31.7\" with image id \"sha256:7f1f6a63d8aa14cf61d0045e912ad312b4ade24637cecccc933b163582eae68c\", repo tag \"registry.k8s.io/kube-scheduler:v1.31.7\", repo digest \"registry.k8s.io/kube-scheduler@sha256:fb80249bcb77ee72b1c9fa5b70bc28a83ed107c9ca71957841ad91db379963bf\", size \"20267994\" in 1.284479865s"
Mar 20 21:30:26.982086 containerd[1505]: time="2025-03-20T21:30:26.982085594Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.7\" returns image reference \"sha256:7f1f6a63d8aa14cf61d0045e912ad312b4ade24637cecccc933b163582eae68c\""
Mar 20 21:30:26.982550 containerd[1505]: time="2025-03-20T21:30:26.982512109Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.7\""
Mar 20 21:30:28.027715 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount586150739.mount: Deactivated successfully.
Mar 20 21:30:28.028876 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1.
Mar 20 21:30:28.030283 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent...
Mar 20 21:30:28.198060 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent.
Mar 20 21:30:28.206915 (kubelet)[2010]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS
Mar 20 21:30:28.246601 kubelet[2010]: E0320 21:30:28.246507    2010 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory"
Mar 20 21:30:28.253440 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE
Mar 20 21:30:28.253770 systemd[1]: kubelet.service: Failed with result 'exit-code'.
Mar 20 21:30:28.254112 systemd[1]: kubelet.service: Consumed 205ms CPU time, 97.5M memory peak.
Mar 20 21:30:28.922135 containerd[1505]: time="2025-03-20T21:30:28.922073867Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.31.7\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Mar 20 21:30:28.922797 containerd[1505]: time="2025-03-20T21:30:28.922751530Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.31.7: active requests=0, bytes read=30354630"
Mar 20 21:30:28.923949 containerd[1505]: time="2025-03-20T21:30:28.923924275Z" level=info msg="ImageCreate event name:\"sha256:dcfc039c372ea285997a302d60e58a75b80905b4c4dba969993b9b22e8ac66d1\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Mar 20 21:30:28.925749 containerd[1505]: time="2025-03-20T21:30:28.925716991Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:e5839270c96c3ad1bea1dce4935126d3281297527f3655408d2970aa4b5cf178\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Mar 20 21:30:28.926216 containerd[1505]: time="2025-03-20T21:30:28.926188937Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.31.7\" with image id \"sha256:dcfc039c372ea285997a302d60e58a75b80905b4c4dba969993b9b22e8ac66d1\", repo tag \"registry.k8s.io/kube-proxy:v1.31.7\", repo digest \"registry.k8s.io/kube-proxy@sha256:e5839270c96c3ad1bea1dce4935126d3281297527f3655408d2970aa4b5cf178\", size \"30353649\" in 1.943645348s"
Mar 20 21:30:28.926244 containerd[1505]: time="2025-03-20T21:30:28.926215614Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.7\" returns image reference \"sha256:dcfc039c372ea285997a302d60e58a75b80905b4c4dba969993b9b22e8ac66d1\""
Mar 20 21:30:28.926695 containerd[1505]: time="2025-03-20T21:30:28.926673868Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\""
Mar 20 21:30:29.435596 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1692368764.mount: Deactivated successfully.
Mar 20 21:30:30.080098 containerd[1505]: time="2025-03-20T21:30:30.080040476Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.1\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Mar 20 21:30:30.080782 containerd[1505]: time="2025-03-20T21:30:30.080705704Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.1: active requests=0, bytes read=18185761"
Mar 20 21:30:30.081949 containerd[1505]: time="2025-03-20T21:30:30.081918427Z" level=info msg="ImageCreate event name:\"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Mar 20 21:30:30.084369 containerd[1505]: time="2025-03-20T21:30:30.084320500Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Mar 20 21:30:30.085176 containerd[1505]: time="2025-03-20T21:30:30.085151495Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.1\" with image id \"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\", size \"18182961\" in 1.158453575s"
Mar 20 21:30:30.085223 containerd[1505]: time="2025-03-20T21:30:30.085178001Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\" returns image reference \"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\""
Mar 20 21:30:30.085806 containerd[1505]: time="2025-03-20T21:30:30.085780274Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\""
Mar 20 21:30:30.570757 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3913767948.mount: Deactivated successfully.
Mar 20 21:30:30.575434 containerd[1505]: time="2025-03-20T21:30:30.575394132Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}  labels:{key:\"io.cri-containerd.pinned\"  value:\"pinned\"}"
Mar 20 21:30:30.576097 containerd[1505]: time="2025-03-20T21:30:30.576016052Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138"
Mar 20 21:30:30.577149 containerd[1505]: time="2025-03-20T21:30:30.577110281Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}  labels:{key:\"io.cri-containerd.pinned\"  value:\"pinned\"}"
Mar 20 21:30:30.578915 containerd[1505]: time="2025-03-20T21:30:30.578886979Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}  labels:{key:\"io.cri-containerd.pinned\"  value:\"pinned\"}"
Mar 20 21:30:30.579478 containerd[1505]: time="2025-03-20T21:30:30.579450836Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 493.640911ms"
Mar 20 21:30:30.579510 containerd[1505]: time="2025-03-20T21:30:30.579476443Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\""
Mar 20 21:30:30.579917 containerd[1505]: time="2025-03-20T21:30:30.579892815Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\""
Mar 20 21:30:31.087256 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4111941095.mount: Deactivated successfully.
Mar 20 21:30:33.172419 containerd[1505]: time="2025-03-20T21:30:33.172348435Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.15-0\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Mar 20 21:30:33.173431 containerd[1505]: time="2025-03-20T21:30:33.173341561Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.15-0: active requests=0, bytes read=56779973"
Mar 20 21:30:33.174379 containerd[1505]: time="2025-03-20T21:30:33.174344912Z" level=info msg="ImageCreate event name:\"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Mar 20 21:30:33.176841 containerd[1505]: time="2025-03-20T21:30:33.176794488Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Mar 20 21:30:33.177892 containerd[1505]: time="2025-03-20T21:30:33.177863687Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.15-0\" with image id \"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\", repo tag \"registry.k8s.io/etcd:3.5.15-0\", repo digest \"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\", size \"56909194\" in 2.597944089s"
Mar 20 21:30:33.177955 containerd[1505]: time="2025-03-20T21:30:33.177892673Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\" returns image reference \"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\""
Mar 20 21:30:35.481039 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent.
Mar 20 21:30:35.481208 systemd[1]: kubelet.service: Consumed 205ms CPU time, 97.5M memory peak.
Mar 20 21:30:35.483307 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent...
Mar 20 21:30:35.506666 systemd[1]: Reload requested from client PID 2151 ('systemctl') (unit session-7.scope)...
Mar 20 21:30:35.506681 systemd[1]: Reloading...
Mar 20 21:30:35.589644 zram_generator::config[2197]: No configuration found.
Mar 20 21:30:35.792913 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly.
Mar 20 21:30:35.892130 systemd[1]: Reloading finished in 385 ms.
Mar 20 21:30:35.961927 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent.
Mar 20 21:30:35.965512 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent...
Mar 20 21:30:35.966048 systemd[1]: kubelet.service: Deactivated successfully.
Mar 20 21:30:35.966328 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent.
Mar 20 21:30:35.966361 systemd[1]: kubelet.service: Consumed 139ms CPU time, 83.6M memory peak.
Mar 20 21:30:35.967942 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent...
Mar 20 21:30:36.123310 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent.
Mar 20 21:30:36.138925 (kubelet)[2244]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS
Mar 20 21:30:36.171186 kubelet[2244]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information.
Mar 20 21:30:36.171186 kubelet[2244]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI.
Mar 20 21:30:36.171186 kubelet[2244]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information.
Mar 20 21:30:36.172181 kubelet[2244]: I0320 21:30:36.172139    2244 server.go:206] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime"
Mar 20 21:30:36.500669 kubelet[2244]: I0320 21:30:36.500558    2244 server.go:486] "Kubelet version" kubeletVersion="v1.31.0"
Mar 20 21:30:36.500669 kubelet[2244]: I0320 21:30:36.500589    2244 server.go:488] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK=""
Mar 20 21:30:36.500888 kubelet[2244]: I0320 21:30:36.500874    2244 server.go:929] "Client rotation is on, will bootstrap in background"
Mar 20 21:30:36.522961 kubelet[2244]: I0320 21:30:36.522929    2244 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt"
Mar 20 21:30:36.522961 kubelet[2244]: E0320 21:30:36.522934    2244 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.0.0.132:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.132:6443: connect: connection refused" logger="UnhandledError"
Mar 20 21:30:36.532744 kubelet[2244]: I0320 21:30:36.532717    2244 server.go:1426] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd"
Mar 20 21:30:36.538706 kubelet[2244]: I0320 21:30:36.538675    2244 server.go:744] "--cgroups-per-qos enabled, but --cgroup-root was not specified.  defaulting to /"
Mar 20 21:30:36.539626 kubelet[2244]: I0320 21:30:36.539589    2244 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority"
Mar 20 21:30:36.539830 kubelet[2244]: I0320 21:30:36.539791    2244 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[]
Mar 20 21:30:36.539971 kubelet[2244]: I0320 21:30:36.539818    2244 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2}
Mar 20 21:30:36.539971 kubelet[2244]: I0320 21:30:36.539967    2244 topology_manager.go:138] "Creating topology manager with none policy"
Mar 20 21:30:36.540093 kubelet[2244]: I0320 21:30:36.539976    2244 container_manager_linux.go:300] "Creating device plugin manager"
Mar 20 21:30:36.540123 kubelet[2244]: I0320 21:30:36.540094    2244 state_mem.go:36] "Initialized new in-memory state store"
Mar 20 21:30:36.541553 kubelet[2244]: I0320 21:30:36.541518    2244 kubelet.go:408] "Attempting to sync node with API server"
Mar 20 21:30:36.541553 kubelet[2244]: I0320 21:30:36.541541    2244 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests"
Mar 20 21:30:36.541818 kubelet[2244]: I0320 21:30:36.541576    2244 kubelet.go:314] "Adding apiserver pod source"
Mar 20 21:30:36.541818 kubelet[2244]: I0320 21:30:36.541634    2244 apiserver.go:42] "Waiting for node sync before watching apiserver pods"
Mar 20 21:30:36.546364 kubelet[2244]: W0320 21:30:36.546318    2244 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.132:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.0.0.132:6443: connect: connection refused
Mar 20 21:30:36.546480 kubelet[2244]: E0320 21:30:36.546449    2244 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.0.0.132:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.132:6443: connect: connection refused" logger="UnhandledError"
Mar 20 21:30:36.547294 kubelet[2244]: I0320 21:30:36.547281    2244 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v2.0.1" apiVersion="v1"
Mar 20 21:30:36.548345 kubelet[2244]: W0320 21:30:36.548301    2244 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.132:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.132:6443: connect: connection refused
Mar 20 21:30:36.548404 kubelet[2244]: E0320 21:30:36.548362    2244 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.0.0.132:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.132:6443: connect: connection refused" logger="UnhandledError"
Mar 20 21:30:36.549024 kubelet[2244]: I0320 21:30:36.549008    2244 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode"
Mar 20 21:30:36.549934 kubelet[2244]: W0320 21:30:36.549909    2244 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating.
Mar 20 21:30:36.550795 kubelet[2244]: I0320 21:30:36.550652    2244 server.go:1269] "Started kubelet"
Mar 20 21:30:36.551518 kubelet[2244]: I0320 21:30:36.550903    2244 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10
Mar 20 21:30:36.551518 kubelet[2244]: I0320 21:30:36.551226    2244 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock"
Mar 20 21:30:36.551518 kubelet[2244]: I0320 21:30:36.551281    2244 server.go:163] "Starting to listen" address="0.0.0.0" port=10250
Mar 20 21:30:36.552249 kubelet[2244]: I0320 21:30:36.552124    2244 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer"
Mar 20 21:30:36.553309 kubelet[2244]: I0320 21:30:36.552534    2244 server.go:460] "Adding debug handlers to kubelet server"
Mar 20 21:30:36.553309 kubelet[2244]: I0320 21:30:36.553167    2244 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key"
Mar 20 21:30:36.554046 kubelet[2244]: E0320 21:30:36.553773    2244 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found"
Mar 20 21:30:36.554046 kubelet[2244]: I0320 21:30:36.553817    2244 volume_manager.go:289] "Starting Kubelet Volume Manager"
Mar 20 21:30:36.554046 kubelet[2244]: I0320 21:30:36.553969    2244 desired_state_of_world_populator.go:146] "Desired state populator starts to run"
Mar 20 21:30:36.554046 kubelet[2244]: I0320 21:30:36.554037    2244 reconciler.go:26] "Reconciler: start to sync state"
Mar 20 21:30:36.555018 kubelet[2244]: I0320 21:30:36.554698    2244 factory.go:221] Registration of the systemd container factory successfully
Mar 20 21:30:36.555018 kubelet[2244]: I0320 21:30:36.554766    2244 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory
Mar 20 21:30:36.555120 kubelet[2244]: E0320 21:30:36.555072    2244 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.132:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.132:6443: connect: connection refused" interval="200ms"
Mar 20 21:30:36.555189 kubelet[2244]: W0320 21:30:36.555155    2244 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.132:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.132:6443: connect: connection refused
Mar 20 21:30:36.555216 kubelet[2244]: E0320 21:30:36.555197    2244 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.0.0.132:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.132:6443: connect: connection refused" logger="UnhandledError"
Mar 20 21:30:36.555408 kubelet[2244]: E0320 21:30:36.553179    2244 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.132:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.132:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.182ea030a07fb8bd  default    0 0001-01-01 00:00:00 +0000 UTC <nil> <nil> map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-03-20 21:30:36.550559933 +0000 UTC m=+0.408140778,LastTimestamp:2025-03-20 21:30:36.550559933 +0000 UTC m=+0.408140778,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}"
Mar 20 21:30:36.555602 kubelet[2244]: E0320 21:30:36.555583    2244 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem"
Mar 20 21:30:36.556366 kubelet[2244]: I0320 21:30:36.556322    2244 factory.go:221] Registration of the containerd container factory successfully
Mar 20 21:30:36.569392 kubelet[2244]: I0320 21:30:36.569355    2244 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4"
Mar 20 21:30:36.570984 kubelet[2244]: I0320 21:30:36.570680    2244 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6"
Mar 20 21:30:36.570984 kubelet[2244]: I0320 21:30:36.570808    2244 status_manager.go:217] "Starting to sync pod status with apiserver"
Mar 20 21:30:36.570984 kubelet[2244]: I0320 21:30:36.570829    2244 kubelet.go:2321] "Starting kubelet main sync loop"
Mar 20 21:30:36.570984 kubelet[2244]: E0320 21:30:36.570865    2244 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]"
Mar 20 21:30:36.571258 kubelet[2244]: I0320 21:30:36.571236    2244 cpu_manager.go:214] "Starting CPU manager" policy="none"
Mar 20 21:30:36.571258 kubelet[2244]: I0320 21:30:36.571254    2244 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s"
Mar 20 21:30:36.571312 kubelet[2244]: I0320 21:30:36.571271    2244 state_mem.go:36] "Initialized new in-memory state store"
Mar 20 21:30:36.571881 kubelet[2244]: W0320 21:30:36.571385    2244 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.132:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.132:6443: connect: connection refused
Mar 20 21:30:36.571881 kubelet[2244]: E0320 21:30:36.571416    2244 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.0.0.132:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.132:6443: connect: connection refused" logger="UnhandledError"
Mar 20 21:30:36.654600 kubelet[2244]: E0320 21:30:36.654569    2244 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found"
Mar 20 21:30:36.671812 kubelet[2244]: E0320 21:30:36.671779    2244 kubelet.go:2345] "Skipping pod synchronization" err="container runtime status check may not have completed yet"
Mar 20 21:30:36.755047 kubelet[2244]: E0320 21:30:36.754893    2244 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found"
Mar 20 21:30:36.756374 kubelet[2244]: E0320 21:30:36.756337    2244 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.132:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.132:6443: connect: connection refused" interval="400ms"
Mar 20 21:30:36.845544 kubelet[2244]: I0320 21:30:36.845495    2244 policy_none.go:49] "None policy: Start"
Mar 20 21:30:36.846363 kubelet[2244]: I0320 21:30:36.846332    2244 memory_manager.go:170] "Starting memorymanager" policy="None"
Mar 20 21:30:36.846363 kubelet[2244]: I0320 21:30:36.846357    2244 state_mem.go:35] "Initializing new in-memory state store"
Mar 20 21:30:36.854450 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice.
Mar 20 21:30:36.854980 kubelet[2244]: E0320 21:30:36.854953    2244 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found"
Mar 20 21:30:36.871988 kubelet[2244]: E0320 21:30:36.871949    2244 kubelet.go:2345] "Skipping pod synchronization" err="container runtime status check may not have completed yet"
Mar 20 21:30:36.874150 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice.
Mar 20 21:30:36.877439 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice.
Mar 20 21:30:36.888491 kubelet[2244]: I0320 21:30:36.888454    2244 manager.go:510] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found"
Mar 20 21:30:36.888711 kubelet[2244]: I0320 21:30:36.888683    2244 eviction_manager.go:189] "Eviction manager: starting control loop"
Mar 20 21:30:36.888776 kubelet[2244]: I0320 21:30:36.888698    2244 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s"
Mar 20 21:30:36.888997 kubelet[2244]: I0320 21:30:36.888931    2244 plugin_manager.go:118] "Starting Kubelet Plugin Manager"
Mar 20 21:30:36.889968 kubelet[2244]: E0320 21:30:36.889925    2244 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found"
Mar 20 21:30:36.990876 kubelet[2244]: I0320 21:30:36.990838    2244 kubelet_node_status.go:72] "Attempting to register node" node="localhost"
Mar 20 21:30:36.991178 kubelet[2244]: E0320 21:30:36.991153    2244 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.0.0.132:6443/api/v1/nodes\": dial tcp 10.0.0.132:6443: connect: connection refused" node="localhost"
Mar 20 21:30:37.157038 kubelet[2244]: E0320 21:30:37.156893    2244 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.132:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.132:6443: connect: connection refused" interval="800ms"
Mar 20 21:30:37.192922 kubelet[2244]: I0320 21:30:37.192888    2244 kubelet_node_status.go:72] "Attempting to register node" node="localhost"
Mar 20 21:30:37.193321 kubelet[2244]: E0320 21:30:37.193236    2244 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.0.0.132:6443/api/v1/nodes\": dial tcp 10.0.0.132:6443: connect: connection refused" node="localhost"
Mar 20 21:30:37.280882 systemd[1]: Created slice kubepods-burstable-poda2d7033018db514077a55f9bbb204c5f.slice - libcontainer container kubepods-burstable-poda2d7033018db514077a55f9bbb204c5f.slice.
Mar 20 21:30:37.304034 systemd[1]: Created slice kubepods-burstable-pod60762308083b5ef6c837b1be48ec53d6.slice - libcontainer container kubepods-burstable-pod60762308083b5ef6c837b1be48ec53d6.slice.
Mar 20 21:30:37.307471 systemd[1]: Created slice kubepods-burstable-pod6f32907a07e55aea05abdc5cd284a8d5.slice - libcontainer container kubepods-burstable-pod6f32907a07e55aea05abdc5cd284a8d5.slice.
Mar 20 21:30:37.357184 kubelet[2244]: I0320 21:30:37.357111    2244 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/60762308083b5ef6c837b1be48ec53d6-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"60762308083b5ef6c837b1be48ec53d6\") " pod="kube-system/kube-controller-manager-localhost"
Mar 20 21:30:37.357184 kubelet[2244]: I0320 21:30:37.357175    2244 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/60762308083b5ef6c837b1be48ec53d6-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"60762308083b5ef6c837b1be48ec53d6\") " pod="kube-system/kube-controller-manager-localhost"
Mar 20 21:30:37.357326 kubelet[2244]: I0320 21:30:37.357199    2244 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/60762308083b5ef6c837b1be48ec53d6-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"60762308083b5ef6c837b1be48ec53d6\") " pod="kube-system/kube-controller-manager-localhost"
Mar 20 21:30:37.357326 kubelet[2244]: I0320 21:30:37.357251    2244 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/60762308083b5ef6c837b1be48ec53d6-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"60762308083b5ef6c837b1be48ec53d6\") " pod="kube-system/kube-controller-manager-localhost"
Mar 20 21:30:37.357326 kubelet[2244]: I0320 21:30:37.357281    2244 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/6f32907a07e55aea05abdc5cd284a8d5-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"6f32907a07e55aea05abdc5cd284a8d5\") " pod="kube-system/kube-scheduler-localhost"
Mar 20 21:30:37.357326 kubelet[2244]: I0320 21:30:37.357301    2244 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/a2d7033018db514077a55f9bbb204c5f-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"a2d7033018db514077a55f9bbb204c5f\") " pod="kube-system/kube-apiserver-localhost"
Mar 20 21:30:37.357453 kubelet[2244]: I0320 21:30:37.357340    2244 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/a2d7033018db514077a55f9bbb204c5f-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"a2d7033018db514077a55f9bbb204c5f\") " pod="kube-system/kube-apiserver-localhost"
Mar 20 21:30:37.357453 kubelet[2244]: I0320 21:30:37.357357    2244 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/a2d7033018db514077a55f9bbb204c5f-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"a2d7033018db514077a55f9bbb204c5f\") " pod="kube-system/kube-apiserver-localhost"
Mar 20 21:30:37.357453 kubelet[2244]: I0320 21:30:37.357371    2244 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/60762308083b5ef6c837b1be48ec53d6-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"60762308083b5ef6c837b1be48ec53d6\") " pod="kube-system/kube-controller-manager-localhost"
Mar 20 21:30:37.548478 kubelet[2244]: W0320 21:30:37.548339    2244 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.132:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.0.0.132:6443: connect: connection refused
Mar 20 21:30:37.548478 kubelet[2244]: E0320 21:30:37.548407    2244 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.0.0.132:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.132:6443: connect: connection refused" logger="UnhandledError"
Mar 20 21:30:37.595177 kubelet[2244]: I0320 21:30:37.595145    2244 kubelet_node_status.go:72] "Attempting to register node" node="localhost"
Mar 20 21:30:37.595395 kubelet[2244]: E0320 21:30:37.595367    2244 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.0.0.132:6443/api/v1/nodes\": dial tcp 10.0.0.132:6443: connect: connection refused" node="localhost"
Mar 20 21:30:37.601711 kubelet[2244]: W0320 21:30:37.601644    2244 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.132:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.132:6443: connect: connection refused
Mar 20 21:30:37.601711 kubelet[2244]: E0320 21:30:37.601699    2244 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.0.0.132:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.132:6443: connect: connection refused" logger="UnhandledError"
Mar 20 21:30:37.602217 containerd[1505]: time="2025-03-20T21:30:37.602171822Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:a2d7033018db514077a55f9bbb204c5f,Namespace:kube-system,Attempt:0,}"
Mar 20 21:30:37.606730 containerd[1505]: time="2025-03-20T21:30:37.606686081Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:60762308083b5ef6c837b1be48ec53d6,Namespace:kube-system,Attempt:0,}"
Mar 20 21:30:37.610282 containerd[1505]: time="2025-03-20T21:30:37.610228492Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:6f32907a07e55aea05abdc5cd284a8d5,Namespace:kube-system,Attempt:0,}"
Mar 20 21:30:37.811841 kubelet[2244]: W0320 21:30:37.811720    2244 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.132:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.132:6443: connect: connection refused
Mar 20 21:30:37.811841 kubelet[2244]: E0320 21:30:37.811798    2244 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.0.0.132:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.132:6443: connect: connection refused" logger="UnhandledError"
Mar 20 21:30:37.880423 containerd[1505]: time="2025-03-20T21:30:37.880366339Z" level=info msg="connecting to shim 1acbc96e90cc1fc081c44728e2937a4df9c0ad25a2e494b00e61d6cbfde5d2cb" address="unix:///run/containerd/s/10168f78ecb1f01a640ecfe27195a6c77483394cea9b1877ea545cd6da9acafa" namespace=k8s.io protocol=ttrpc version=3
Mar 20 21:30:37.881972 containerd[1505]: time="2025-03-20T21:30:37.881924184Z" level=info msg="connecting to shim 6c371fc3102dce8e21df3fe97f75d955412fe830dd5680d8cc9ad7ec374d09c3" address="unix:///run/containerd/s/899d5b378485a32fafd34a56977733f41875e57bc2c05492e0676e30ef48655d" namespace=k8s.io protocol=ttrpc version=3
Mar 20 21:30:37.890343 containerd[1505]: time="2025-03-20T21:30:37.889944296Z" level=info msg="connecting to shim 70aa7335aa6cc4f6e19ac7dcb6a714f2a195449a2b10e68f241a6c702909d8c7" address="unix:///run/containerd/s/4651a013f7322c344011014ca65cb734463e899a0524270a32c3722c4457c661" namespace=k8s.io protocol=ttrpc version=3
Mar 20 21:30:37.904867 systemd[1]: Started cri-containerd-6c371fc3102dce8e21df3fe97f75d955412fe830dd5680d8cc9ad7ec374d09c3.scope - libcontainer container 6c371fc3102dce8e21df3fe97f75d955412fe830dd5680d8cc9ad7ec374d09c3.
Mar 20 21:30:37.922759 systemd[1]: Started cri-containerd-70aa7335aa6cc4f6e19ac7dcb6a714f2a195449a2b10e68f241a6c702909d8c7.scope - libcontainer container 70aa7335aa6cc4f6e19ac7dcb6a714f2a195449a2b10e68f241a6c702909d8c7.
Mar 20 21:30:37.925998 systemd[1]: Started cri-containerd-1acbc96e90cc1fc081c44728e2937a4df9c0ad25a2e494b00e61d6cbfde5d2cb.scope - libcontainer container 1acbc96e90cc1fc081c44728e2937a4df9c0ad25a2e494b00e61d6cbfde5d2cb.
Mar 20 21:30:37.958284 kubelet[2244]: E0320 21:30:37.958163    2244 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.132:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.132:6443: connect: connection refused" interval="1.6s"
Mar 20 21:30:37.968066 containerd[1505]: time="2025-03-20T21:30:37.968016866Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:6f32907a07e55aea05abdc5cd284a8d5,Namespace:kube-system,Attempt:0,} returns sandbox id \"6c371fc3102dce8e21df3fe97f75d955412fe830dd5680d8cc9ad7ec374d09c3\""
Mar 20 21:30:37.970812 containerd[1505]: time="2025-03-20T21:30:37.970774001Z" level=info msg="CreateContainer within sandbox \"6c371fc3102dce8e21df3fe97f75d955412fe830dd5680d8cc9ad7ec374d09c3\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}"
Mar 20 21:30:37.971464 containerd[1505]: time="2025-03-20T21:30:37.971343862Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:a2d7033018db514077a55f9bbb204c5f,Namespace:kube-system,Attempt:0,} returns sandbox id \"1acbc96e90cc1fc081c44728e2937a4df9c0ad25a2e494b00e61d6cbfde5d2cb\""
Mar 20 21:30:37.973132 containerd[1505]: time="2025-03-20T21:30:37.973110610Z" level=info msg="CreateContainer within sandbox \"1acbc96e90cc1fc081c44728e2937a4df9c0ad25a2e494b00e61d6cbfde5d2cb\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}"
Mar 20 21:30:37.978690 containerd[1505]: time="2025-03-20T21:30:37.978671134Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:60762308083b5ef6c837b1be48ec53d6,Namespace:kube-system,Attempt:0,} returns sandbox id \"70aa7335aa6cc4f6e19ac7dcb6a714f2a195449a2b10e68f241a6c702909d8c7\""
Mar 20 21:30:37.980869 containerd[1505]: time="2025-03-20T21:30:37.980845132Z" level=info msg="CreateContainer within sandbox \"70aa7335aa6cc4f6e19ac7dcb6a714f2a195449a2b10e68f241a6c702909d8c7\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}"
Mar 20 21:30:37.983134 containerd[1505]: time="2025-03-20T21:30:37.983115709Z" level=info msg="Container ab69ff09a1e13c79148ae93d09b20041583947ea6c4b300c5031673061ca67e8: CDI devices from CRI Config.CDIDevices: []"
Mar 20 21:30:37.986867 containerd[1505]: time="2025-03-20T21:30:37.986763672Z" level=info msg="Container 9ebe67afe3f8fd6ed0c565746cb4d5fdfaf775cfc9a3a9701c0347757988c75a: CDI devices from CRI Config.CDIDevices: []"
Mar 20 21:30:37.991907 containerd[1505]: time="2025-03-20T21:30:37.991886640Z" level=info msg="CreateContainer within sandbox \"6c371fc3102dce8e21df3fe97f75d955412fe830dd5680d8cc9ad7ec374d09c3\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"ab69ff09a1e13c79148ae93d09b20041583947ea6c4b300c5031673061ca67e8\""
Mar 20 21:30:37.992455 containerd[1505]: time="2025-03-20T21:30:37.992432118Z" level=info msg="StartContainer for \"ab69ff09a1e13c79148ae93d09b20041583947ea6c4b300c5031673061ca67e8\""
Mar 20 21:30:37.993396 containerd[1505]: time="2025-03-20T21:30:37.993376232Z" level=info msg="connecting to shim ab69ff09a1e13c79148ae93d09b20041583947ea6c4b300c5031673061ca67e8" address="unix:///run/containerd/s/899d5b378485a32fafd34a56977733f41875e57bc2c05492e0676e30ef48655d" protocol=ttrpc version=3
Mar 20 21:30:37.997058 containerd[1505]: time="2025-03-20T21:30:37.997015201Z" level=info msg="Container b0b6a6ed7a2cfb80df91a87bd4df59d50ac762e42fd67c65b57e7b0071d97c26: CDI devices from CRI Config.CDIDevices: []"
Mar 20 21:30:38.000735 containerd[1505]: time="2025-03-20T21:30:38.000713426Z" level=info msg="CreateContainer within sandbox \"1acbc96e90cc1fc081c44728e2937a4df9c0ad25a2e494b00e61d6cbfde5d2cb\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"9ebe67afe3f8fd6ed0c565746cb4d5fdfaf775cfc9a3a9701c0347757988c75a\""
Mar 20 21:30:38.001045 containerd[1505]: time="2025-03-20T21:30:38.001027794Z" level=info msg="StartContainer for \"9ebe67afe3f8fd6ed0c565746cb4d5fdfaf775cfc9a3a9701c0347757988c75a\""
Mar 20 21:30:38.002509 containerd[1505]: time="2025-03-20T21:30:38.002140205Z" level=info msg="connecting to shim 9ebe67afe3f8fd6ed0c565746cb4d5fdfaf775cfc9a3a9701c0347757988c75a" address="unix:///run/containerd/s/10168f78ecb1f01a640ecfe27195a6c77483394cea9b1877ea545cd6da9acafa" protocol=ttrpc version=3
Mar 20 21:30:38.003379 containerd[1505]: time="2025-03-20T21:30:38.003332617Z" level=info msg="CreateContainer within sandbox \"70aa7335aa6cc4f6e19ac7dcb6a714f2a195449a2b10e68f241a6c702909d8c7\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"b0b6a6ed7a2cfb80df91a87bd4df59d50ac762e42fd67c65b57e7b0071d97c26\""
Mar 20 21:30:38.003765 containerd[1505]: time="2025-03-20T21:30:38.003723306Z" level=info msg="StartContainer for \"b0b6a6ed7a2cfb80df91a87bd4df59d50ac762e42fd67c65b57e7b0071d97c26\""
Mar 20 21:30:38.005793 containerd[1505]: time="2025-03-20T21:30:38.005764953Z" level=info msg="connecting to shim b0b6a6ed7a2cfb80df91a87bd4df59d50ac762e42fd67c65b57e7b0071d97c26" address="unix:///run/containerd/s/4651a013f7322c344011014ca65cb734463e899a0524270a32c3722c4457c661" protocol=ttrpc version=3
Mar 20 21:30:38.010778 systemd[1]: Started cri-containerd-ab69ff09a1e13c79148ae93d09b20041583947ea6c4b300c5031673061ca67e8.scope - libcontainer container ab69ff09a1e13c79148ae93d09b20041583947ea6c4b300c5031673061ca67e8.
Mar 20 21:30:38.027746 systemd[1]: Started cri-containerd-b0b6a6ed7a2cfb80df91a87bd4df59d50ac762e42fd67c65b57e7b0071d97c26.scope - libcontainer container b0b6a6ed7a2cfb80df91a87bd4df59d50ac762e42fd67c65b57e7b0071d97c26.
Mar 20 21:30:38.032348 systemd[1]: Started cri-containerd-9ebe67afe3f8fd6ed0c565746cb4d5fdfaf775cfc9a3a9701c0347757988c75a.scope - libcontainer container 9ebe67afe3f8fd6ed0c565746cb4d5fdfaf775cfc9a3a9701c0347757988c75a.
Mar 20 21:30:38.075957 containerd[1505]: time="2025-03-20T21:30:38.075752986Z" level=info msg="StartContainer for \"ab69ff09a1e13c79148ae93d09b20041583947ea6c4b300c5031673061ca67e8\" returns successfully"
Mar 20 21:30:38.082057 containerd[1505]: time="2025-03-20T21:30:38.081949815Z" level=info msg="StartContainer for \"b0b6a6ed7a2cfb80df91a87bd4df59d50ac762e42fd67c65b57e7b0071d97c26\" returns successfully"
Mar 20 21:30:38.092875 containerd[1505]: time="2025-03-20T21:30:38.092842056Z" level=info msg="StartContainer for \"9ebe67afe3f8fd6ed0c565746cb4d5fdfaf775cfc9a3a9701c0347757988c75a\" returns successfully"
Mar 20 21:30:38.121916 kubelet[2244]: W0320 21:30:38.121805    2244 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.132:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.132:6443: connect: connection refused
Mar 20 21:30:38.121916 kubelet[2244]: E0320 21:30:38.121873    2244 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.0.0.132:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.132:6443: connect: connection refused" logger="UnhandledError"
Mar 20 21:30:38.397375 kubelet[2244]: I0320 21:30:38.397261    2244 kubelet_node_status.go:72] "Attempting to register node" node="localhost"
Mar 20 21:30:39.049766 kubelet[2244]: I0320 21:30:39.049718    2244 kubelet_node_status.go:75] "Successfully registered node" node="localhost"
Mar 20 21:30:39.049766 kubelet[2244]: E0320 21:30:39.049763    2244 kubelet_node_status.go:535] "Error updating node status, will retry" err="error getting node \"localhost\": node \"localhost\" not found"
Mar 20 21:30:39.057776 kubelet[2244]: E0320 21:30:39.057740    2244 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found"
Mar 20 21:30:39.158156 kubelet[2244]: E0320 21:30:39.158104    2244 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found"
Mar 20 21:30:39.258676 kubelet[2244]: E0320 21:30:39.258636    2244 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found"
Mar 20 21:30:39.358999 kubelet[2244]: E0320 21:30:39.358883    2244 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found"
Mar 20 21:30:39.459413 kubelet[2244]: E0320 21:30:39.459388    2244 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found"
Mar 20 21:30:39.559632 kubelet[2244]: E0320 21:30:39.559591    2244 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found"
Mar 20 21:30:39.660438 kubelet[2244]: E0320 21:30:39.660302    2244 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found"
Mar 20 21:30:39.761173 kubelet[2244]: E0320 21:30:39.761131    2244 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found"
Mar 20 21:30:39.861744 kubelet[2244]: E0320 21:30:39.861704    2244 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found"
Mar 20 21:30:39.962678 kubelet[2244]: E0320 21:30:39.962520    2244 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found"
Mar 20 21:30:40.063264 kubelet[2244]: E0320 21:30:40.063211    2244 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found"
Mar 20 21:30:40.163832 kubelet[2244]: E0320 21:30:40.163756    2244 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found"
Mar 20 21:30:40.264396 kubelet[2244]: E0320 21:30:40.264277    2244 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found"
Mar 20 21:30:40.364637 kubelet[2244]: E0320 21:30:40.364587    2244 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found"
Mar 20 21:30:40.465141 kubelet[2244]: E0320 21:30:40.465097    2244 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found"
Mar 20 21:30:40.565827 kubelet[2244]: E0320 21:30:40.565720    2244 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found"
Mar 20 21:30:41.062134 systemd[1]: Reload requested from client PID 2516 ('systemctl') (unit session-7.scope)...
Mar 20 21:30:41.062151 systemd[1]: Reloading...
Mar 20 21:30:41.150660 zram_generator::config[2563]: No configuration found.
Mar 20 21:30:41.378946 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly.
Mar 20 21:30:41.497434 systemd[1]: Reloading finished in 434 ms.
Mar 20 21:30:41.522200 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent...
Mar 20 21:30:41.547177 systemd[1]: kubelet.service: Deactivated successfully.
Mar 20 21:30:41.547480 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent.
Mar 20 21:30:41.547535 systemd[1]: kubelet.service: Consumed 821ms CPU time, 118.9M memory peak.
Mar 20 21:30:41.550411 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent...
Mar 20 21:30:41.724181 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent.
Mar 20 21:30:41.729033 (kubelet)[2605]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS
Mar 20 21:30:41.764142 kubelet[2605]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information.
Mar 20 21:30:41.764142 kubelet[2605]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI.
Mar 20 21:30:41.764142 kubelet[2605]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information.
Mar 20 21:30:41.764540 kubelet[2605]: I0320 21:30:41.764185    2605 server.go:206] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime"
Mar 20 21:30:41.770945 kubelet[2605]: I0320 21:30:41.770911    2605 server.go:486] "Kubelet version" kubeletVersion="v1.31.0"
Mar 20 21:30:41.770945 kubelet[2605]: I0320 21:30:41.770936    2605 server.go:488] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK=""
Mar 20 21:30:41.771188 kubelet[2605]: I0320 21:30:41.771171    2605 server.go:929] "Client rotation is on, will bootstrap in background"
Mar 20 21:30:41.772386 kubelet[2605]: I0320 21:30:41.772366    2605 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem".
Mar 20 21:30:41.774117 kubelet[2605]: I0320 21:30:41.774098    2605 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt"
Mar 20 21:30:41.777753 kubelet[2605]: I0320 21:30:41.777715    2605 server.go:1426] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd"
Mar 20 21:30:41.782211 kubelet[2605]: I0320 21:30:41.782181    2605 server.go:744] "--cgroups-per-qos enabled, but --cgroup-root was not specified.  defaulting to /"
Mar 20 21:30:41.782307 kubelet[2605]: I0320 21:30:41.782286    2605 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority"
Mar 20 21:30:41.782441 kubelet[2605]: I0320 21:30:41.782407    2605 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[]
Mar 20 21:30:41.782586 kubelet[2605]: I0320 21:30:41.782434    2605 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2}
Mar 20 21:30:41.782679 kubelet[2605]: I0320 21:30:41.782585    2605 topology_manager.go:138] "Creating topology manager with none policy"
Mar 20 21:30:41.782679 kubelet[2605]: I0320 21:30:41.782594    2605 container_manager_linux.go:300] "Creating device plugin manager"
Mar 20 21:30:41.782679 kubelet[2605]: I0320 21:30:41.782635    2605 state_mem.go:36] "Initialized new in-memory state store"
Mar 20 21:30:41.782781 kubelet[2605]: I0320 21:30:41.782733    2605 kubelet.go:408] "Attempting to sync node with API server"
Mar 20 21:30:41.782781 kubelet[2605]: I0320 21:30:41.782744    2605 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests"
Mar 20 21:30:41.782781 kubelet[2605]: I0320 21:30:41.782771    2605 kubelet.go:314] "Adding apiserver pod source"
Mar 20 21:30:41.782781 kubelet[2605]: I0320 21:30:41.782790    2605 apiserver.go:42] "Waiting for node sync before watching apiserver pods"
Mar 20 21:30:41.783723 kubelet[2605]: I0320 21:30:41.783704    2605 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v2.0.1" apiVersion="v1"
Mar 20 21:30:41.786917 kubelet[2605]: I0320 21:30:41.786894    2605 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode"
Mar 20 21:30:41.787412 kubelet[2605]: I0320 21:30:41.787393    2605 server.go:1269] "Started kubelet"
Mar 20 21:30:41.788305 kubelet[2605]: I0320 21:30:41.788230    2605 server.go:163] "Starting to listen" address="0.0.0.0" port=10250
Mar 20 21:30:41.791634 kubelet[2605]: I0320 21:30:41.789491    2605 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer"
Mar 20 21:30:41.791634 kubelet[2605]: I0320 21:30:41.791555    2605 server.go:460] "Adding debug handlers to kubelet server"
Mar 20 21:30:41.793422 kubelet[2605]: I0320 21:30:41.793377    2605 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10
Mar 20 21:30:41.794305 kubelet[2605]: I0320 21:30:41.794279    2605 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key"
Mar 20 21:30:41.794703 kubelet[2605]: I0320 21:30:41.794689    2605 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock"
Mar 20 21:30:41.795704 kubelet[2605]: E0320 21:30:41.795688    2605 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem"
Mar 20 21:30:41.796752 kubelet[2605]: I0320 21:30:41.796731    2605 volume_manager.go:289] "Starting Kubelet Volume Manager"
Mar 20 21:30:41.796850 kubelet[2605]: I0320 21:30:41.796833    2605 desired_state_of_world_populator.go:146] "Desired state populator starts to run"
Mar 20 21:30:41.796999 kubelet[2605]: I0320 21:30:41.796975    2605 reconciler.go:26] "Reconciler: start to sync state"
Mar 20 21:30:41.797522 kubelet[2605]: I0320 21:30:41.797508    2605 factory.go:221] Registration of the systemd container factory successfully
Mar 20 21:30:41.797685 kubelet[2605]: I0320 21:30:41.797669    2605 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory
Mar 20 21:30:41.798843 kubelet[2605]: I0320 21:30:41.798823    2605 factory.go:221] Registration of the containerd container factory successfully
Mar 20 21:30:41.803537 kubelet[2605]: I0320 21:30:41.803424    2605 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4"
Mar 20 21:30:41.804659 kubelet[2605]: I0320 21:30:41.804518    2605 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6"
Mar 20 21:30:41.804659 kubelet[2605]: I0320 21:30:41.804545    2605 status_manager.go:217] "Starting to sync pod status with apiserver"
Mar 20 21:30:41.804659 kubelet[2605]: I0320 21:30:41.804559    2605 kubelet.go:2321] "Starting kubelet main sync loop"
Mar 20 21:30:41.804659 kubelet[2605]: E0320 21:30:41.804591    2605 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]"
Mar 20 21:30:41.834173 kubelet[2605]: I0320 21:30:41.834140    2605 cpu_manager.go:214] "Starting CPU manager" policy="none"
Mar 20 21:30:41.834173 kubelet[2605]: I0320 21:30:41.834159    2605 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s"
Mar 20 21:30:41.834173 kubelet[2605]: I0320 21:30:41.834175    2605 state_mem.go:36] "Initialized new in-memory state store"
Mar 20 21:30:41.834339 kubelet[2605]: I0320 21:30:41.834307    2605 state_mem.go:88] "Updated default CPUSet" cpuSet=""
Mar 20 21:30:41.834339 kubelet[2605]: I0320 21:30:41.834316    2605 state_mem.go:96] "Updated CPUSet assignments" assignments={}
Mar 20 21:30:41.834339 kubelet[2605]: I0320 21:30:41.834334    2605 policy_none.go:49] "None policy: Start"
Mar 20 21:30:41.835003 kubelet[2605]: I0320 21:30:41.834953    2605 memory_manager.go:170] "Starting memorymanager" policy="None"
Mar 20 21:30:41.835003 kubelet[2605]: I0320 21:30:41.834983    2605 state_mem.go:35] "Initializing new in-memory state store"
Mar 20 21:30:41.835179 kubelet[2605]: I0320 21:30:41.835161    2605 state_mem.go:75] "Updated machine memory state"
Mar 20 21:30:41.839109 kubelet[2605]: I0320 21:30:41.839081    2605 manager.go:510] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found"
Mar 20 21:30:41.839327 kubelet[2605]: I0320 21:30:41.839298    2605 eviction_manager.go:189] "Eviction manager: starting control loop"
Mar 20 21:30:41.839626 kubelet[2605]: I0320 21:30:41.839314    2605 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s"
Mar 20 21:30:41.839626 kubelet[2605]: I0320 21:30:41.839469    2605 plugin_manager.go:118] "Starting Kubelet Plugin Manager"
Mar 20 21:30:41.942565 kubelet[2605]: I0320 21:30:41.942513    2605 kubelet_node_status.go:72] "Attempting to register node" node="localhost"
Mar 20 21:30:41.998017 kubelet[2605]: I0320 21:30:41.997894    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/60762308083b5ef6c837b1be48ec53d6-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"60762308083b5ef6c837b1be48ec53d6\") " pod="kube-system/kube-controller-manager-localhost"
Mar 20 21:30:42.044824 kubelet[2605]: I0320 21:30:42.044797    2605 kubelet_node_status.go:111] "Node was previously registered" node="localhost"
Mar 20 21:30:42.044920 kubelet[2605]: I0320 21:30:42.044870    2605 kubelet_node_status.go:75] "Successfully registered node" node="localhost"
Mar 20 21:30:42.097526 sudo[2638]:     root : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin
Mar 20 21:30:42.097901 sudo[2638]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=0)
Mar 20 21:30:42.098260 kubelet[2605]: I0320 21:30:42.098216    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/60762308083b5ef6c837b1be48ec53d6-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"60762308083b5ef6c837b1be48ec53d6\") " pod="kube-system/kube-controller-manager-localhost"
Mar 20 21:30:42.098260 kubelet[2605]: I0320 21:30:42.098253    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/60762308083b5ef6c837b1be48ec53d6-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"60762308083b5ef6c837b1be48ec53d6\") " pod="kube-system/kube-controller-manager-localhost"
Mar 20 21:30:42.098475 kubelet[2605]: I0320 21:30:42.098273    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/6f32907a07e55aea05abdc5cd284a8d5-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"6f32907a07e55aea05abdc5cd284a8d5\") " pod="kube-system/kube-scheduler-localhost"
Mar 20 21:30:42.098475 kubelet[2605]: I0320 21:30:42.098289    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/a2d7033018db514077a55f9bbb204c5f-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"a2d7033018db514077a55f9bbb204c5f\") " pod="kube-system/kube-apiserver-localhost"
Mar 20 21:30:42.098475 kubelet[2605]: I0320 21:30:42.098304    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/a2d7033018db514077a55f9bbb204c5f-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"a2d7033018db514077a55f9bbb204c5f\") " pod="kube-system/kube-apiserver-localhost"
Mar 20 21:30:42.098475 kubelet[2605]: I0320 21:30:42.098337    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/60762308083b5ef6c837b1be48ec53d6-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"60762308083b5ef6c837b1be48ec53d6\") " pod="kube-system/kube-controller-manager-localhost"
Mar 20 21:30:42.098475 kubelet[2605]: I0320 21:30:42.098352    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/60762308083b5ef6c837b1be48ec53d6-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"60762308083b5ef6c837b1be48ec53d6\") " pod="kube-system/kube-controller-manager-localhost"
Mar 20 21:30:42.098595 kubelet[2605]: I0320 21:30:42.098368    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/a2d7033018db514077a55f9bbb204c5f-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"a2d7033018db514077a55f9bbb204c5f\") " pod="kube-system/kube-apiserver-localhost"
Mar 20 21:30:42.550865 sudo[2638]: pam_unix(sudo:session): session closed for user root
Mar 20 21:30:42.783608 kubelet[2605]: I0320 21:30:42.783564    2605 apiserver.go:52] "Watching apiserver"
Mar 20 21:30:42.797084 kubelet[2605]: I0320 21:30:42.797058    2605 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world"
Mar 20 21:30:42.831910 kubelet[2605]: E0320 21:30:42.831788    2605 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost"
Mar 20 21:30:42.836642 kubelet[2605]: I0320 21:30:42.836570    2605 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.836555127 podStartE2EDuration="1.836555127s" podCreationTimestamp="2025-03-20 21:30:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-20 21:30:42.836264839 +0000 UTC m=+1.103571387" watchObservedRunningTime="2025-03-20 21:30:42.836555127 +0000 UTC m=+1.103861676"
Mar 20 21:30:42.842992 kubelet[2605]: I0320 21:30:42.842933    2605 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=1.842922831 podStartE2EDuration="1.842922831s" podCreationTimestamp="2025-03-20 21:30:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-20 21:30:42.842788695 +0000 UTC m=+1.110095263" watchObservedRunningTime="2025-03-20 21:30:42.842922831 +0000 UTC m=+1.110229379"
Mar 20 21:30:42.849520 kubelet[2605]: I0320 21:30:42.849471    2605 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.849420066 podStartE2EDuration="1.849420066s" podCreationTimestamp="2025-03-20 21:30:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-20 21:30:42.849129497 +0000 UTC m=+1.116436045" watchObservedRunningTime="2025-03-20 21:30:42.849420066 +0000 UTC m=+1.116726614"
Mar 20 21:30:43.761078 sudo[1704]: pam_unix(sudo:session): session closed for user root
Mar 20 21:30:43.762401 sshd[1703]: Connection closed by 10.0.0.1 port 47644
Mar 20 21:30:43.762745 sshd-session[1700]: pam_unix(sshd:session): session closed for user core
Mar 20 21:30:43.766809 systemd[1]: sshd@6-10.0.0.132:22-10.0.0.1:47644.service: Deactivated successfully.
Mar 20 21:30:43.768960 systemd[1]: session-7.scope: Deactivated successfully.
Mar 20 21:30:43.769181 systemd[1]: session-7.scope: Consumed 4.137s CPU time, 261.2M memory peak.
Mar 20 21:30:43.770424 systemd-logind[1487]: Session 7 logged out. Waiting for processes to exit.
Mar 20 21:30:43.771333 systemd-logind[1487]: Removed session 7.
Mar 20 21:30:47.185586 kubelet[2605]: I0320 21:30:47.185544    2605 kuberuntime_manager.go:1633] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24"
Mar 20 21:30:47.186279 kubelet[2605]: I0320 21:30:47.186013    2605 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24"
Mar 20 21:30:47.186322 containerd[1505]: time="2025-03-20T21:30:47.185845352Z" level=info msg="No cni config template is specified, wait for other system components to drop the config."
Mar 20 21:30:48.050927 systemd[1]: Created slice kubepods-besteffort-podc695062b_61cf_4132_bfd4_fa5893629989.slice - libcontainer container kubepods-besteffort-podc695062b_61cf_4132_bfd4_fa5893629989.slice.
Mar 20 21:30:48.075932 systemd[1]: Created slice kubepods-burstable-pod80e00c81_b62f_4915_bd2e_8b9b6e12ff4c.slice - libcontainer container kubepods-burstable-pod80e00c81_b62f_4915_bd2e_8b9b6e12ff4c.slice.
Mar 20 21:30:48.140582 kubelet[2605]: I0320 21:30:48.140550    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c695062b-61cf-4132-bfd4-fa5893629989-lib-modules\") pod \"kube-proxy-59wfc\" (UID: \"c695062b-61cf-4132-bfd4-fa5893629989\") " pod="kube-system/kube-proxy-59wfc"
Mar 20 21:30:48.140582 kubelet[2605]: I0320 21:30:48.140581    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-clustermesh-secrets\") pod \"cilium-kvbxq\" (UID: \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\") " pod="kube-system/cilium-kvbxq"
Mar 20 21:30:48.140748 kubelet[2605]: I0320 21:30:48.140596    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5j5v\" (UniqueName: \"kubernetes.io/projected/c695062b-61cf-4132-bfd4-fa5893629989-kube-api-access-x5j5v\") pod \"kube-proxy-59wfc\" (UID: \"c695062b-61cf-4132-bfd4-fa5893629989\") " pod="kube-system/kube-proxy-59wfc"
Mar 20 21:30:48.140748 kubelet[2605]: I0320 21:30:48.140634    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-xtables-lock\") pod \"cilium-kvbxq\" (UID: \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\") " pod="kube-system/cilium-kvbxq"
Mar 20 21:30:48.140748 kubelet[2605]: I0320 21:30:48.140649    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-hostproc\") pod \"cilium-kvbxq\" (UID: \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\") " pod="kube-system/cilium-kvbxq"
Mar 20 21:30:48.140748 kubelet[2605]: I0320 21:30:48.140665    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/c695062b-61cf-4132-bfd4-fa5893629989-xtables-lock\") pod \"kube-proxy-59wfc\" (UID: \"c695062b-61cf-4132-bfd4-fa5893629989\") " pod="kube-system/kube-proxy-59wfc"
Mar 20 21:30:48.140748 kubelet[2605]: I0320 21:30:48.140690    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-cilium-cgroup\") pod \"cilium-kvbxq\" (UID: \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\") " pod="kube-system/cilium-kvbxq"
Mar 20 21:30:48.140748 kubelet[2605]: I0320 21:30:48.140711    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-etc-cni-netd\") pod \"cilium-kvbxq\" (UID: \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\") " pod="kube-system/cilium-kvbxq"
Mar 20 21:30:48.140982 kubelet[2605]: I0320 21:30:48.140725    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-cilium-config-path\") pod \"cilium-kvbxq\" (UID: \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\") " pod="kube-system/cilium-kvbxq"
Mar 20 21:30:48.140982 kubelet[2605]: I0320 21:30:48.140740    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-host-proc-sys-net\") pod \"cilium-kvbxq\" (UID: \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\") " pod="kube-system/cilium-kvbxq"
Mar 20 21:30:48.140982 kubelet[2605]: I0320 21:30:48.140756    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-host-proc-sys-kernel\") pod \"cilium-kvbxq\" (UID: \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\") " pod="kube-system/cilium-kvbxq"
Mar 20 21:30:48.140982 kubelet[2605]: I0320 21:30:48.140779    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/c695062b-61cf-4132-bfd4-fa5893629989-kube-proxy\") pod \"kube-proxy-59wfc\" (UID: \"c695062b-61cf-4132-bfd4-fa5893629989\") " pod="kube-system/kube-proxy-59wfc"
Mar 20 21:30:48.140982 kubelet[2605]: I0320 21:30:48.140881    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-cilium-run\") pod \"cilium-kvbxq\" (UID: \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\") " pod="kube-system/cilium-kvbxq"
Mar 20 21:30:48.140982 kubelet[2605]: I0320 21:30:48.140952    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-bpf-maps\") pod \"cilium-kvbxq\" (UID: \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\") " pod="kube-system/cilium-kvbxq"
Mar 20 21:30:48.141127 kubelet[2605]: I0320 21:30:48.140983    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-cni-path\") pod \"cilium-kvbxq\" (UID: \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\") " pod="kube-system/cilium-kvbxq"
Mar 20 21:30:48.141127 kubelet[2605]: I0320 21:30:48.141006    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhlh9\" (UniqueName: \"kubernetes.io/projected/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-kube-api-access-zhlh9\") pod \"cilium-kvbxq\" (UID: \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\") " pod="kube-system/cilium-kvbxq"
Mar 20 21:30:48.141127 kubelet[2605]: I0320 21:30:48.141035    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-lib-modules\") pod \"cilium-kvbxq\" (UID: \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\") " pod="kube-system/cilium-kvbxq"
Mar 20 21:30:48.141127 kubelet[2605]: I0320 21:30:48.141060    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-hubble-tls\") pod \"cilium-kvbxq\" (UID: \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\") " pod="kube-system/cilium-kvbxq"
Mar 20 21:30:48.448038 kubelet[2605]: E0320 21:30:48.447507    2605 projected.go:288] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found
Mar 20 21:30:48.448038 kubelet[2605]: E0320 21:30:48.447553    2605 projected.go:194] Error preparing data for projected volume kube-api-access-zhlh9 for pod kube-system/cilium-kvbxq: configmap "kube-root-ca.crt" not found
Mar 20 21:30:48.448038 kubelet[2605]: E0320 21:30:48.447667    2605 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-kube-api-access-zhlh9 podName:80e00c81-b62f-4915-bd2e-8b9b6e12ff4c nodeName:}" failed. No retries permitted until 2025-03-20 21:30:48.947641864 +0000 UTC m=+7.214948402 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-zhlh9" (UniqueName: "kubernetes.io/projected/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-kube-api-access-zhlh9") pod "cilium-kvbxq" (UID: "80e00c81-b62f-4915-bd2e-8b9b6e12ff4c") : configmap "kube-root-ca.crt" not found
Mar 20 21:30:48.448038 kubelet[2605]: E0320 21:30:48.447857    2605 projected.go:288] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found
Mar 20 21:30:48.448038 kubelet[2605]: E0320 21:30:48.447874    2605 projected.go:194] Error preparing data for projected volume kube-api-access-x5j5v for pod kube-system/kube-proxy-59wfc: configmap "kube-root-ca.crt" not found
Mar 20 21:30:48.448038 kubelet[2605]: E0320 21:30:48.447912    2605 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c695062b-61cf-4132-bfd4-fa5893629989-kube-api-access-x5j5v podName:c695062b-61cf-4132-bfd4-fa5893629989 nodeName:}" failed. No retries permitted until 2025-03-20 21:30:48.947900901 +0000 UTC m=+7.215207449 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-x5j5v" (UniqueName: "kubernetes.io/projected/c695062b-61cf-4132-bfd4-fa5893629989-kube-api-access-x5j5v") pod "kube-proxy-59wfc" (UID: "c695062b-61cf-4132-bfd4-fa5893629989") : configmap "kube-root-ca.crt" not found
Mar 20 21:30:48.482148 systemd[1]: Created slice kubepods-besteffort-pod2edb670f_4b11_4193_9194_56522899f17d.slice - libcontainer container kubepods-besteffort-pod2edb670f_4b11_4193_9194_56522899f17d.slice.
Mar 20 21:30:48.645887 kubelet[2605]: I0320 21:30:48.645818    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gzjk\" (UniqueName: \"kubernetes.io/projected/2edb670f-4b11-4193-9194-56522899f17d-kube-api-access-4gzjk\") pod \"cilium-operator-5d85765b45-h4496\" (UID: \"2edb670f-4b11-4193-9194-56522899f17d\") " pod="kube-system/cilium-operator-5d85765b45-h4496"
Mar 20 21:30:48.645887 kubelet[2605]: I0320 21:30:48.645887    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/2edb670f-4b11-4193-9194-56522899f17d-cilium-config-path\") pod \"cilium-operator-5d85765b45-h4496\" (UID: \"2edb670f-4b11-4193-9194-56522899f17d\") " pod="kube-system/cilium-operator-5d85765b45-h4496"
Mar 20 21:30:48.786960 containerd[1505]: time="2025-03-20T21:30:48.786906011Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-5d85765b45-h4496,Uid:2edb670f-4b11-4193-9194-56522899f17d,Namespace:kube-system,Attempt:0,}"
Mar 20 21:30:48.806955 containerd[1505]: time="2025-03-20T21:30:48.806899543Z" level=info msg="connecting to shim 512c6fd7d4ffe0f78b9b8897bf2679d5eeeaf851306b1867d446d6dfa9e49ca0" address="unix:///run/containerd/s/b8cca4cd6cd3fe608e5ce35f8a76f2255148327041a0aa17ff12f88dc7f5886b" namespace=k8s.io protocol=ttrpc version=3
Mar 20 21:30:48.831746 systemd[1]: Started cri-containerd-512c6fd7d4ffe0f78b9b8897bf2679d5eeeaf851306b1867d446d6dfa9e49ca0.scope - libcontainer container 512c6fd7d4ffe0f78b9b8897bf2679d5eeeaf851306b1867d446d6dfa9e49ca0.
Mar 20 21:30:48.872387 containerd[1505]: time="2025-03-20T21:30:48.872327331Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-5d85765b45-h4496,Uid:2edb670f-4b11-4193-9194-56522899f17d,Namespace:kube-system,Attempt:0,} returns sandbox id \"512c6fd7d4ffe0f78b9b8897bf2679d5eeeaf851306b1867d446d6dfa9e49ca0\""
Mar 20 21:30:48.875070 containerd[1505]: time="2025-03-20T21:30:48.875010584Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\""
Mar 20 21:30:48.973973 containerd[1505]: time="2025-03-20T21:30:48.973929344Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-59wfc,Uid:c695062b-61cf-4132-bfd4-fa5893629989,Namespace:kube-system,Attempt:0,}"
Mar 20 21:30:48.979527 containerd[1505]: time="2025-03-20T21:30:48.979490794Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-kvbxq,Uid:80e00c81-b62f-4915-bd2e-8b9b6e12ff4c,Namespace:kube-system,Attempt:0,}"
Mar 20 21:30:48.999185 containerd[1505]: time="2025-03-20T21:30:48.999133363Z" level=info msg="connecting to shim 7d217f1ed89ed965e2602cb6d137bbb56eeada05d8d84d90f979effd81b10b40" address="unix:///run/containerd/s/4d7755752333d4dfd6b0b973f2cb3c128295b058f03a7c3270f7ea505dc250da" namespace=k8s.io protocol=ttrpc version=3
Mar 20 21:30:49.008892 containerd[1505]: time="2025-03-20T21:30:49.008845520Z" level=info msg="connecting to shim bd74444e833f33dc7dc44cf1755614cabdcda78497ceca51625f9284ac3f126a" address="unix:///run/containerd/s/dbe14d736a188f1725a075e0a6385e478e409a0dcc9e3849b0558f210e750834" namespace=k8s.io protocol=ttrpc version=3
Mar 20 21:30:49.021756 systemd[1]: Started cri-containerd-7d217f1ed89ed965e2602cb6d137bbb56eeada05d8d84d90f979effd81b10b40.scope - libcontainer container 7d217f1ed89ed965e2602cb6d137bbb56eeada05d8d84d90f979effd81b10b40.
Mar 20 21:30:49.024734 systemd[1]: Started cri-containerd-bd74444e833f33dc7dc44cf1755614cabdcda78497ceca51625f9284ac3f126a.scope - libcontainer container bd74444e833f33dc7dc44cf1755614cabdcda78497ceca51625f9284ac3f126a.
Mar 20 21:30:49.049173 containerd[1505]: time="2025-03-20T21:30:49.048899861Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-59wfc,Uid:c695062b-61cf-4132-bfd4-fa5893629989,Namespace:kube-system,Attempt:0,} returns sandbox id \"7d217f1ed89ed965e2602cb6d137bbb56eeada05d8d84d90f979effd81b10b40\""
Mar 20 21:30:49.051905 containerd[1505]: time="2025-03-20T21:30:49.051850138Z" level=info msg="CreateContainer within sandbox \"7d217f1ed89ed965e2602cb6d137bbb56eeada05d8d84d90f979effd81b10b40\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}"
Mar 20 21:30:49.055585 containerd[1505]: time="2025-03-20T21:30:49.055531075Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-kvbxq,Uid:80e00c81-b62f-4915-bd2e-8b9b6e12ff4c,Namespace:kube-system,Attempt:0,} returns sandbox id \"bd74444e833f33dc7dc44cf1755614cabdcda78497ceca51625f9284ac3f126a\""
Mar 20 21:30:49.064965 containerd[1505]: time="2025-03-20T21:30:49.064925380Z" level=info msg="Container 2c4d32da447b25fae7ce0dce8c0e2fa9b18f088191f0454d1ecd90f211f93a21: CDI devices from CRI Config.CDIDevices: []"
Mar 20 21:30:49.073711 containerd[1505]: time="2025-03-20T21:30:49.073675101Z" level=info msg="CreateContainer within sandbox \"7d217f1ed89ed965e2602cb6d137bbb56eeada05d8d84d90f979effd81b10b40\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"2c4d32da447b25fae7ce0dce8c0e2fa9b18f088191f0454d1ecd90f211f93a21\""
Mar 20 21:30:49.074343 containerd[1505]: time="2025-03-20T21:30:49.074146012Z" level=info msg="StartContainer for \"2c4d32da447b25fae7ce0dce8c0e2fa9b18f088191f0454d1ecd90f211f93a21\""
Mar 20 21:30:49.075438 containerd[1505]: time="2025-03-20T21:30:49.075404541Z" level=info msg="connecting to shim 2c4d32da447b25fae7ce0dce8c0e2fa9b18f088191f0454d1ecd90f211f93a21" address="unix:///run/containerd/s/4d7755752333d4dfd6b0b973f2cb3c128295b058f03a7c3270f7ea505dc250da" protocol=ttrpc version=3
Mar 20 21:30:49.096748 systemd[1]: Started cri-containerd-2c4d32da447b25fae7ce0dce8c0e2fa9b18f088191f0454d1ecd90f211f93a21.scope - libcontainer container 2c4d32da447b25fae7ce0dce8c0e2fa9b18f088191f0454d1ecd90f211f93a21.
Mar 20 21:30:49.136134 containerd[1505]: time="2025-03-20T21:30:49.136092142Z" level=info msg="StartContainer for \"2c4d32da447b25fae7ce0dce8c0e2fa9b18f088191f0454d1ecd90f211f93a21\" returns successfully"
Mar 20 21:30:49.838876 kubelet[2605]: I0320 21:30:49.838806    2605 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-59wfc" podStartSLOduration=1.838790044 podStartE2EDuration="1.838790044s" podCreationTimestamp="2025-03-20 21:30:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-20 21:30:49.838516371 +0000 UTC m=+8.105822919" watchObservedRunningTime="2025-03-20 21:30:49.838790044 +0000 UTC m=+8.106096592"
Mar 20 21:30:51.003066 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2441052074.mount: Deactivated successfully.
Mar 20 21:30:51.328219 containerd[1505]: time="2025-03-20T21:30:51.328079383Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Mar 20 21:30:51.328817 containerd[1505]: time="2025-03-20T21:30:51.328773048Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=18904197"
Mar 20 21:30:51.329889 containerd[1505]: time="2025-03-20T21:30:51.329862880Z" level=info msg="ImageCreate event name:\"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Mar 20 21:30:51.331007 containerd[1505]: time="2025-03-20T21:30:51.330978942Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"18897442\" in 2.455933942s"
Mar 20 21:30:51.331040 containerd[1505]: time="2025-03-20T21:30:51.331008679Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\""
Mar 20 21:30:51.332088 containerd[1505]: time="2025-03-20T21:30:51.332024580Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\""
Mar 20 21:30:51.332949 containerd[1505]: time="2025-03-20T21:30:51.332913808Z" level=info msg="CreateContainer within sandbox \"512c6fd7d4ffe0f78b9b8897bf2679d5eeeaf851306b1867d446d6dfa9e49ca0\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}"
Mar 20 21:30:51.342201 containerd[1505]: time="2025-03-20T21:30:51.341625489Z" level=info msg="Container 3d97c67b7858b412ccbd97394b377bd9d015d155fc32e9e6a3f745ad4db55eee: CDI devices from CRI Config.CDIDevices: []"
Mar 20 21:30:51.347987 containerd[1505]: time="2025-03-20T21:30:51.347951513Z" level=info msg="CreateContainer within sandbox \"512c6fd7d4ffe0f78b9b8897bf2679d5eeeaf851306b1867d446d6dfa9e49ca0\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"3d97c67b7858b412ccbd97394b377bd9d015d155fc32e9e6a3f745ad4db55eee\""
Mar 20 21:30:51.348391 containerd[1505]: time="2025-03-20T21:30:51.348368490Z" level=info msg="StartContainer for \"3d97c67b7858b412ccbd97394b377bd9d015d155fc32e9e6a3f745ad4db55eee\""
Mar 20 21:30:51.349247 containerd[1505]: time="2025-03-20T21:30:51.349217521Z" level=info msg="connecting to shim 3d97c67b7858b412ccbd97394b377bd9d015d155fc32e9e6a3f745ad4db55eee" address="unix:///run/containerd/s/b8cca4cd6cd3fe608e5ce35f8a76f2255148327041a0aa17ff12f88dc7f5886b" protocol=ttrpc version=3
Mar 20 21:30:51.389754 systemd[1]: Started cri-containerd-3d97c67b7858b412ccbd97394b377bd9d015d155fc32e9e6a3f745ad4db55eee.scope - libcontainer container 3d97c67b7858b412ccbd97394b377bd9d015d155fc32e9e6a3f745ad4db55eee.
Mar 20 21:30:51.420163 containerd[1505]: time="2025-03-20T21:30:51.420126582Z" level=info msg="StartContainer for \"3d97c67b7858b412ccbd97394b377bd9d015d155fc32e9e6a3f745ad4db55eee\" returns successfully"
Mar 20 21:30:54.786998 kubelet[2605]: I0320 21:30:54.786933    2605 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-5d85765b45-h4496" podStartSLOduration=4.329549338 podStartE2EDuration="6.786909458s" podCreationTimestamp="2025-03-20 21:30:48 +0000 UTC" firstStartedPulling="2025-03-20 21:30:48.874454809 +0000 UTC m=+7.141761357" lastFinishedPulling="2025-03-20 21:30:51.331814929 +0000 UTC m=+9.599121477" observedRunningTime="2025-03-20 21:30:51.856872724 +0000 UTC m=+10.124179272" watchObservedRunningTime="2025-03-20 21:30:54.786909458 +0000 UTC m=+13.054216016"
Mar 20 21:30:59.597567 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount257154787.mount: Deactivated successfully.
Mar 20 21:31:00.988309 update_engine[1489]: I20250320 21:31:00.988232  1489 update_attempter.cc:509] Updating boot flags...
Mar 20 21:31:01.079200 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 39 scanned by (udev-worker) (3045)
Mar 20 21:31:01.113679 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 39 scanned by (udev-worker) (3043)
Mar 20 21:31:01.144652 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 39 scanned by (udev-worker) (3043)
Mar 20 21:31:03.335766 containerd[1505]: time="2025-03-20T21:31:03.335711681Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Mar 20 21:31:03.336489 containerd[1505]: time="2025-03-20T21:31:03.336437176Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=166730503"
Mar 20 21:31:03.337561 containerd[1505]: time="2025-03-20T21:31:03.337532720Z" level=info msg="ImageCreate event name:\"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Mar 20 21:31:03.339054 containerd[1505]: time="2025-03-20T21:31:03.339008344Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"166719855\" in 12.006933599s"
Mar 20 21:31:03.339054 containerd[1505]: time="2025-03-20T21:31:03.339051586Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\""
Mar 20 21:31:03.366849 containerd[1505]: time="2025-03-20T21:31:03.366786408Z" level=info msg="CreateContainer within sandbox \"bd74444e833f33dc7dc44cf1755614cabdcda78497ceca51625f9284ac3f126a\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}"
Mar 20 21:31:03.375696 containerd[1505]: time="2025-03-20T21:31:03.375600640Z" level=info msg="Container cb5cccb990530e77805b5371aa20a2d743334ab726fb5f69adfaaaa8c2c84c37: CDI devices from CRI Config.CDIDevices: []"
Mar 20 21:31:03.382326 containerd[1505]: time="2025-03-20T21:31:03.382285209Z" level=info msg="CreateContainer within sandbox \"bd74444e833f33dc7dc44cf1755614cabdcda78497ceca51625f9284ac3f126a\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"cb5cccb990530e77805b5371aa20a2d743334ab726fb5f69adfaaaa8c2c84c37\""
Mar 20 21:31:03.385584 containerd[1505]: time="2025-03-20T21:31:03.385534602Z" level=info msg="StartContainer for \"cb5cccb990530e77805b5371aa20a2d743334ab726fb5f69adfaaaa8c2c84c37\""
Mar 20 21:31:03.386530 containerd[1505]: time="2025-03-20T21:31:03.386505631Z" level=info msg="connecting to shim cb5cccb990530e77805b5371aa20a2d743334ab726fb5f69adfaaaa8c2c84c37" address="unix:///run/containerd/s/dbe14d736a188f1725a075e0a6385e478e409a0dcc9e3849b0558f210e750834" protocol=ttrpc version=3
Mar 20 21:31:03.409848 systemd[1]: Started cri-containerd-cb5cccb990530e77805b5371aa20a2d743334ab726fb5f69adfaaaa8c2c84c37.scope - libcontainer container cb5cccb990530e77805b5371aa20a2d743334ab726fb5f69adfaaaa8c2c84c37.
Mar 20 21:31:03.440354 containerd[1505]: time="2025-03-20T21:31:03.440307697Z" level=info msg="StartContainer for \"cb5cccb990530e77805b5371aa20a2d743334ab726fb5f69adfaaaa8c2c84c37\" returns successfully"
Mar 20 21:31:03.449428 systemd[1]: cri-containerd-cb5cccb990530e77805b5371aa20a2d743334ab726fb5f69adfaaaa8c2c84c37.scope: Deactivated successfully.
Mar 20 21:31:03.450973 containerd[1505]: time="2025-03-20T21:31:03.450928741Z" level=info msg="received exit event container_id:\"cb5cccb990530e77805b5371aa20a2d743334ab726fb5f69adfaaaa8c2c84c37\"  id:\"cb5cccb990530e77805b5371aa20a2d743334ab726fb5f69adfaaaa8c2c84c37\"  pid:3091  exited_at:{seconds:1742506263  nanos:450516641}"
Mar 20 21:31:03.451068 containerd[1505]: time="2025-03-20T21:31:03.451035353Z" level=info msg="TaskExit event in podsandbox handler container_id:\"cb5cccb990530e77805b5371aa20a2d743334ab726fb5f69adfaaaa8c2c84c37\"  id:\"cb5cccb990530e77805b5371aa20a2d743334ab726fb5f69adfaaaa8c2c84c37\"  pid:3091  exited_at:{seconds:1742506263  nanos:450516641}"
Mar 20 21:31:03.470753 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-cb5cccb990530e77805b5371aa20a2d743334ab726fb5f69adfaaaa8c2c84c37-rootfs.mount: Deactivated successfully.
Mar 20 21:31:04.645521 systemd[1]: Started sshd@7-10.0.0.132:22-10.0.0.1:47228.service - OpenSSH per-connection server daemon (10.0.0.1:47228).
Mar 20 21:31:04.697609 sshd[3127]: Accepted publickey for core from 10.0.0.1 port 47228 ssh2: RSA SHA256:KJ7ck8imsv1/sWVS7eR1M7V7NSskkAYjKibngyOtAC0
Mar 20 21:31:04.698969 sshd-session[3127]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Mar 20 21:31:04.703575 systemd-logind[1487]: New session 8 of user core.
Mar 20 21:31:04.717764 systemd[1]: Started session-8.scope - Session 8 of User core.
Mar 20 21:31:04.844493 sshd[3130]: Connection closed by 10.0.0.1 port 47228
Mar 20 21:31:04.844799 sshd-session[3127]: pam_unix(sshd:session): session closed for user core
Mar 20 21:31:04.849043 systemd[1]: sshd@7-10.0.0.132:22-10.0.0.1:47228.service: Deactivated successfully.
Mar 20 21:31:04.851226 systemd[1]: session-8.scope: Deactivated successfully.
Mar 20 21:31:04.852008 systemd-logind[1487]: Session 8 logged out. Waiting for processes to exit.
Mar 20 21:31:04.852955 systemd-logind[1487]: Removed session 8.
Mar 20 21:31:05.326747 containerd[1505]: time="2025-03-20T21:31:05.326706167Z" level=info msg="CreateContainer within sandbox \"bd74444e833f33dc7dc44cf1755614cabdcda78497ceca51625f9284ac3f126a\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}"
Mar 20 21:31:05.338830 containerd[1505]: time="2025-03-20T21:31:05.338780181Z" level=info msg="Container 97939959568ed3aae7628f5d62b3b233027c5c6ff1c16a18a8a2dd67839d9fbd: CDI devices from CRI Config.CDIDevices: []"
Mar 20 21:31:05.345321 containerd[1505]: time="2025-03-20T21:31:05.345273897Z" level=info msg="CreateContainer within sandbox \"bd74444e833f33dc7dc44cf1755614cabdcda78497ceca51625f9284ac3f126a\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"97939959568ed3aae7628f5d62b3b233027c5c6ff1c16a18a8a2dd67839d9fbd\""
Mar 20 21:31:05.345768 containerd[1505]: time="2025-03-20T21:31:05.345736222Z" level=info msg="StartContainer for \"97939959568ed3aae7628f5d62b3b233027c5c6ff1c16a18a8a2dd67839d9fbd\""
Mar 20 21:31:05.346602 containerd[1505]: time="2025-03-20T21:31:05.346572253Z" level=info msg="connecting to shim 97939959568ed3aae7628f5d62b3b233027c5c6ff1c16a18a8a2dd67839d9fbd" address="unix:///run/containerd/s/dbe14d736a188f1725a075e0a6385e478e409a0dcc9e3849b0558f210e750834" protocol=ttrpc version=3
Mar 20 21:31:05.365746 systemd[1]: Started cri-containerd-97939959568ed3aae7628f5d62b3b233027c5c6ff1c16a18a8a2dd67839d9fbd.scope - libcontainer container 97939959568ed3aae7628f5d62b3b233027c5c6ff1c16a18a8a2dd67839d9fbd.
Mar 20 21:31:05.393894 containerd[1505]: time="2025-03-20T21:31:05.393860333Z" level=info msg="StartContainer for \"97939959568ed3aae7628f5d62b3b233027c5c6ff1c16a18a8a2dd67839d9fbd\" returns successfully"
Mar 20 21:31:05.405914 systemd[1]: systemd-sysctl.service: Deactivated successfully.
Mar 20 21:31:05.406148 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables.
Mar 20 21:31:05.406376 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables...
Mar 20 21:31:05.408103 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables...
Mar 20 21:31:05.410004 containerd[1505]: time="2025-03-20T21:31:05.409968100Z" level=info msg="TaskExit event in podsandbox handler container_id:\"97939959568ed3aae7628f5d62b3b233027c5c6ff1c16a18a8a2dd67839d9fbd\"  id:\"97939959568ed3aae7628f5d62b3b233027c5c6ff1c16a18a8a2dd67839d9fbd\"  pid:3160  exited_at:{seconds:1742506265  nanos:409630080}"
Mar 20 21:31:05.410233 containerd[1505]: time="2025-03-20T21:31:05.410108314Z" level=info msg="received exit event container_id:\"97939959568ed3aae7628f5d62b3b233027c5c6ff1c16a18a8a2dd67839d9fbd\"  id:\"97939959568ed3aae7628f5d62b3b233027c5c6ff1c16a18a8a2dd67839d9fbd\"  pid:3160  exited_at:{seconds:1742506265  nanos:409630080}"
Mar 20 21:31:05.410002 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully.
Mar 20 21:31:05.410406 systemd[1]: cri-containerd-97939959568ed3aae7628f5d62b3b233027c5c6ff1c16a18a8a2dd67839d9fbd.scope: Deactivated successfully.
Mar 20 21:31:05.430199 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables.
Mar 20 21:31:06.329237 containerd[1505]: time="2025-03-20T21:31:06.329194228Z" level=info msg="CreateContainer within sandbox \"bd74444e833f33dc7dc44cf1755614cabdcda78497ceca51625f9284ac3f126a\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}"
Mar 20 21:31:06.339680 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-97939959568ed3aae7628f5d62b3b233027c5c6ff1c16a18a8a2dd67839d9fbd-rootfs.mount: Deactivated successfully.
Mar 20 21:31:06.342751 containerd[1505]: time="2025-03-20T21:31:06.342704860Z" level=info msg="Container 43a3e9daa8b637bb9ca9efd72c1f81f1f2072fba03af0871655bd90bc07a1d80: CDI devices from CRI Config.CDIDevices: []"
Mar 20 21:31:06.355167 containerd[1505]: time="2025-03-20T21:31:06.355109290Z" level=info msg="CreateContainer within sandbox \"bd74444e833f33dc7dc44cf1755614cabdcda78497ceca51625f9284ac3f126a\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"43a3e9daa8b637bb9ca9efd72c1f81f1f2072fba03af0871655bd90bc07a1d80\""
Mar 20 21:31:06.355634 containerd[1505]: time="2025-03-20T21:31:06.355587535Z" level=info msg="StartContainer for \"43a3e9daa8b637bb9ca9efd72c1f81f1f2072fba03af0871655bd90bc07a1d80\""
Mar 20 21:31:06.356927 containerd[1505]: time="2025-03-20T21:31:06.356886251Z" level=info msg="connecting to shim 43a3e9daa8b637bb9ca9efd72c1f81f1f2072fba03af0871655bd90bc07a1d80" address="unix:///run/containerd/s/dbe14d736a188f1725a075e0a6385e478e409a0dcc9e3849b0558f210e750834" protocol=ttrpc version=3
Mar 20 21:31:06.382775 systemd[1]: Started cri-containerd-43a3e9daa8b637bb9ca9efd72c1f81f1f2072fba03af0871655bd90bc07a1d80.scope - libcontainer container 43a3e9daa8b637bb9ca9efd72c1f81f1f2072fba03af0871655bd90bc07a1d80.
Mar 20 21:31:06.425834 systemd[1]: cri-containerd-43a3e9daa8b637bb9ca9efd72c1f81f1f2072fba03af0871655bd90bc07a1d80.scope: Deactivated successfully.
Mar 20 21:31:06.427018 containerd[1505]: time="2025-03-20T21:31:06.426776661Z" level=info msg="received exit event container_id:\"43a3e9daa8b637bb9ca9efd72c1f81f1f2072fba03af0871655bd90bc07a1d80\"  id:\"43a3e9daa8b637bb9ca9efd72c1f81f1f2072fba03af0871655bd90bc07a1d80\"  pid:3209  exited_at:{seconds:1742506266  nanos:426538270}"
Mar 20 21:31:06.427018 containerd[1505]: time="2025-03-20T21:31:06.426946422Z" level=info msg="TaskExit event in podsandbox handler container_id:\"43a3e9daa8b637bb9ca9efd72c1f81f1f2072fba03af0871655bd90bc07a1d80\"  id:\"43a3e9daa8b637bb9ca9efd72c1f81f1f2072fba03af0871655bd90bc07a1d80\"  pid:3209  exited_at:{seconds:1742506266  nanos:426538270}"
Mar 20 21:31:06.427161 containerd[1505]: time="2025-03-20T21:31:06.427130840Z" level=info msg="StartContainer for \"43a3e9daa8b637bb9ca9efd72c1f81f1f2072fba03af0871655bd90bc07a1d80\" returns successfully"
Mar 20 21:31:06.446008 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-43a3e9daa8b637bb9ca9efd72c1f81f1f2072fba03af0871655bd90bc07a1d80-rootfs.mount: Deactivated successfully.
Mar 20 21:31:07.332633 containerd[1505]: time="2025-03-20T21:31:07.332584039Z" level=info msg="CreateContainer within sandbox \"bd74444e833f33dc7dc44cf1755614cabdcda78497ceca51625f9284ac3f126a\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}"
Mar 20 21:31:07.546148 containerd[1505]: time="2025-03-20T21:31:07.546081199Z" level=info msg="Container 7d7d70dae2773a709a180aa31e705b7948b49abcacac51a54048a6dfb2d459fb: CDI devices from CRI Config.CDIDevices: []"
Mar 20 21:31:07.586049 containerd[1505]: time="2025-03-20T21:31:07.585916198Z" level=info msg="CreateContainer within sandbox \"bd74444e833f33dc7dc44cf1755614cabdcda78497ceca51625f9284ac3f126a\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"7d7d70dae2773a709a180aa31e705b7948b49abcacac51a54048a6dfb2d459fb\""
Mar 20 21:31:07.586731 containerd[1505]: time="2025-03-20T21:31:07.586392929Z" level=info msg="StartContainer for \"7d7d70dae2773a709a180aa31e705b7948b49abcacac51a54048a6dfb2d459fb\""
Mar 20 21:31:07.587398 containerd[1505]: time="2025-03-20T21:31:07.587376318Z" level=info msg="connecting to shim 7d7d70dae2773a709a180aa31e705b7948b49abcacac51a54048a6dfb2d459fb" address="unix:///run/containerd/s/dbe14d736a188f1725a075e0a6385e478e409a0dcc9e3849b0558f210e750834" protocol=ttrpc version=3
Mar 20 21:31:07.609799 systemd[1]: Started cri-containerd-7d7d70dae2773a709a180aa31e705b7948b49abcacac51a54048a6dfb2d459fb.scope - libcontainer container 7d7d70dae2773a709a180aa31e705b7948b49abcacac51a54048a6dfb2d459fb.
Mar 20 21:31:07.636433 systemd[1]: cri-containerd-7d7d70dae2773a709a180aa31e705b7948b49abcacac51a54048a6dfb2d459fb.scope: Deactivated successfully.
Mar 20 21:31:07.637018 containerd[1505]: time="2025-03-20T21:31:07.636981378Z" level=info msg="TaskExit event in podsandbox handler container_id:\"7d7d70dae2773a709a180aa31e705b7948b49abcacac51a54048a6dfb2d459fb\"  id:\"7d7d70dae2773a709a180aa31e705b7948b49abcacac51a54048a6dfb2d459fb\"  pid:3247  exited_at:{seconds:1742506267  nanos:636608413}"
Mar 20 21:31:07.639354 containerd[1505]: time="2025-03-20T21:31:07.639318015Z" level=info msg="received exit event container_id:\"7d7d70dae2773a709a180aa31e705b7948b49abcacac51a54048a6dfb2d459fb\"  id:\"7d7d70dae2773a709a180aa31e705b7948b49abcacac51a54048a6dfb2d459fb\"  pid:3247  exited_at:{seconds:1742506267  nanos:636608413}"
Mar 20 21:31:07.647345 containerd[1505]: time="2025-03-20T21:31:07.647302690Z" level=info msg="StartContainer for \"7d7d70dae2773a709a180aa31e705b7948b49abcacac51a54048a6dfb2d459fb\" returns successfully"
Mar 20 21:31:07.660001 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-7d7d70dae2773a709a180aa31e705b7948b49abcacac51a54048a6dfb2d459fb-rootfs.mount: Deactivated successfully.
Mar 20 21:31:08.337399 containerd[1505]: time="2025-03-20T21:31:08.337352535Z" level=info msg="CreateContainer within sandbox \"bd74444e833f33dc7dc44cf1755614cabdcda78497ceca51625f9284ac3f126a\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}"
Mar 20 21:31:08.350324 containerd[1505]: time="2025-03-20T21:31:08.350277641Z" level=info msg="Container 9d0628577cc16ff5b8bfd1c78c11a5ff4cfc3ff46fc5f579fb65f7339975cef2: CDI devices from CRI Config.CDIDevices: []"
Mar 20 21:31:08.357311 containerd[1505]: time="2025-03-20T21:31:08.357281296Z" level=info msg="CreateContainer within sandbox \"bd74444e833f33dc7dc44cf1755614cabdcda78497ceca51625f9284ac3f126a\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"9d0628577cc16ff5b8bfd1c78c11a5ff4cfc3ff46fc5f579fb65f7339975cef2\""
Mar 20 21:31:08.357845 containerd[1505]: time="2025-03-20T21:31:08.357822139Z" level=info msg="StartContainer for \"9d0628577cc16ff5b8bfd1c78c11a5ff4cfc3ff46fc5f579fb65f7339975cef2\""
Mar 20 21:31:08.358818 containerd[1505]: time="2025-03-20T21:31:08.358777635Z" level=info msg="connecting to shim 9d0628577cc16ff5b8bfd1c78c11a5ff4cfc3ff46fc5f579fb65f7339975cef2" address="unix:///run/containerd/s/dbe14d736a188f1725a075e0a6385e478e409a0dcc9e3849b0558f210e750834" protocol=ttrpc version=3
Mar 20 21:31:08.379734 systemd[1]: Started cri-containerd-9d0628577cc16ff5b8bfd1c78c11a5ff4cfc3ff46fc5f579fb65f7339975cef2.scope - libcontainer container 9d0628577cc16ff5b8bfd1c78c11a5ff4cfc3ff46fc5f579fb65f7339975cef2.
Mar 20 21:31:08.415978 containerd[1505]: time="2025-03-20T21:31:08.414629446Z" level=info msg="StartContainer for \"9d0628577cc16ff5b8bfd1c78c11a5ff4cfc3ff46fc5f579fb65f7339975cef2\" returns successfully"
Mar 20 21:31:08.478376 containerd[1505]: time="2025-03-20T21:31:08.478331222Z" level=info msg="TaskExit event in podsandbox handler container_id:\"9d0628577cc16ff5b8bfd1c78c11a5ff4cfc3ff46fc5f579fb65f7339975cef2\"  id:\"a72e790af42ca2f47ed8611d746dcb71ee25da17ff72466ac96406148e342014\"  pid:3315  exited_at:{seconds:1742506268  nanos:477983044}"
Mar 20 21:31:08.507639 kubelet[2605]: I0320 21:31:08.507580    2605 kubelet_node_status.go:488] "Fast updating node status as it just became ready"
Mar 20 21:31:08.534038 systemd[1]: Created slice kubepods-burstable-pod754226cb_8a4f_4730_aa0e_c8424595350d.slice - libcontainer container kubepods-burstable-pod754226cb_8a4f_4730_aa0e_c8424595350d.slice.
Mar 20 21:31:08.539899 systemd[1]: Created slice kubepods-burstable-pod5a02776e_3f1a_421b_b49a_fc6db4e6040f.slice - libcontainer container kubepods-burstable-pod5a02776e_3f1a_421b_b49a_fc6db4e6040f.slice.
Mar 20 21:31:08.548357 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2010812474.mount: Deactivated successfully.
Mar 20 21:31:08.681149 kubelet[2605]: I0320 21:31:08.681024    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/754226cb-8a4f-4730-aa0e-c8424595350d-config-volume\") pod \"coredns-6f6b679f8f-g9n5l\" (UID: \"754226cb-8a4f-4730-aa0e-c8424595350d\") " pod="kube-system/coredns-6f6b679f8f-g9n5l"
Mar 20 21:31:08.681149 kubelet[2605]: I0320 21:31:08.681074    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5a02776e-3f1a-421b-b49a-fc6db4e6040f-config-volume\") pod \"coredns-6f6b679f8f-696q4\" (UID: \"5a02776e-3f1a-421b-b49a-fc6db4e6040f\") " pod="kube-system/coredns-6f6b679f8f-696q4"
Mar 20 21:31:08.681149 kubelet[2605]: I0320 21:31:08.681091    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjh45\" (UniqueName: \"kubernetes.io/projected/5a02776e-3f1a-421b-b49a-fc6db4e6040f-kube-api-access-gjh45\") pod \"coredns-6f6b679f8f-696q4\" (UID: \"5a02776e-3f1a-421b-b49a-fc6db4e6040f\") " pod="kube-system/coredns-6f6b679f8f-696q4"
Mar 20 21:31:08.681149 kubelet[2605]: I0320 21:31:08.681111    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4f4bx\" (UniqueName: \"kubernetes.io/projected/754226cb-8a4f-4730-aa0e-c8424595350d-kube-api-access-4f4bx\") pod \"coredns-6f6b679f8f-g9n5l\" (UID: \"754226cb-8a4f-4730-aa0e-c8424595350d\") " pod="kube-system/coredns-6f6b679f8f-g9n5l"
Mar 20 21:31:08.837589 containerd[1505]: time="2025-03-20T21:31:08.837539198Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-g9n5l,Uid:754226cb-8a4f-4730-aa0e-c8424595350d,Namespace:kube-system,Attempt:0,}"
Mar 20 21:31:08.843356 containerd[1505]: time="2025-03-20T21:31:08.843298130Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-696q4,Uid:5a02776e-3f1a-421b-b49a-fc6db4e6040f,Namespace:kube-system,Attempt:0,}"
Mar 20 21:31:09.538360 kubelet[2605]: I0320 21:31:09.538272    2605 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-kvbxq" podStartSLOduration=7.246859757 podStartE2EDuration="21.538248823s" podCreationTimestamp="2025-03-20 21:30:48 +0000 UTC" firstStartedPulling="2025-03-20 21:30:49.056311048 +0000 UTC m=+7.323617596" lastFinishedPulling="2025-03-20 21:31:03.347700114 +0000 UTC m=+21.615006662" observedRunningTime="2025-03-20 21:31:09.538113698 +0000 UTC m=+27.805420246" watchObservedRunningTime="2025-03-20 21:31:09.538248823 +0000 UTC m=+27.805555371"
Mar 20 21:31:09.860748 systemd[1]: Started sshd@8-10.0.0.132:22-10.0.0.1:60072.service - OpenSSH per-connection server daemon (10.0.0.1:60072).
Mar 20 21:31:09.933031 sshd[3411]: Accepted publickey for core from 10.0.0.1 port 60072 ssh2: RSA SHA256:KJ7ck8imsv1/sWVS7eR1M7V7NSskkAYjKibngyOtAC0
Mar 20 21:31:09.934281 sshd-session[3411]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Mar 20 21:31:09.940464 systemd-logind[1487]: New session 9 of user core.
Mar 20 21:31:09.949737 systemd[1]: Started session-9.scope - Session 9 of User core.
Mar 20 21:31:10.099284 sshd[3413]: Connection closed by 10.0.0.1 port 60072
Mar 20 21:31:10.099585 sshd-session[3411]: pam_unix(sshd:session): session closed for user core
Mar 20 21:31:10.103869 systemd[1]: sshd@8-10.0.0.132:22-10.0.0.1:60072.service: Deactivated successfully.
Mar 20 21:31:10.106005 systemd[1]: session-9.scope: Deactivated successfully.
Mar 20 21:31:10.106682 systemd-logind[1487]: Session 9 logged out. Waiting for processes to exit.
Mar 20 21:31:10.107502 systemd-logind[1487]: Removed session 9.
Mar 20 21:31:10.618740 systemd-networkd[1424]: cilium_host: Link UP
Mar 20 21:31:10.618934 systemd-networkd[1424]: cilium_net: Link UP
Mar 20 21:31:10.619117 systemd-networkd[1424]: cilium_net: Gained carrier
Mar 20 21:31:10.619300 systemd-networkd[1424]: cilium_host: Gained carrier
Mar 20 21:31:10.725765 systemd-networkd[1424]: cilium_vxlan: Link UP
Mar 20 21:31:10.725907 systemd-networkd[1424]: cilium_vxlan: Gained carrier
Mar 20 21:31:10.927657 kernel: NET: Registered PF_ALG protocol family
Mar 20 21:31:11.076802 systemd-networkd[1424]: cilium_host: Gained IPv6LL
Mar 20 21:31:11.253209 systemd-networkd[1424]: cilium_net: Gained IPv6LL
Mar 20 21:31:11.603347 systemd-networkd[1424]: lxc_health: Link UP
Mar 20 21:31:11.612003 systemd-networkd[1424]: lxc_health: Gained carrier
Mar 20 21:31:11.993214 systemd-networkd[1424]: lxc72029ac25a96: Link UP
Mar 20 21:31:12.001642 kernel: eth0: renamed from tmp9c295
Mar 20 21:31:12.007499 systemd-networkd[1424]: lxc72029ac25a96: Gained carrier
Mar 20 21:31:12.182547 systemd-networkd[1424]: lxc2c778aad39fa: Link UP
Mar 20 21:31:12.185648 kernel: eth0: renamed from tmpd0dec
Mar 20 21:31:12.192258 systemd-networkd[1424]: lxc2c778aad39fa: Gained carrier
Mar 20 21:31:12.340781 systemd-networkd[1424]: cilium_vxlan: Gained IPv6LL
Mar 20 21:31:13.044782 systemd-networkd[1424]: lxc_health: Gained IPv6LL
Mar 20 21:31:13.555817 systemd-networkd[1424]: lxc2c778aad39fa: Gained IPv6LL
Mar 20 21:31:13.875767 systemd-networkd[1424]: lxc72029ac25a96: Gained IPv6LL
Mar 20 21:31:15.115323 systemd[1]: Started sshd@9-10.0.0.132:22-10.0.0.1:60080.service - OpenSSH per-connection server daemon (10.0.0.1:60080).
Mar 20 21:31:15.164652 sshd[3811]: Accepted publickey for core from 10.0.0.1 port 60080 ssh2: RSA SHA256:KJ7ck8imsv1/sWVS7eR1M7V7NSskkAYjKibngyOtAC0
Mar 20 21:31:15.164711 sshd-session[3811]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Mar 20 21:31:15.169123 systemd-logind[1487]: New session 10 of user core.
Mar 20 21:31:15.176746 systemd[1]: Started session-10.scope - Session 10 of User core.
Mar 20 21:31:15.321081 sshd[3816]: Connection closed by 10.0.0.1 port 60080
Mar 20 21:31:15.321718 sshd-session[3811]: pam_unix(sshd:session): session closed for user core
Mar 20 21:31:15.327189 systemd[1]: sshd@9-10.0.0.132:22-10.0.0.1:60080.service: Deactivated successfully.
Mar 20 21:31:15.329099 systemd[1]: session-10.scope: Deactivated successfully.
Mar 20 21:31:15.329780 systemd-logind[1487]: Session 10 logged out. Waiting for processes to exit.
Mar 20 21:31:15.330661 systemd-logind[1487]: Removed session 10.
Mar 20 21:31:15.619418 containerd[1505]: time="2025-03-20T21:31:15.619358494Z" level=info msg="connecting to shim 9c2956023041480b5e266a35cca815ca6220bb781dc2b1323bb1cbed45f15244" address="unix:///run/containerd/s/0371531366f88e50adfe34655b2e0f835f737c9cce03884e11a955670186c868" namespace=k8s.io protocol=ttrpc version=3
Mar 20 21:31:15.620767 containerd[1505]: time="2025-03-20T21:31:15.620736082Z" level=info msg="connecting to shim d0dec458760ffc3b7544460a322017e7142b6e7fd4f479509e036110606f7947" address="unix:///run/containerd/s/a5fe467cd55e0753ff71144625678de0d955db9676af90e2f3cc52e9b254b5ff" namespace=k8s.io protocol=ttrpc version=3
Mar 20 21:31:15.651902 systemd[1]: Started cri-containerd-9c2956023041480b5e266a35cca815ca6220bb781dc2b1323bb1cbed45f15244.scope - libcontainer container 9c2956023041480b5e266a35cca815ca6220bb781dc2b1323bb1cbed45f15244.
Mar 20 21:31:15.653981 systemd[1]: Started cri-containerd-d0dec458760ffc3b7544460a322017e7142b6e7fd4f479509e036110606f7947.scope - libcontainer container d0dec458760ffc3b7544460a322017e7142b6e7fd4f479509e036110606f7947.
Mar 20 21:31:15.666462 systemd-resolved[1341]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address
Mar 20 21:31:15.667917 systemd-resolved[1341]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address
Mar 20 21:31:15.824393 containerd[1505]: time="2025-03-20T21:31:15.824340507Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-696q4,Uid:5a02776e-3f1a-421b-b49a-fc6db4e6040f,Namespace:kube-system,Attempt:0,} returns sandbox id \"9c2956023041480b5e266a35cca815ca6220bb781dc2b1323bb1cbed45f15244\""
Mar 20 21:31:15.826853 containerd[1505]: time="2025-03-20T21:31:15.826519617Z" level=info msg="CreateContainer within sandbox \"9c2956023041480b5e266a35cca815ca6220bb781dc2b1323bb1cbed45f15244\" for container &ContainerMetadata{Name:coredns,Attempt:0,}"
Mar 20 21:31:15.900572 containerd[1505]: time="2025-03-20T21:31:15.900461851Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-g9n5l,Uid:754226cb-8a4f-4730-aa0e-c8424595350d,Namespace:kube-system,Attempt:0,} returns sandbox id \"d0dec458760ffc3b7544460a322017e7142b6e7fd4f479509e036110606f7947\""
Mar 20 21:31:15.903029 containerd[1505]: time="2025-03-20T21:31:15.902981453Z" level=info msg="CreateContainer within sandbox \"d0dec458760ffc3b7544460a322017e7142b6e7fd4f479509e036110606f7947\" for container &ContainerMetadata{Name:coredns,Attempt:0,}"
Mar 20 21:31:16.030044 containerd[1505]: time="2025-03-20T21:31:16.029991961Z" level=info msg="Container 9718d9b2cecc34e7a24cbc628c27996a102035b2b216e6373d8b1a0c77285e5e: CDI devices from CRI Config.CDIDevices: []"
Mar 20 21:31:16.032311 containerd[1505]: time="2025-03-20T21:31:16.032248627Z" level=info msg="Container 0a9a156ab0dd286e699514f29e3d6cd72015613e22b9ebcd3209b75a18809e6d: CDI devices from CRI Config.CDIDevices: []"
Mar 20 21:31:16.038958 containerd[1505]: time="2025-03-20T21:31:16.038908385Z" level=info msg="CreateContainer within sandbox \"9c2956023041480b5e266a35cca815ca6220bb781dc2b1323bb1cbed45f15244\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"9718d9b2cecc34e7a24cbc628c27996a102035b2b216e6373d8b1a0c77285e5e\""
Mar 20 21:31:16.039498 containerd[1505]: time="2025-03-20T21:31:16.039453623Z" level=info msg="StartContainer for \"9718d9b2cecc34e7a24cbc628c27996a102035b2b216e6373d8b1a0c77285e5e\""
Mar 20 21:31:16.040350 containerd[1505]: time="2025-03-20T21:31:16.040324806Z" level=info msg="connecting to shim 9718d9b2cecc34e7a24cbc628c27996a102035b2b216e6373d8b1a0c77285e5e" address="unix:///run/containerd/s/0371531366f88e50adfe34655b2e0f835f737c9cce03884e11a955670186c868" protocol=ttrpc version=3
Mar 20 21:31:16.042909 containerd[1505]: time="2025-03-20T21:31:16.042867862Z" level=info msg="CreateContainer within sandbox \"d0dec458760ffc3b7544460a322017e7142b6e7fd4f479509e036110606f7947\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"0a9a156ab0dd286e699514f29e3d6cd72015613e22b9ebcd3209b75a18809e6d\""
Mar 20 21:31:16.043380 containerd[1505]: time="2025-03-20T21:31:16.043328511Z" level=info msg="StartContainer for \"0a9a156ab0dd286e699514f29e3d6cd72015613e22b9ebcd3209b75a18809e6d\""
Mar 20 21:31:16.044243 containerd[1505]: time="2025-03-20T21:31:16.044183052Z" level=info msg="connecting to shim 0a9a156ab0dd286e699514f29e3d6cd72015613e22b9ebcd3209b75a18809e6d" address="unix:///run/containerd/s/a5fe467cd55e0753ff71144625678de0d955db9676af90e2f3cc52e9b254b5ff" protocol=ttrpc version=3
Mar 20 21:31:16.063760 systemd[1]: Started cri-containerd-0a9a156ab0dd286e699514f29e3d6cd72015613e22b9ebcd3209b75a18809e6d.scope - libcontainer container 0a9a156ab0dd286e699514f29e3d6cd72015613e22b9ebcd3209b75a18809e6d.
Mar 20 21:31:16.066351 systemd[1]: Started cri-containerd-9718d9b2cecc34e7a24cbc628c27996a102035b2b216e6373d8b1a0c77285e5e.scope - libcontainer container 9718d9b2cecc34e7a24cbc628c27996a102035b2b216e6373d8b1a0c77285e5e.
Mar 20 21:31:16.098847 containerd[1505]: time="2025-03-20T21:31:16.098640303Z" level=info msg="StartContainer for \"0a9a156ab0dd286e699514f29e3d6cd72015613e22b9ebcd3209b75a18809e6d\" returns successfully"
Mar 20 21:31:16.099411 containerd[1505]: time="2025-03-20T21:31:16.099362525Z" level=info msg="StartContainer for \"9718d9b2cecc34e7a24cbc628c27996a102035b2b216e6373d8b1a0c77285e5e\" returns successfully"
Mar 20 21:31:16.365508 kubelet[2605]: I0320 21:31:16.365454    2605 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-6f6b679f8f-g9n5l" podStartSLOduration=28.365435606 podStartE2EDuration="28.365435606s" podCreationTimestamp="2025-03-20 21:30:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-20 21:31:16.365040591 +0000 UTC m=+34.632347139" watchObservedRunningTime="2025-03-20 21:31:16.365435606 +0000 UTC m=+34.632742154"
Mar 20 21:31:16.385073 kubelet[2605]: I0320 21:31:16.384994    2605 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-6f6b679f8f-696q4" podStartSLOduration=28.384972838 podStartE2EDuration="28.384972838s" podCreationTimestamp="2025-03-20 21:30:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-20 21:31:16.384828756 +0000 UTC m=+34.652135304" watchObservedRunningTime="2025-03-20 21:31:16.384972838 +0000 UTC m=+34.652279386"
Mar 20 21:31:20.334786 systemd[1]: Started sshd@10-10.0.0.132:22-10.0.0.1:50066.service - OpenSSH per-connection server daemon (10.0.0.1:50066).
Mar 20 21:31:20.388807 sshd[4005]: Accepted publickey for core from 10.0.0.1 port 50066 ssh2: RSA SHA256:KJ7ck8imsv1/sWVS7eR1M7V7NSskkAYjKibngyOtAC0
Mar 20 21:31:20.390422 sshd-session[4005]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Mar 20 21:31:20.394570 systemd-logind[1487]: New session 11 of user core.
Mar 20 21:31:20.407762 systemd[1]: Started session-11.scope - Session 11 of User core.
Mar 20 21:31:20.528007 sshd[4007]: Connection closed by 10.0.0.1 port 50066
Mar 20 21:31:20.528334 sshd-session[4005]: pam_unix(sshd:session): session closed for user core
Mar 20 21:31:20.531938 systemd[1]: sshd@10-10.0.0.132:22-10.0.0.1:50066.service: Deactivated successfully.
Mar 20 21:31:20.534384 systemd[1]: session-11.scope: Deactivated successfully.
Mar 20 21:31:20.535210 systemd-logind[1487]: Session 11 logged out. Waiting for processes to exit.
Mar 20 21:31:20.536260 systemd-logind[1487]: Removed session 11.
Mar 20 21:31:25.540735 systemd[1]: Started sshd@11-10.0.0.132:22-10.0.0.1:50074.service - OpenSSH per-connection server daemon (10.0.0.1:50074).
Mar 20 21:31:25.586063 sshd[4021]: Accepted publickey for core from 10.0.0.1 port 50074 ssh2: RSA SHA256:KJ7ck8imsv1/sWVS7eR1M7V7NSskkAYjKibngyOtAC0
Mar 20 21:31:25.587446 sshd-session[4021]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Mar 20 21:31:25.591521 systemd-logind[1487]: New session 12 of user core.
Mar 20 21:31:25.599726 systemd[1]: Started session-12.scope - Session 12 of User core.
Mar 20 21:31:25.716096 sshd[4023]: Connection closed by 10.0.0.1 port 50074
Mar 20 21:31:25.716454 sshd-session[4021]: pam_unix(sshd:session): session closed for user core
Mar 20 21:31:25.728386 systemd[1]: sshd@11-10.0.0.132:22-10.0.0.1:50074.service: Deactivated successfully.
Mar 20 21:31:25.730349 systemd[1]: session-12.scope: Deactivated successfully.
Mar 20 21:31:25.731699 systemd-logind[1487]: Session 12 logged out. Waiting for processes to exit.
Mar 20 21:31:25.733135 systemd[1]: Started sshd@12-10.0.0.132:22-10.0.0.1:60300.service - OpenSSH per-connection server daemon (10.0.0.1:60300).
Mar 20 21:31:25.734265 systemd-logind[1487]: Removed session 12.
Mar 20 21:31:25.779162 sshd[4036]: Accepted publickey for core from 10.0.0.1 port 60300 ssh2: RSA SHA256:KJ7ck8imsv1/sWVS7eR1M7V7NSskkAYjKibngyOtAC0
Mar 20 21:31:25.780693 sshd-session[4036]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Mar 20 21:31:25.784831 systemd-logind[1487]: New session 13 of user core.
Mar 20 21:31:25.794757 systemd[1]: Started session-13.scope - Session 13 of User core.
Mar 20 21:31:25.938703 sshd[4039]: Connection closed by 10.0.0.1 port 60300
Mar 20 21:31:25.940359 sshd-session[4036]: pam_unix(sshd:session): session closed for user core
Mar 20 21:31:25.950278 systemd[1]: sshd@12-10.0.0.132:22-10.0.0.1:60300.service: Deactivated successfully.
Mar 20 21:31:25.953276 systemd[1]: session-13.scope: Deactivated successfully.
Mar 20 21:31:25.954173 systemd-logind[1487]: Session 13 logged out. Waiting for processes to exit.
Mar 20 21:31:25.956976 systemd[1]: Started sshd@13-10.0.0.132:22-10.0.0.1:60304.service - OpenSSH per-connection server daemon (10.0.0.1:60304).
Mar 20 21:31:25.959155 systemd-logind[1487]: Removed session 13.
Mar 20 21:31:26.002638 sshd[4049]: Accepted publickey for core from 10.0.0.1 port 60304 ssh2: RSA SHA256:KJ7ck8imsv1/sWVS7eR1M7V7NSskkAYjKibngyOtAC0
Mar 20 21:31:26.004070 sshd-session[4049]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Mar 20 21:31:26.008330 systemd-logind[1487]: New session 14 of user core.
Mar 20 21:31:26.020734 systemd[1]: Started session-14.scope - Session 14 of User core.
Mar 20 21:31:26.137335 sshd[4052]: Connection closed by 10.0.0.1 port 60304
Mar 20 21:31:26.137605 sshd-session[4049]: pam_unix(sshd:session): session closed for user core
Mar 20 21:31:26.141873 systemd[1]: sshd@13-10.0.0.132:22-10.0.0.1:60304.service: Deactivated successfully.
Mar 20 21:31:26.143917 systemd[1]: session-14.scope: Deactivated successfully.
Mar 20 21:31:26.144722 systemd-logind[1487]: Session 14 logged out. Waiting for processes to exit.
Mar 20 21:31:26.145660 systemd-logind[1487]: Removed session 14.
Mar 20 21:31:31.149544 systemd[1]: Started sshd@14-10.0.0.132:22-10.0.0.1:60312.service - OpenSSH per-connection server daemon (10.0.0.1:60312).
Mar 20 21:31:31.200393 sshd[4066]: Accepted publickey for core from 10.0.0.1 port 60312 ssh2: RSA SHA256:KJ7ck8imsv1/sWVS7eR1M7V7NSskkAYjKibngyOtAC0
Mar 20 21:31:31.202224 sshd-session[4066]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Mar 20 21:31:31.206254 systemd-logind[1487]: New session 15 of user core.
Mar 20 21:31:31.212747 systemd[1]: Started session-15.scope - Session 15 of User core.
Mar 20 21:31:31.326257 sshd[4068]: Connection closed by 10.0.0.1 port 60312
Mar 20 21:31:31.326596 sshd-session[4066]: pam_unix(sshd:session): session closed for user core
Mar 20 21:31:31.331108 systemd[1]: sshd@14-10.0.0.132:22-10.0.0.1:60312.service: Deactivated successfully.
Mar 20 21:31:31.333233 systemd[1]: session-15.scope: Deactivated successfully.
Mar 20 21:31:31.333945 systemd-logind[1487]: Session 15 logged out. Waiting for processes to exit.
Mar 20 21:31:31.334886 systemd-logind[1487]: Removed session 15.
Mar 20 21:31:36.338748 systemd[1]: Started sshd@15-10.0.0.132:22-10.0.0.1:60494.service - OpenSSH per-connection server daemon (10.0.0.1:60494).
Mar 20 21:31:36.377675 sshd[4082]: Accepted publickey for core from 10.0.0.1 port 60494 ssh2: RSA SHA256:KJ7ck8imsv1/sWVS7eR1M7V7NSskkAYjKibngyOtAC0
Mar 20 21:31:36.379356 sshd-session[4082]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Mar 20 21:31:36.383407 systemd-logind[1487]: New session 16 of user core.
Mar 20 21:31:36.393829 systemd[1]: Started session-16.scope - Session 16 of User core.
Mar 20 21:31:36.501852 sshd[4084]: Connection closed by 10.0.0.1 port 60494
Mar 20 21:31:36.502294 sshd-session[4082]: pam_unix(sshd:session): session closed for user core
Mar 20 21:31:36.514183 systemd[1]: sshd@15-10.0.0.132:22-10.0.0.1:60494.service: Deactivated successfully.
Mar 20 21:31:36.515898 systemd[1]: session-16.scope: Deactivated successfully.
Mar 20 21:31:36.517401 systemd-logind[1487]: Session 16 logged out. Waiting for processes to exit.
Mar 20 21:31:36.518610 systemd[1]: Started sshd@16-10.0.0.132:22-10.0.0.1:60510.service - OpenSSH per-connection server daemon (10.0.0.1:60510).
Mar 20 21:31:36.519580 systemd-logind[1487]: Removed session 16.
Mar 20 21:31:36.567056 sshd[4096]: Accepted publickey for core from 10.0.0.1 port 60510 ssh2: RSA SHA256:KJ7ck8imsv1/sWVS7eR1M7V7NSskkAYjKibngyOtAC0
Mar 20 21:31:36.568379 sshd-session[4096]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Mar 20 21:31:36.572586 systemd-logind[1487]: New session 17 of user core.
Mar 20 21:31:36.587743 systemd[1]: Started session-17.scope - Session 17 of User core.
Mar 20 21:31:36.810205 sshd[4099]: Connection closed by 10.0.0.1 port 60510
Mar 20 21:31:36.810596 sshd-session[4096]: pam_unix(sshd:session): session closed for user core
Mar 20 21:31:36.823414 systemd[1]: sshd@16-10.0.0.132:22-10.0.0.1:60510.service: Deactivated successfully.
Mar 20 21:31:36.825331 systemd[1]: session-17.scope: Deactivated successfully.
Mar 20 21:31:36.826795 systemd-logind[1487]: Session 17 logged out. Waiting for processes to exit.
Mar 20 21:31:36.828278 systemd[1]: Started sshd@17-10.0.0.132:22-10.0.0.1:60516.service - OpenSSH per-connection server daemon (10.0.0.1:60516).
Mar 20 21:31:36.829433 systemd-logind[1487]: Removed session 17.
Mar 20 21:31:36.877092 sshd[4109]: Accepted publickey for core from 10.0.0.1 port 60516 ssh2: RSA SHA256:KJ7ck8imsv1/sWVS7eR1M7V7NSskkAYjKibngyOtAC0
Mar 20 21:31:36.878341 sshd-session[4109]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Mar 20 21:31:36.882448 systemd-logind[1487]: New session 18 of user core.
Mar 20 21:31:36.887753 systemd[1]: Started session-18.scope - Session 18 of User core.
Mar 20 21:31:38.153295 sshd[4112]: Connection closed by 10.0.0.1 port 60516
Mar 20 21:31:38.153717 sshd-session[4109]: pam_unix(sshd:session): session closed for user core
Mar 20 21:31:38.165691 systemd[1]: sshd@17-10.0.0.132:22-10.0.0.1:60516.service: Deactivated successfully.
Mar 20 21:31:38.168812 systemd[1]: session-18.scope: Deactivated successfully.
Mar 20 21:31:38.170570 systemd-logind[1487]: Session 18 logged out. Waiting for processes to exit.
Mar 20 21:31:38.173443 systemd[1]: Started sshd@18-10.0.0.132:22-10.0.0.1:60520.service - OpenSSH per-connection server daemon (10.0.0.1:60520).
Mar 20 21:31:38.174140 systemd-logind[1487]: Removed session 18.
Mar 20 21:31:38.219781 sshd[4131]: Accepted publickey for core from 10.0.0.1 port 60520 ssh2: RSA SHA256:KJ7ck8imsv1/sWVS7eR1M7V7NSskkAYjKibngyOtAC0
Mar 20 21:31:38.221188 sshd-session[4131]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Mar 20 21:31:38.225150 systemd-logind[1487]: New session 19 of user core.
Mar 20 21:31:38.238744 systemd[1]: Started session-19.scope - Session 19 of User core.
Mar 20 21:31:38.453373 sshd[4134]: Connection closed by 10.0.0.1 port 60520
Mar 20 21:31:38.454347 sshd-session[4131]: pam_unix(sshd:session): session closed for user core
Mar 20 21:31:38.466541 systemd[1]: sshd@18-10.0.0.132:22-10.0.0.1:60520.service: Deactivated successfully.
Mar 20 21:31:38.468372 systemd[1]: session-19.scope: Deactivated successfully.
Mar 20 21:31:38.469069 systemd-logind[1487]: Session 19 logged out. Waiting for processes to exit.
Mar 20 21:31:38.470985 systemd[1]: Started sshd@19-10.0.0.132:22-10.0.0.1:60532.service - OpenSSH per-connection server daemon (10.0.0.1:60532).
Mar 20 21:31:38.472166 systemd-logind[1487]: Removed session 19.
Mar 20 21:31:38.514765 sshd[4144]: Accepted publickey for core from 10.0.0.1 port 60532 ssh2: RSA SHA256:KJ7ck8imsv1/sWVS7eR1M7V7NSskkAYjKibngyOtAC0
Mar 20 21:31:38.516078 sshd-session[4144]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Mar 20 21:31:38.520571 systemd-logind[1487]: New session 20 of user core.
Mar 20 21:31:38.531723 systemd[1]: Started session-20.scope - Session 20 of User core.
Mar 20 21:31:38.657986 sshd[4147]: Connection closed by 10.0.0.1 port 60532
Mar 20 21:31:38.658343 sshd-session[4144]: pam_unix(sshd:session): session closed for user core
Mar 20 21:31:38.662551 systemd[1]: sshd@19-10.0.0.132:22-10.0.0.1:60532.service: Deactivated successfully.
Mar 20 21:31:38.665201 systemd[1]: session-20.scope: Deactivated successfully.
Mar 20 21:31:38.666004 systemd-logind[1487]: Session 20 logged out. Waiting for processes to exit.
Mar 20 21:31:38.667090 systemd-logind[1487]: Removed session 20.
Mar 20 21:31:43.674407 systemd[1]: Started sshd@20-10.0.0.132:22-10.0.0.1:60540.service - OpenSSH per-connection server daemon (10.0.0.1:60540).
Mar 20 21:31:43.724257 sshd[4166]: Accepted publickey for core from 10.0.0.1 port 60540 ssh2: RSA SHA256:KJ7ck8imsv1/sWVS7eR1M7V7NSskkAYjKibngyOtAC0
Mar 20 21:31:43.725718 sshd-session[4166]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Mar 20 21:31:43.729865 systemd-logind[1487]: New session 21 of user core.
Mar 20 21:31:43.740752 systemd[1]: Started session-21.scope - Session 21 of User core.
Mar 20 21:31:43.849717 sshd[4168]: Connection closed by 10.0.0.1 port 60540
Mar 20 21:31:43.850001 sshd-session[4166]: pam_unix(sshd:session): session closed for user core
Mar 20 21:31:43.854139 systemd[1]: sshd@20-10.0.0.132:22-10.0.0.1:60540.service: Deactivated successfully.
Mar 20 21:31:43.856275 systemd[1]: session-21.scope: Deactivated successfully.
Mar 20 21:31:43.856982 systemd-logind[1487]: Session 21 logged out. Waiting for processes to exit.
Mar 20 21:31:43.857902 systemd-logind[1487]: Removed session 21.
Mar 20 21:31:48.863263 systemd[1]: Started sshd@21-10.0.0.132:22-10.0.0.1:44634.service - OpenSSH per-connection server daemon (10.0.0.1:44634).
Mar 20 21:31:48.909505 sshd[4182]: Accepted publickey for core from 10.0.0.1 port 44634 ssh2: RSA SHA256:KJ7ck8imsv1/sWVS7eR1M7V7NSskkAYjKibngyOtAC0
Mar 20 21:31:48.911162 sshd-session[4182]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Mar 20 21:31:48.915503 systemd-logind[1487]: New session 22 of user core.
Mar 20 21:31:48.921755 systemd[1]: Started session-22.scope - Session 22 of User core.
Mar 20 21:31:49.032437 sshd[4184]: Connection closed by 10.0.0.1 port 44634
Mar 20 21:31:49.032809 sshd-session[4182]: pam_unix(sshd:session): session closed for user core
Mar 20 21:31:49.036736 systemd[1]: sshd@21-10.0.0.132:22-10.0.0.1:44634.service: Deactivated successfully.
Mar 20 21:31:49.038825 systemd[1]: session-22.scope: Deactivated successfully.
Mar 20 21:31:49.039515 systemd-logind[1487]: Session 22 logged out. Waiting for processes to exit.
Mar 20 21:31:49.040367 systemd-logind[1487]: Removed session 22.
Mar 20 21:31:54.046791 systemd[1]: Started sshd@22-10.0.0.132:22-10.0.0.1:44646.service - OpenSSH per-connection server daemon (10.0.0.1:44646).
Mar 20 21:31:54.094927 sshd[4200]: Accepted publickey for core from 10.0.0.1 port 44646 ssh2: RSA SHA256:KJ7ck8imsv1/sWVS7eR1M7V7NSskkAYjKibngyOtAC0
Mar 20 21:31:54.096322 sshd-session[4200]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Mar 20 21:31:54.100386 systemd-logind[1487]: New session 23 of user core.
Mar 20 21:31:54.109730 systemd[1]: Started session-23.scope - Session 23 of User core.
Mar 20 21:31:54.214904 sshd[4202]: Connection closed by 10.0.0.1 port 44646
Mar 20 21:31:54.215242 sshd-session[4200]: pam_unix(sshd:session): session closed for user core
Mar 20 21:31:54.219288 systemd[1]: sshd@22-10.0.0.132:22-10.0.0.1:44646.service: Deactivated successfully.
Mar 20 21:31:54.221523 systemd[1]: session-23.scope: Deactivated successfully.
Mar 20 21:31:54.222223 systemd-logind[1487]: Session 23 logged out. Waiting for processes to exit.
Mar 20 21:31:54.223060 systemd-logind[1487]: Removed session 23.
Mar 20 21:31:59.231213 systemd[1]: Started sshd@23-10.0.0.132:22-10.0.0.1:48870.service - OpenSSH per-connection server daemon (10.0.0.1:48870).
Mar 20 21:31:59.277858 sshd[4215]: Accepted publickey for core from 10.0.0.1 port 48870 ssh2: RSA SHA256:KJ7ck8imsv1/sWVS7eR1M7V7NSskkAYjKibngyOtAC0
Mar 20 21:31:59.279229 sshd-session[4215]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Mar 20 21:31:59.283383 systemd-logind[1487]: New session 24 of user core.
Mar 20 21:31:59.293716 systemd[1]: Started session-24.scope - Session 24 of User core.
Mar 20 21:31:59.395973 sshd[4217]: Connection closed by 10.0.0.1 port 48870
Mar 20 21:31:59.396320 sshd-session[4215]: pam_unix(sshd:session): session closed for user core
Mar 20 21:31:59.412345 systemd[1]: sshd@23-10.0.0.132:22-10.0.0.1:48870.service: Deactivated successfully.
Mar 20 21:31:59.414365 systemd[1]: session-24.scope: Deactivated successfully.
Mar 20 21:31:59.415740 systemd-logind[1487]: Session 24 logged out. Waiting for processes to exit.
Mar 20 21:31:59.417419 systemd[1]: Started sshd@24-10.0.0.132:22-10.0.0.1:48884.service - OpenSSH per-connection server daemon (10.0.0.1:48884).
Mar 20 21:31:59.418169 systemd-logind[1487]: Removed session 24.
Mar 20 21:31:59.463746 sshd[4229]: Accepted publickey for core from 10.0.0.1 port 48884 ssh2: RSA SHA256:KJ7ck8imsv1/sWVS7eR1M7V7NSskkAYjKibngyOtAC0
Mar 20 21:31:59.465109 sshd-session[4229]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Mar 20 21:31:59.469182 systemd-logind[1487]: New session 25 of user core.
Mar 20 21:31:59.478732 systemd[1]: Started session-25.scope - Session 25 of User core.
Mar 20 21:32:01.148281 containerd[1505]: time="2025-03-20T21:32:01.147922393Z" level=info msg="TaskExit event in podsandbox handler container_id:\"9d0628577cc16ff5b8bfd1c78c11a5ff4cfc3ff46fc5f579fb65f7339975cef2\"  id:\"46cb5d2cb4a7b526bc3033eb62f7684f9b564b18f7ca74b6556c8b5476053f98\"  pid:4251  exited_at:{seconds:1742506321  nanos:147448231}"
Mar 20 21:32:01.150554 containerd[1505]: time="2025-03-20T21:32:01.150129436Z" level=info msg="StopContainer for \"9d0628577cc16ff5b8bfd1c78c11a5ff4cfc3ff46fc5f579fb65f7339975cef2\" with timeout 2 (s)"
Mar 20 21:32:01.151054 containerd[1505]: time="2025-03-20T21:32:01.151012355Z" level=info msg="Stop container \"9d0628577cc16ff5b8bfd1c78c11a5ff4cfc3ff46fc5f579fb65f7339975cef2\" with signal terminated"
Mar 20 21:32:01.151541 containerd[1505]: time="2025-03-20T21:32:01.151501144Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE        \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config"
Mar 20 21:32:01.154163 containerd[1505]: time="2025-03-20T21:32:01.153788232Z" level=info msg="StopContainer for \"3d97c67b7858b412ccbd97394b377bd9d015d155fc32e9e6a3f745ad4db55eee\" with timeout 30 (s)"
Mar 20 21:32:01.155600 containerd[1505]: time="2025-03-20T21:32:01.155544780Z" level=info msg="Stop container \"3d97c67b7858b412ccbd97394b377bd9d015d155fc32e9e6a3f745ad4db55eee\" with signal terminated"
Mar 20 21:32:01.158309 systemd-networkd[1424]: lxc_health: Link DOWN
Mar 20 21:32:01.158319 systemd-networkd[1424]: lxc_health: Lost carrier
Mar 20 21:32:01.169558 systemd[1]: cri-containerd-3d97c67b7858b412ccbd97394b377bd9d015d155fc32e9e6a3f745ad4db55eee.scope: Deactivated successfully.
Mar 20 21:32:01.171698 containerd[1505]: time="2025-03-20T21:32:01.171663742Z" level=info msg="received exit event container_id:\"3d97c67b7858b412ccbd97394b377bd9d015d155fc32e9e6a3f745ad4db55eee\"  id:\"3d97c67b7858b412ccbd97394b377bd9d015d155fc32e9e6a3f745ad4db55eee\"  pid:3012  exited_at:{seconds:1742506321  nanos:171390287}"
Mar 20 21:32:01.171834 containerd[1505]: time="2025-03-20T21:32:01.171806236Z" level=info msg="TaskExit event in podsandbox handler container_id:\"3d97c67b7858b412ccbd97394b377bd9d015d155fc32e9e6a3f745ad4db55eee\"  id:\"3d97c67b7858b412ccbd97394b377bd9d015d155fc32e9e6a3f745ad4db55eee\"  pid:3012  exited_at:{seconds:1742506321  nanos:171390287}"
Mar 20 21:32:01.182941 systemd[1]: cri-containerd-9d0628577cc16ff5b8bfd1c78c11a5ff4cfc3ff46fc5f579fb65f7339975cef2.scope: Deactivated successfully.
Mar 20 21:32:01.183288 systemd[1]: cri-containerd-9d0628577cc16ff5b8bfd1c78c11a5ff4cfc3ff46fc5f579fb65f7339975cef2.scope: Consumed 6.699s CPU time, 122M memory peak, 156K read from disk, 13.3M written to disk.
Mar 20 21:32:01.184975 containerd[1505]: time="2025-03-20T21:32:01.184813016Z" level=info msg="received exit event container_id:\"9d0628577cc16ff5b8bfd1c78c11a5ff4cfc3ff46fc5f579fb65f7339975cef2\"  id:\"9d0628577cc16ff5b8bfd1c78c11a5ff4cfc3ff46fc5f579fb65f7339975cef2\"  pid:3284  exited_at:{seconds:1742506321  nanos:184353492}"
Mar 20 21:32:01.185261 containerd[1505]: time="2025-03-20T21:32:01.185226581Z" level=info msg="TaskExit event in podsandbox handler container_id:\"9d0628577cc16ff5b8bfd1c78c11a5ff4cfc3ff46fc5f579fb65f7339975cef2\"  id:\"9d0628577cc16ff5b8bfd1c78c11a5ff4cfc3ff46fc5f579fb65f7339975cef2\"  pid:3284  exited_at:{seconds:1742506321  nanos:184353492}"
Mar 20 21:32:01.193699 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-3d97c67b7858b412ccbd97394b377bd9d015d155fc32e9e6a3f745ad4db55eee-rootfs.mount: Deactivated successfully.
Mar 20 21:32:01.205949 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-9d0628577cc16ff5b8bfd1c78c11a5ff4cfc3ff46fc5f579fb65f7339975cef2-rootfs.mount: Deactivated successfully.
Mar 20 21:32:01.211211 containerd[1505]: time="2025-03-20T21:32:01.211170746Z" level=info msg="StopContainer for \"9d0628577cc16ff5b8bfd1c78c11a5ff4cfc3ff46fc5f579fb65f7339975cef2\" returns successfully"
Mar 20 21:32:01.211810 containerd[1505]: time="2025-03-20T21:32:01.211768796Z" level=info msg="StopPodSandbox for \"bd74444e833f33dc7dc44cf1755614cabdcda78497ceca51625f9284ac3f126a\""
Mar 20 21:32:01.221333 containerd[1505]: time="2025-03-20T21:32:01.221273321Z" level=info msg="Container to stop \"97939959568ed3aae7628f5d62b3b233027c5c6ff1c16a18a8a2dd67839d9fbd\" must be in running or unknown state, current state \"CONTAINER_EXITED\""
Mar 20 21:32:01.221333 containerd[1505]: time="2025-03-20T21:32:01.221299842Z" level=info msg="Container to stop \"7d7d70dae2773a709a180aa31e705b7948b49abcacac51a54048a6dfb2d459fb\" must be in running or unknown state, current state \"CONTAINER_EXITED\""
Mar 20 21:32:01.221333 containerd[1505]: time="2025-03-20T21:32:01.221310212Z" level=info msg="Container to stop \"cb5cccb990530e77805b5371aa20a2d743334ab726fb5f69adfaaaa8c2c84c37\" must be in running or unknown state, current state \"CONTAINER_EXITED\""
Mar 20 21:32:01.221333 containerd[1505]: time="2025-03-20T21:32:01.221319680Z" level=info msg="Container to stop \"43a3e9daa8b637bb9ca9efd72c1f81f1f2072fba03af0871655bd90bc07a1d80\" must be in running or unknown state, current state \"CONTAINER_EXITED\""
Mar 20 21:32:01.221333 containerd[1505]: time="2025-03-20T21:32:01.221328076Z" level=info msg="Container to stop \"9d0628577cc16ff5b8bfd1c78c11a5ff4cfc3ff46fc5f579fb65f7339975cef2\" must be in running or unknown state, current state \"CONTAINER_EXITED\""
Mar 20 21:32:01.221799 containerd[1505]: time="2025-03-20T21:32:01.221709229Z" level=info msg="StopContainer for \"3d97c67b7858b412ccbd97394b377bd9d015d155fc32e9e6a3f745ad4db55eee\" returns successfully"
Mar 20 21:32:01.222362 containerd[1505]: time="2025-03-20T21:32:01.222171578Z" level=info msg="StopPodSandbox for \"512c6fd7d4ffe0f78b9b8897bf2679d5eeeaf851306b1867d446d6dfa9e49ca0\""
Mar 20 21:32:01.222362 containerd[1505]: time="2025-03-20T21:32:01.222214400Z" level=info msg="Container to stop \"3d97c67b7858b412ccbd97394b377bd9d015d155fc32e9e6a3f745ad4db55eee\" must be in running or unknown state, current state \"CONTAINER_EXITED\""
Mar 20 21:32:01.227784 systemd[1]: cri-containerd-bd74444e833f33dc7dc44cf1755614cabdcda78497ceca51625f9284ac3f126a.scope: Deactivated successfully.
Mar 20 21:32:01.230190 systemd[1]: cri-containerd-512c6fd7d4ffe0f78b9b8897bf2679d5eeeaf851306b1867d446d6dfa9e49ca0.scope: Deactivated successfully.
Mar 20 21:32:01.233545 containerd[1505]: time="2025-03-20T21:32:01.233502956Z" level=info msg="TaskExit event in podsandbox handler container_id:\"bd74444e833f33dc7dc44cf1755614cabdcda78497ceca51625f9284ac3f126a\"  id:\"bd74444e833f33dc7dc44cf1755614cabdcda78497ceca51625f9284ac3f126a\"  pid:2804  exit_status:137  exited_at:{seconds:1742506321  nanos:232965533}"
Mar 20 21:32:01.254131 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-512c6fd7d4ffe0f78b9b8897bf2679d5eeeaf851306b1867d446d6dfa9e49ca0-rootfs.mount: Deactivated successfully.
Mar 20 21:32:01.259023 containerd[1505]: time="2025-03-20T21:32:01.258921511Z" level=info msg="shim disconnected" id=512c6fd7d4ffe0f78b9b8897bf2679d5eeeaf851306b1867d446d6dfa9e49ca0 namespace=k8s.io
Mar 20 21:32:01.261758 containerd[1505]: time="2025-03-20T21:32:01.260054439Z" level=warning msg="cleaning up after shim disconnected" id=512c6fd7d4ffe0f78b9b8897bf2679d5eeeaf851306b1867d446d6dfa9e49ca0 namespace=k8s.io
Mar 20 21:32:01.262076 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-bd74444e833f33dc7dc44cf1755614cabdcda78497ceca51625f9284ac3f126a-rootfs.mount: Deactivated successfully.
Mar 20 21:32:01.264308 containerd[1505]: time="2025-03-20T21:32:01.260074407Z" level=info msg="cleaning up dead shim" namespace=k8s.io
Mar 20 21:32:01.265511 containerd[1505]: time="2025-03-20T21:32:01.265481885Z" level=info msg="shim disconnected" id=bd74444e833f33dc7dc44cf1755614cabdcda78497ceca51625f9284ac3f126a namespace=k8s.io
Mar 20 21:32:01.265740 containerd[1505]: time="2025-03-20T21:32:01.265605151Z" level=warning msg="cleaning up after shim disconnected" id=bd74444e833f33dc7dc44cf1755614cabdcda78497ceca51625f9284ac3f126a namespace=k8s.io
Mar 20 21:32:01.265849 containerd[1505]: time="2025-03-20T21:32:01.265723138Z" level=info msg="cleaning up dead shim" namespace=k8s.io
Mar 20 21:32:01.285532 containerd[1505]: time="2025-03-20T21:32:01.285418067Z" level=info msg="TaskExit event in podsandbox handler container_id:\"512c6fd7d4ffe0f78b9b8897bf2679d5eeeaf851306b1867d446d6dfa9e49ca0\"  id:\"512c6fd7d4ffe0f78b9b8897bf2679d5eeeaf851306b1867d446d6dfa9e49ca0\"  pid:2718  exit_status:137  exited_at:{seconds:1742506321  nanos:233154145}"
Mar 20 21:32:01.287462 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-bd74444e833f33dc7dc44cf1755614cabdcda78497ceca51625f9284ac3f126a-shm.mount: Deactivated successfully.
Mar 20 21:32:01.287586 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-512c6fd7d4ffe0f78b9b8897bf2679d5eeeaf851306b1867d446d6dfa9e49ca0-shm.mount: Deactivated successfully.
Mar 20 21:32:01.294564 containerd[1505]: time="2025-03-20T21:32:01.294520540Z" level=info msg="TearDown network for sandbox \"512c6fd7d4ffe0f78b9b8897bf2679d5eeeaf851306b1867d446d6dfa9e49ca0\" successfully"
Mar 20 21:32:01.294564 containerd[1505]: time="2025-03-20T21:32:01.294557240Z" level=info msg="StopPodSandbox for \"512c6fd7d4ffe0f78b9b8897bf2679d5eeeaf851306b1867d446d6dfa9e49ca0\" returns successfully"
Mar 20 21:32:01.296886 containerd[1505]: time="2025-03-20T21:32:01.296847994Z" level=info msg="TearDown network for sandbox \"bd74444e833f33dc7dc44cf1755614cabdcda78497ceca51625f9284ac3f126a\" successfully"
Mar 20 21:32:01.296935 containerd[1505]: time="2025-03-20T21:32:01.296886459Z" level=info msg="StopPodSandbox for \"bd74444e833f33dc7dc44cf1755614cabdcda78497ceca51625f9284ac3f126a\" returns successfully"
Mar 20 21:32:01.301751 containerd[1505]: time="2025-03-20T21:32:01.301707078Z" level=info msg="received exit event sandbox_id:\"512c6fd7d4ffe0f78b9b8897bf2679d5eeeaf851306b1867d446d6dfa9e49ca0\"  exit_status:137  exited_at:{seconds:1742506321  nanos:233154145}"
Mar 20 21:32:01.301878 containerd[1505]: time="2025-03-20T21:32:01.301848901Z" level=info msg="received exit event sandbox_id:\"bd74444e833f33dc7dc44cf1755614cabdcda78497ceca51625f9284ac3f126a\"  exit_status:137  exited_at:{seconds:1742506321  nanos:232965533}"
Mar 20 21:32:01.438315 kubelet[2605]: I0320 21:32:01.438162    2605 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4gzjk\" (UniqueName: \"kubernetes.io/projected/2edb670f-4b11-4193-9194-56522899f17d-kube-api-access-4gzjk\") pod \"2edb670f-4b11-4193-9194-56522899f17d\" (UID: \"2edb670f-4b11-4193-9194-56522899f17d\") "
Mar 20 21:32:01.438315 kubelet[2605]: I0320 21:32:01.438205    2605 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/2edb670f-4b11-4193-9194-56522899f17d-cilium-config-path\") pod \"2edb670f-4b11-4193-9194-56522899f17d\" (UID: \"2edb670f-4b11-4193-9194-56522899f17d\") "
Mar 20 21:32:01.438315 kubelet[2605]: I0320 21:32:01.438224    2605 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-bpf-maps\") pod \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\" (UID: \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\") "
Mar 20 21:32:01.438315 kubelet[2605]: I0320 21:32:01.438251    2605 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-xtables-lock\") pod \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\" (UID: \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\") "
Mar 20 21:32:01.438315 kubelet[2605]: I0320 21:32:01.438265    2605 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-host-proc-sys-net\") pod \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\" (UID: \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\") "
Mar 20 21:32:01.438315 kubelet[2605]: I0320 21:32:01.438278    2605 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-etc-cni-netd\") pod \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\" (UID: \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\") "
Mar 20 21:32:01.438926 kubelet[2605]: I0320 21:32:01.438292    2605 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-cilium-config-path\") pod \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\" (UID: \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\") "
Mar 20 21:32:01.438926 kubelet[2605]: I0320 21:32:01.438305    2605 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhlh9\" (UniqueName: \"kubernetes.io/projected/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-kube-api-access-zhlh9\") pod \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\" (UID: \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\") "
Mar 20 21:32:01.438926 kubelet[2605]: I0320 21:32:01.438320    2605 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-hostproc\") pod \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\" (UID: \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\") "
Mar 20 21:32:01.438926 kubelet[2605]: I0320 21:32:01.438336    2605 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-clustermesh-secrets\") pod \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\" (UID: \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\") "
Mar 20 21:32:01.438926 kubelet[2605]: I0320 21:32:01.438353    2605 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-cni-path\") pod \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\" (UID: \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\") "
Mar 20 21:32:01.438926 kubelet[2605]: I0320 21:32:01.438368    2605 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-lib-modules\") pod \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\" (UID: \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\") "
Mar 20 21:32:01.439091 kubelet[2605]: I0320 21:32:01.438384    2605 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-hubble-tls\") pod \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\" (UID: \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\") "
Mar 20 21:32:01.439091 kubelet[2605]: I0320 21:32:01.438398    2605 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-cilium-cgroup\") pod \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\" (UID: \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\") "
Mar 20 21:32:01.439091 kubelet[2605]: I0320 21:32:01.438411    2605 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-host-proc-sys-kernel\") pod \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\" (UID: \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\") "
Mar 20 21:32:01.439091 kubelet[2605]: I0320 21:32:01.438424    2605 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-cilium-run\") pod \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\" (UID: \"80e00c81-b62f-4915-bd2e-8b9b6e12ff4c\") "
Mar 20 21:32:01.439091 kubelet[2605]: I0320 21:32:01.438490    2605 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "80e00c81-b62f-4915-bd2e-8b9b6e12ff4c" (UID: "80e00c81-b62f-4915-bd2e-8b9b6e12ff4c"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Mar 20 21:32:01.439091 kubelet[2605]: I0320 21:32:01.438525    2605 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "80e00c81-b62f-4915-bd2e-8b9b6e12ff4c" (UID: "80e00c81-b62f-4915-bd2e-8b9b6e12ff4c"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Mar 20 21:32:01.439265 kubelet[2605]: I0320 21:32:01.438540    2605 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "80e00c81-b62f-4915-bd2e-8b9b6e12ff4c" (UID: "80e00c81-b62f-4915-bd2e-8b9b6e12ff4c"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Mar 20 21:32:01.439265 kubelet[2605]: I0320 21:32:01.438555    2605 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "80e00c81-b62f-4915-bd2e-8b9b6e12ff4c" (UID: "80e00c81-b62f-4915-bd2e-8b9b6e12ff4c"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Mar 20 21:32:01.439265 kubelet[2605]: I0320 21:32:01.438568    2605 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "80e00c81-b62f-4915-bd2e-8b9b6e12ff4c" (UID: "80e00c81-b62f-4915-bd2e-8b9b6e12ff4c"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Mar 20 21:32:01.439265 kubelet[2605]: I0320 21:32:01.438600    2605 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-cni-path" (OuterVolumeSpecName: "cni-path") pod "80e00c81-b62f-4915-bd2e-8b9b6e12ff4c" (UID: "80e00c81-b62f-4915-bd2e-8b9b6e12ff4c"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Mar 20 21:32:01.442723 kubelet[2605]: I0320 21:32:01.442472    2605 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2edb670f-4b11-4193-9194-56522899f17d-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "2edb670f-4b11-4193-9194-56522899f17d" (UID: "2edb670f-4b11-4193-9194-56522899f17d"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Mar 20 21:32:01.442723 kubelet[2605]: I0320 21:32:01.442723    2605 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "80e00c81-b62f-4915-bd2e-8b9b6e12ff4c" (UID: "80e00c81-b62f-4915-bd2e-8b9b6e12ff4c"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Mar 20 21:32:01.442860 kubelet[2605]: I0320 21:32:01.442751    2605 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "80e00c81-b62f-4915-bd2e-8b9b6e12ff4c" (UID: "80e00c81-b62f-4915-bd2e-8b9b6e12ff4c"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Mar 20 21:32:01.442860 kubelet[2605]: I0320 21:32:01.442767    2605 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "80e00c81-b62f-4915-bd2e-8b9b6e12ff4c" (UID: "80e00c81-b62f-4915-bd2e-8b9b6e12ff4c"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Mar 20 21:32:01.442860 kubelet[2605]: I0320 21:32:01.442784    2605 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "80e00c81-b62f-4915-bd2e-8b9b6e12ff4c" (UID: "80e00c81-b62f-4915-bd2e-8b9b6e12ff4c"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Mar 20 21:32:01.442860 kubelet[2605]: I0320 21:32:01.442798    2605 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-hostproc" (OuterVolumeSpecName: "hostproc") pod "80e00c81-b62f-4915-bd2e-8b9b6e12ff4c" (UID: "80e00c81-b62f-4915-bd2e-8b9b6e12ff4c"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Mar 20 21:32:01.444376 kubelet[2605]: I0320 21:32:01.444353    2605 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2edb670f-4b11-4193-9194-56522899f17d-kube-api-access-4gzjk" (OuterVolumeSpecName: "kube-api-access-4gzjk") pod "2edb670f-4b11-4193-9194-56522899f17d" (UID: "2edb670f-4b11-4193-9194-56522899f17d"). InnerVolumeSpecName "kube-api-access-4gzjk". PluginName "kubernetes.io/projected", VolumeGidValue ""
Mar 20 21:32:01.444670 kubelet[2605]: I0320 21:32:01.444634    2605 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-kube-api-access-zhlh9" (OuterVolumeSpecName: "kube-api-access-zhlh9") pod "80e00c81-b62f-4915-bd2e-8b9b6e12ff4c" (UID: "80e00c81-b62f-4915-bd2e-8b9b6e12ff4c"). InnerVolumeSpecName "kube-api-access-zhlh9". PluginName "kubernetes.io/projected", VolumeGidValue ""
Mar 20 21:32:01.445414 kubelet[2605]: I0320 21:32:01.445393    2605 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "80e00c81-b62f-4915-bd2e-8b9b6e12ff4c" (UID: "80e00c81-b62f-4915-bd2e-8b9b6e12ff4c"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue ""
Mar 20 21:32:01.445797 kubelet[2605]: I0320 21:32:01.445722    2605 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "80e00c81-b62f-4915-bd2e-8b9b6e12ff4c" (UID: "80e00c81-b62f-4915-bd2e-8b9b6e12ff4c"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue ""
Mar 20 21:32:01.450067 kubelet[2605]: I0320 21:32:01.449963    2605 scope.go:117] "RemoveContainer" containerID="9d0628577cc16ff5b8bfd1c78c11a5ff4cfc3ff46fc5f579fb65f7339975cef2"
Mar 20 21:32:01.452871 containerd[1505]: time="2025-03-20T21:32:01.452247044Z" level=info msg="RemoveContainer for \"9d0628577cc16ff5b8bfd1c78c11a5ff4cfc3ff46fc5f579fb65f7339975cef2\""
Mar 20 21:32:01.456414 systemd[1]: Removed slice kubepods-burstable-pod80e00c81_b62f_4915_bd2e_8b9b6e12ff4c.slice - libcontainer container kubepods-burstable-pod80e00c81_b62f_4915_bd2e_8b9b6e12ff4c.slice.
Mar 20 21:32:01.456508 systemd[1]: kubepods-burstable-pod80e00c81_b62f_4915_bd2e_8b9b6e12ff4c.slice: Consumed 6.806s CPU time, 122.4M memory peak, 168K read from disk, 13.3M written to disk.
Mar 20 21:32:01.457719 systemd[1]: Removed slice kubepods-besteffort-pod2edb670f_4b11_4193_9194_56522899f17d.slice - libcontainer container kubepods-besteffort-pod2edb670f_4b11_4193_9194_56522899f17d.slice.
Mar 20 21:32:01.474341 containerd[1505]: time="2025-03-20T21:32:01.474291730Z" level=info msg="RemoveContainer for \"9d0628577cc16ff5b8bfd1c78c11a5ff4cfc3ff46fc5f579fb65f7339975cef2\" returns successfully"
Mar 20 21:32:01.474575 kubelet[2605]: I0320 21:32:01.474547    2605 scope.go:117] "RemoveContainer" containerID="7d7d70dae2773a709a180aa31e705b7948b49abcacac51a54048a6dfb2d459fb"
Mar 20 21:32:01.476358 containerd[1505]: time="2025-03-20T21:32:01.476317265Z" level=info msg="RemoveContainer for \"7d7d70dae2773a709a180aa31e705b7948b49abcacac51a54048a6dfb2d459fb\""
Mar 20 21:32:01.480695 containerd[1505]: time="2025-03-20T21:32:01.480665155Z" level=info msg="RemoveContainer for \"7d7d70dae2773a709a180aa31e705b7948b49abcacac51a54048a6dfb2d459fb\" returns successfully"
Mar 20 21:32:01.480824 kubelet[2605]: I0320 21:32:01.480800    2605 scope.go:117] "RemoveContainer" containerID="43a3e9daa8b637bb9ca9efd72c1f81f1f2072fba03af0871655bd90bc07a1d80"
Mar 20 21:32:01.482542 containerd[1505]: time="2025-03-20T21:32:01.482522126Z" level=info msg="RemoveContainer for \"43a3e9daa8b637bb9ca9efd72c1f81f1f2072fba03af0871655bd90bc07a1d80\""
Mar 20 21:32:01.486506 containerd[1505]: time="2025-03-20T21:32:01.486478983Z" level=info msg="RemoveContainer for \"43a3e9daa8b637bb9ca9efd72c1f81f1f2072fba03af0871655bd90bc07a1d80\" returns successfully"
Mar 20 21:32:01.486650 kubelet[2605]: I0320 21:32:01.486631    2605 scope.go:117] "RemoveContainer" containerID="97939959568ed3aae7628f5d62b3b233027c5c6ff1c16a18a8a2dd67839d9fbd"
Mar 20 21:32:01.487992 containerd[1505]: time="2025-03-20T21:32:01.487968679Z" level=info msg="RemoveContainer for \"97939959568ed3aae7628f5d62b3b233027c5c6ff1c16a18a8a2dd67839d9fbd\""
Mar 20 21:32:01.492642 containerd[1505]: time="2025-03-20T21:32:01.491781319Z" level=info msg="RemoveContainer for \"97939959568ed3aae7628f5d62b3b233027c5c6ff1c16a18a8a2dd67839d9fbd\" returns successfully"
Mar 20 21:32:01.492819 kubelet[2605]: I0320 21:32:01.492761    2605 scope.go:117] "RemoveContainer" containerID="cb5cccb990530e77805b5371aa20a2d743334ab726fb5f69adfaaaa8c2c84c37"
Mar 20 21:32:01.494397 containerd[1505]: time="2025-03-20T21:32:01.494356170Z" level=info msg="RemoveContainer for \"cb5cccb990530e77805b5371aa20a2d743334ab726fb5f69adfaaaa8c2c84c37\""
Mar 20 21:32:01.497801 containerd[1505]: time="2025-03-20T21:32:01.497781305Z" level=info msg="RemoveContainer for \"cb5cccb990530e77805b5371aa20a2d743334ab726fb5f69adfaaaa8c2c84c37\" returns successfully"
Mar 20 21:32:01.498136 kubelet[2605]: I0320 21:32:01.498093    2605 scope.go:117] "RemoveContainer" containerID="9d0628577cc16ff5b8bfd1c78c11a5ff4cfc3ff46fc5f579fb65f7339975cef2"
Mar 20 21:32:01.500271 containerd[1505]: time="2025-03-20T21:32:01.500217129Z" level=error msg="ContainerStatus for \"9d0628577cc16ff5b8bfd1c78c11a5ff4cfc3ff46fc5f579fb65f7339975cef2\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"9d0628577cc16ff5b8bfd1c78c11a5ff4cfc3ff46fc5f579fb65f7339975cef2\": not found"
Mar 20 21:32:01.502940 kubelet[2605]: E0320 21:32:01.502887    2605 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"9d0628577cc16ff5b8bfd1c78c11a5ff4cfc3ff46fc5f579fb65f7339975cef2\": not found" containerID="9d0628577cc16ff5b8bfd1c78c11a5ff4cfc3ff46fc5f579fb65f7339975cef2"
Mar 20 21:32:01.503059 kubelet[2605]: I0320 21:32:01.502941    2605 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"9d0628577cc16ff5b8bfd1c78c11a5ff4cfc3ff46fc5f579fb65f7339975cef2"} err="failed to get container status \"9d0628577cc16ff5b8bfd1c78c11a5ff4cfc3ff46fc5f579fb65f7339975cef2\": rpc error: code = NotFound desc = an error occurred when try to find container \"9d0628577cc16ff5b8bfd1c78c11a5ff4cfc3ff46fc5f579fb65f7339975cef2\": not found"
Mar 20 21:32:01.503059 kubelet[2605]: I0320 21:32:01.503023    2605 scope.go:117] "RemoveContainer" containerID="7d7d70dae2773a709a180aa31e705b7948b49abcacac51a54048a6dfb2d459fb"
Mar 20 21:32:01.503233 containerd[1505]: time="2025-03-20T21:32:01.503204503Z" level=error msg="ContainerStatus for \"7d7d70dae2773a709a180aa31e705b7948b49abcacac51a54048a6dfb2d459fb\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"7d7d70dae2773a709a180aa31e705b7948b49abcacac51a54048a6dfb2d459fb\": not found"
Mar 20 21:32:01.503473 kubelet[2605]: E0320 21:32:01.503332    2605 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"7d7d70dae2773a709a180aa31e705b7948b49abcacac51a54048a6dfb2d459fb\": not found" containerID="7d7d70dae2773a709a180aa31e705b7948b49abcacac51a54048a6dfb2d459fb"
Mar 20 21:32:01.503473 kubelet[2605]: I0320 21:32:01.503365    2605 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"7d7d70dae2773a709a180aa31e705b7948b49abcacac51a54048a6dfb2d459fb"} err="failed to get container status \"7d7d70dae2773a709a180aa31e705b7948b49abcacac51a54048a6dfb2d459fb\": rpc error: code = NotFound desc = an error occurred when try to find container \"7d7d70dae2773a709a180aa31e705b7948b49abcacac51a54048a6dfb2d459fb\": not found"
Mar 20 21:32:01.503473 kubelet[2605]: I0320 21:32:01.503387    2605 scope.go:117] "RemoveContainer" containerID="43a3e9daa8b637bb9ca9efd72c1f81f1f2072fba03af0871655bd90bc07a1d80"
Mar 20 21:32:01.503719 containerd[1505]: time="2025-03-20T21:32:01.503665900Z" level=error msg="ContainerStatus for \"43a3e9daa8b637bb9ca9efd72c1f81f1f2072fba03af0871655bd90bc07a1d80\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"43a3e9daa8b637bb9ca9efd72c1f81f1f2072fba03af0871655bd90bc07a1d80\": not found"
Mar 20 21:32:01.503829 kubelet[2605]: E0320 21:32:01.503807    2605 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"43a3e9daa8b637bb9ca9efd72c1f81f1f2072fba03af0871655bd90bc07a1d80\": not found" containerID="43a3e9daa8b637bb9ca9efd72c1f81f1f2072fba03af0871655bd90bc07a1d80"
Mar 20 21:32:01.503872 kubelet[2605]: I0320 21:32:01.503832    2605 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"43a3e9daa8b637bb9ca9efd72c1f81f1f2072fba03af0871655bd90bc07a1d80"} err="failed to get container status \"43a3e9daa8b637bb9ca9efd72c1f81f1f2072fba03af0871655bd90bc07a1d80\": rpc error: code = NotFound desc = an error occurred when try to find container \"43a3e9daa8b637bb9ca9efd72c1f81f1f2072fba03af0871655bd90bc07a1d80\": not found"
Mar 20 21:32:01.503872 kubelet[2605]: I0320 21:32:01.503848    2605 scope.go:117] "RemoveContainer" containerID="97939959568ed3aae7628f5d62b3b233027c5c6ff1c16a18a8a2dd67839d9fbd"
Mar 20 21:32:01.504113 containerd[1505]: time="2025-03-20T21:32:01.504046051Z" level=error msg="ContainerStatus for \"97939959568ed3aae7628f5d62b3b233027c5c6ff1c16a18a8a2dd67839d9fbd\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"97939959568ed3aae7628f5d62b3b233027c5c6ff1c16a18a8a2dd67839d9fbd\": not found"
Mar 20 21:32:01.504204 kubelet[2605]: E0320 21:32:01.504180    2605 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"97939959568ed3aae7628f5d62b3b233027c5c6ff1c16a18a8a2dd67839d9fbd\": not found" containerID="97939959568ed3aae7628f5d62b3b233027c5c6ff1c16a18a8a2dd67839d9fbd"
Mar 20 21:32:01.504254 kubelet[2605]: I0320 21:32:01.504210    2605 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"97939959568ed3aae7628f5d62b3b233027c5c6ff1c16a18a8a2dd67839d9fbd"} err="failed to get container status \"97939959568ed3aae7628f5d62b3b233027c5c6ff1c16a18a8a2dd67839d9fbd\": rpc error: code = NotFound desc = an error occurred when try to find container \"97939959568ed3aae7628f5d62b3b233027c5c6ff1c16a18a8a2dd67839d9fbd\": not found"
Mar 20 21:32:01.504254 kubelet[2605]: I0320 21:32:01.504231    2605 scope.go:117] "RemoveContainer" containerID="cb5cccb990530e77805b5371aa20a2d743334ab726fb5f69adfaaaa8c2c84c37"
Mar 20 21:32:01.504451 containerd[1505]: time="2025-03-20T21:32:01.504419730Z" level=error msg="ContainerStatus for \"cb5cccb990530e77805b5371aa20a2d743334ab726fb5f69adfaaaa8c2c84c37\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"cb5cccb990530e77805b5371aa20a2d743334ab726fb5f69adfaaaa8c2c84c37\": not found"
Mar 20 21:32:01.504546 kubelet[2605]: E0320 21:32:01.504526    2605 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"cb5cccb990530e77805b5371aa20a2d743334ab726fb5f69adfaaaa8c2c84c37\": not found" containerID="cb5cccb990530e77805b5371aa20a2d743334ab726fb5f69adfaaaa8c2c84c37"
Mar 20 21:32:01.504576 kubelet[2605]: I0320 21:32:01.504547    2605 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"cb5cccb990530e77805b5371aa20a2d743334ab726fb5f69adfaaaa8c2c84c37"} err="failed to get container status \"cb5cccb990530e77805b5371aa20a2d743334ab726fb5f69adfaaaa8c2c84c37\": rpc error: code = NotFound desc = an error occurred when try to find container \"cb5cccb990530e77805b5371aa20a2d743334ab726fb5f69adfaaaa8c2c84c37\": not found"
Mar 20 21:32:01.504576 kubelet[2605]: I0320 21:32:01.504559    2605 scope.go:117] "RemoveContainer" containerID="3d97c67b7858b412ccbd97394b377bd9d015d155fc32e9e6a3f745ad4db55eee"
Mar 20 21:32:01.507447 containerd[1505]: time="2025-03-20T21:32:01.507415450Z" level=info msg="RemoveContainer for \"3d97c67b7858b412ccbd97394b377bd9d015d155fc32e9e6a3f745ad4db55eee\""
Mar 20 21:32:01.511275 containerd[1505]: time="2025-03-20T21:32:01.511229293Z" level=info msg="RemoveContainer for \"3d97c67b7858b412ccbd97394b377bd9d015d155fc32e9e6a3f745ad4db55eee\" returns successfully"
Mar 20 21:32:01.511488 kubelet[2605]: I0320 21:32:01.511406    2605 scope.go:117] "RemoveContainer" containerID="3d97c67b7858b412ccbd97394b377bd9d015d155fc32e9e6a3f745ad4db55eee"
Mar 20 21:32:01.511647 containerd[1505]: time="2025-03-20T21:32:01.511567534Z" level=error msg="ContainerStatus for \"3d97c67b7858b412ccbd97394b377bd9d015d155fc32e9e6a3f745ad4db55eee\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"3d97c67b7858b412ccbd97394b377bd9d015d155fc32e9e6a3f745ad4db55eee\": not found"
Mar 20 21:32:01.511718 kubelet[2605]: E0320 21:32:01.511693    2605 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"3d97c67b7858b412ccbd97394b377bd9d015d155fc32e9e6a3f745ad4db55eee\": not found" containerID="3d97c67b7858b412ccbd97394b377bd9d015d155fc32e9e6a3f745ad4db55eee"
Mar 20 21:32:01.511750 kubelet[2605]: I0320 21:32:01.511719    2605 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"3d97c67b7858b412ccbd97394b377bd9d015d155fc32e9e6a3f745ad4db55eee"} err="failed to get container status \"3d97c67b7858b412ccbd97394b377bd9d015d155fc32e9e6a3f745ad4db55eee\": rpc error: code = NotFound desc = an error occurred when try to find container \"3d97c67b7858b412ccbd97394b377bd9d015d155fc32e9e6a3f745ad4db55eee\": not found"
Mar 20 21:32:01.538910 kubelet[2605]: I0320 21:32:01.538891    2605 reconciler_common.go:288] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-xtables-lock\") on node \"localhost\" DevicePath \"\""
Mar 20 21:32:01.538910 kubelet[2605]: I0320 21:32:01.538907    2605 reconciler_common.go:288] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-host-proc-sys-net\") on node \"localhost\" DevicePath \"\""
Mar 20 21:32:01.538910 kubelet[2605]: I0320 21:32:01.538918    2605 reconciler_common.go:288] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-cilium-config-path\") on node \"localhost\" DevicePath \"\""
Mar 20 21:32:01.539072 kubelet[2605]: I0320 21:32:01.538926    2605 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-zhlh9\" (UniqueName: \"kubernetes.io/projected/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-kube-api-access-zhlh9\") on node \"localhost\" DevicePath \"\""
Mar 20 21:32:01.539072 kubelet[2605]: I0320 21:32:01.538935    2605 reconciler_common.go:288] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-hostproc\") on node \"localhost\" DevicePath \"\""
Mar 20 21:32:01.539072 kubelet[2605]: I0320 21:32:01.538943    2605 reconciler_common.go:288] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-etc-cni-netd\") on node \"localhost\" DevicePath \"\""
Mar 20 21:32:01.539072 kubelet[2605]: I0320 21:32:01.538950    2605 reconciler_common.go:288] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-clustermesh-secrets\") on node \"localhost\" DevicePath \"\""
Mar 20 21:32:01.539072 kubelet[2605]: I0320 21:32:01.538959    2605 reconciler_common.go:288] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-lib-modules\") on node \"localhost\" DevicePath \"\""
Mar 20 21:32:01.539072 kubelet[2605]: I0320 21:32:01.538966    2605 reconciler_common.go:288] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-hubble-tls\") on node \"localhost\" DevicePath \"\""
Mar 20 21:32:01.539072 kubelet[2605]: I0320 21:32:01.538975    2605 reconciler_common.go:288] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-cilium-cgroup\") on node \"localhost\" DevicePath \"\""
Mar 20 21:32:01.539072 kubelet[2605]: I0320 21:32:01.538983    2605 reconciler_common.go:288] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-host-proc-sys-kernel\") on node \"localhost\" DevicePath \"\""
Mar 20 21:32:01.539258 kubelet[2605]: I0320 21:32:01.538991    2605 reconciler_common.go:288] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-cilium-run\") on node \"localhost\" DevicePath \"\""
Mar 20 21:32:01.539258 kubelet[2605]: I0320 21:32:01.539000    2605 reconciler_common.go:288] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-cni-path\") on node \"localhost\" DevicePath \"\""
Mar 20 21:32:01.539258 kubelet[2605]: I0320 21:32:01.539008    2605 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-4gzjk\" (UniqueName: \"kubernetes.io/projected/2edb670f-4b11-4193-9194-56522899f17d-kube-api-access-4gzjk\") on node \"localhost\" DevicePath \"\""
Mar 20 21:32:01.539258 kubelet[2605]: I0320 21:32:01.539017    2605 reconciler_common.go:288] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/2edb670f-4b11-4193-9194-56522899f17d-cilium-config-path\") on node \"localhost\" DevicePath \"\""
Mar 20 21:32:01.539258 kubelet[2605]: I0320 21:32:01.539026    2605 reconciler_common.go:288] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c-bpf-maps\") on node \"localhost\" DevicePath \"\""
Mar 20 21:32:01.807483 kubelet[2605]: I0320 21:32:01.807436    2605 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2edb670f-4b11-4193-9194-56522899f17d" path="/var/lib/kubelet/pods/2edb670f-4b11-4193-9194-56522899f17d/volumes"
Mar 20 21:32:01.808078 kubelet[2605]: I0320 21:32:01.808049    2605 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80e00c81-b62f-4915-bd2e-8b9b6e12ff4c" path="/var/lib/kubelet/pods/80e00c81-b62f-4915-bd2e-8b9b6e12ff4c/volumes"
Mar 20 21:32:01.856099 kubelet[2605]: E0320 21:32:01.856065    2605 kubelet.go:2901] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"
Mar 20 21:32:02.194053 systemd[1]: var-lib-kubelet-pods-80e00c81\x2db62f\x2d4915\x2dbd2e\x2d8b9b6e12ff4c-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dzhlh9.mount: Deactivated successfully.
Mar 20 21:32:02.194179 systemd[1]: var-lib-kubelet-pods-2edb670f\x2d4b11\x2d4193\x2d9194\x2d56522899f17d-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d4gzjk.mount: Deactivated successfully.
Mar 20 21:32:02.194280 systemd[1]: var-lib-kubelet-pods-80e00c81\x2db62f\x2d4915\x2dbd2e\x2d8b9b6e12ff4c-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully.
Mar 20 21:32:02.194360 systemd[1]: var-lib-kubelet-pods-80e00c81\x2db62f\x2d4915\x2dbd2e\x2d8b9b6e12ff4c-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully.
Mar 20 21:32:02.769874 sshd[4232]: Connection closed by 10.0.0.1 port 48884
Mar 20 21:32:02.770469 sshd-session[4229]: pam_unix(sshd:session): session closed for user core
Mar 20 21:32:02.780438 systemd[1]: sshd@24-10.0.0.132:22-10.0.0.1:48884.service: Deactivated successfully.
Mar 20 21:32:02.782569 systemd[1]: session-25.scope: Deactivated successfully.
Mar 20 21:32:02.783428 systemd-logind[1487]: Session 25 logged out. Waiting for processes to exit.
Mar 20 21:32:02.786022 systemd[1]: Started sshd@25-10.0.0.132:22-10.0.0.1:48888.service - OpenSSH per-connection server daemon (10.0.0.1:48888).
Mar 20 21:32:02.786502 systemd-logind[1487]: Removed session 25.
Mar 20 21:32:02.833678 sshd[4378]: Accepted publickey for core from 10.0.0.1 port 48888 ssh2: RSA SHA256:KJ7ck8imsv1/sWVS7eR1M7V7NSskkAYjKibngyOtAC0
Mar 20 21:32:02.835085 sshd-session[4378]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Mar 20 21:32:02.839489 systemd-logind[1487]: New session 26 of user core.
Mar 20 21:32:02.847721 systemd[1]: Started session-26.scope - Session 26 of User core.
Mar 20 21:32:03.580131 sshd[4381]: Connection closed by 10.0.0.1 port 48888
Mar 20 21:32:03.580937 sshd-session[4378]: pam_unix(sshd:session): session closed for user core
Mar 20 21:32:03.593567 kubelet[2605]: E0320 21:32:03.591744    2605 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="80e00c81-b62f-4915-bd2e-8b9b6e12ff4c" containerName="mount-cgroup"
Mar 20 21:32:03.593567 kubelet[2605]: E0320 21:32:03.591770    2605 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="80e00c81-b62f-4915-bd2e-8b9b6e12ff4c" containerName="mount-bpf-fs"
Mar 20 21:32:03.593567 kubelet[2605]: E0320 21:32:03.591776    2605 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="80e00c81-b62f-4915-bd2e-8b9b6e12ff4c" containerName="clean-cilium-state"
Mar 20 21:32:03.593567 kubelet[2605]: E0320 21:32:03.591783    2605 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="2edb670f-4b11-4193-9194-56522899f17d" containerName="cilium-operator"
Mar 20 21:32:03.593567 kubelet[2605]: E0320 21:32:03.591789    2605 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="80e00c81-b62f-4915-bd2e-8b9b6e12ff4c" containerName="apply-sysctl-overwrites"
Mar 20 21:32:03.593567 kubelet[2605]: E0320 21:32:03.591795    2605 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="80e00c81-b62f-4915-bd2e-8b9b6e12ff4c" containerName="cilium-agent"
Mar 20 21:32:03.593567 kubelet[2605]: I0320 21:32:03.591816    2605 memory_manager.go:354] "RemoveStaleState removing state" podUID="80e00c81-b62f-4915-bd2e-8b9b6e12ff4c" containerName="cilium-agent"
Mar 20 21:32:03.593567 kubelet[2605]: I0320 21:32:03.591823    2605 memory_manager.go:354] "RemoveStaleState removing state" podUID="2edb670f-4b11-4193-9194-56522899f17d" containerName="cilium-operator"
Mar 20 21:32:03.596154 systemd[1]: sshd@25-10.0.0.132:22-10.0.0.1:48888.service: Deactivated successfully.
Mar 20 21:32:03.599254 systemd[1]: session-26.scope: Deactivated successfully.
Mar 20 21:32:03.601517 systemd-logind[1487]: Session 26 logged out. Waiting for processes to exit.
Mar 20 21:32:03.607984 systemd[1]: Started sshd@26-10.0.0.132:22-10.0.0.1:48892.service - OpenSSH per-connection server daemon (10.0.0.1:48892).
Mar 20 21:32:03.611494 systemd-logind[1487]: Removed session 26.
Mar 20 21:32:03.623494 systemd[1]: Created slice kubepods-burstable-pod306848d8_bc51_4b4c_94a3_8f52ad2d110b.slice - libcontainer container kubepods-burstable-pod306848d8_bc51_4b4c_94a3_8f52ad2d110b.slice.
Mar 20 21:32:03.651014 kubelet[2605]: I0320 21:32:03.650970    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/306848d8-bc51-4b4c-94a3-8f52ad2d110b-etc-cni-netd\") pod \"cilium-zhh5g\" (UID: \"306848d8-bc51-4b4c-94a3-8f52ad2d110b\") " pod="kube-system/cilium-zhh5g"
Mar 20 21:32:03.651014 kubelet[2605]: I0320 21:32:03.651007    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52tc4\" (UniqueName: \"kubernetes.io/projected/306848d8-bc51-4b4c-94a3-8f52ad2d110b-kube-api-access-52tc4\") pod \"cilium-zhh5g\" (UID: \"306848d8-bc51-4b4c-94a3-8f52ad2d110b\") " pod="kube-system/cilium-zhh5g"
Mar 20 21:32:03.651125 kubelet[2605]: I0320 21:32:03.651024    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/306848d8-bc51-4b4c-94a3-8f52ad2d110b-cilium-cgroup\") pod \"cilium-zhh5g\" (UID: \"306848d8-bc51-4b4c-94a3-8f52ad2d110b\") " pod="kube-system/cilium-zhh5g"
Mar 20 21:32:03.651125 kubelet[2605]: I0320 21:32:03.651043    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/306848d8-bc51-4b4c-94a3-8f52ad2d110b-lib-modules\") pod \"cilium-zhh5g\" (UID: \"306848d8-bc51-4b4c-94a3-8f52ad2d110b\") " pod="kube-system/cilium-zhh5g"
Mar 20 21:32:03.651125 kubelet[2605]: I0320 21:32:03.651057    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/306848d8-bc51-4b4c-94a3-8f52ad2d110b-hubble-tls\") pod \"cilium-zhh5g\" (UID: \"306848d8-bc51-4b4c-94a3-8f52ad2d110b\") " pod="kube-system/cilium-zhh5g"
Mar 20 21:32:03.651125 kubelet[2605]: I0320 21:32:03.651070    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/306848d8-bc51-4b4c-94a3-8f52ad2d110b-cni-path\") pod \"cilium-zhh5g\" (UID: \"306848d8-bc51-4b4c-94a3-8f52ad2d110b\") " pod="kube-system/cilium-zhh5g"
Mar 20 21:32:03.651125 kubelet[2605]: I0320 21:32:03.651082    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/306848d8-bc51-4b4c-94a3-8f52ad2d110b-host-proc-sys-kernel\") pod \"cilium-zhh5g\" (UID: \"306848d8-bc51-4b4c-94a3-8f52ad2d110b\") " pod="kube-system/cilium-zhh5g"
Mar 20 21:32:03.651125 kubelet[2605]: I0320 21:32:03.651096    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/306848d8-bc51-4b4c-94a3-8f52ad2d110b-cilium-config-path\") pod \"cilium-zhh5g\" (UID: \"306848d8-bc51-4b4c-94a3-8f52ad2d110b\") " pod="kube-system/cilium-zhh5g"
Mar 20 21:32:03.651287 kubelet[2605]: I0320 21:32:03.651110    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/306848d8-bc51-4b4c-94a3-8f52ad2d110b-host-proc-sys-net\") pod \"cilium-zhh5g\" (UID: \"306848d8-bc51-4b4c-94a3-8f52ad2d110b\") " pod="kube-system/cilium-zhh5g"
Mar 20 21:32:03.651287 kubelet[2605]: I0320 21:32:03.651130    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/306848d8-bc51-4b4c-94a3-8f52ad2d110b-xtables-lock\") pod \"cilium-zhh5g\" (UID: \"306848d8-bc51-4b4c-94a3-8f52ad2d110b\") " pod="kube-system/cilium-zhh5g"
Mar 20 21:32:03.651287 kubelet[2605]: I0320 21:32:03.651143    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/306848d8-bc51-4b4c-94a3-8f52ad2d110b-hostproc\") pod \"cilium-zhh5g\" (UID: \"306848d8-bc51-4b4c-94a3-8f52ad2d110b\") " pod="kube-system/cilium-zhh5g"
Mar 20 21:32:03.651287 kubelet[2605]: I0320 21:32:03.651158    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/306848d8-bc51-4b4c-94a3-8f52ad2d110b-cilium-ipsec-secrets\") pod \"cilium-zhh5g\" (UID: \"306848d8-bc51-4b4c-94a3-8f52ad2d110b\") " pod="kube-system/cilium-zhh5g"
Mar 20 21:32:03.651287 kubelet[2605]: I0320 21:32:03.651173    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/306848d8-bc51-4b4c-94a3-8f52ad2d110b-cilium-run\") pod \"cilium-zhh5g\" (UID: \"306848d8-bc51-4b4c-94a3-8f52ad2d110b\") " pod="kube-system/cilium-zhh5g"
Mar 20 21:32:03.651287 kubelet[2605]: I0320 21:32:03.651186    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/306848d8-bc51-4b4c-94a3-8f52ad2d110b-clustermesh-secrets\") pod \"cilium-zhh5g\" (UID: \"306848d8-bc51-4b4c-94a3-8f52ad2d110b\") " pod="kube-system/cilium-zhh5g"
Mar 20 21:32:03.651431 kubelet[2605]: I0320 21:32:03.651199    2605 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/306848d8-bc51-4b4c-94a3-8f52ad2d110b-bpf-maps\") pod \"cilium-zhh5g\" (UID: \"306848d8-bc51-4b4c-94a3-8f52ad2d110b\") " pod="kube-system/cilium-zhh5g"
Mar 20 21:32:03.656578 sshd[4392]: Accepted publickey for core from 10.0.0.1 port 48892 ssh2: RSA SHA256:KJ7ck8imsv1/sWVS7eR1M7V7NSskkAYjKibngyOtAC0
Mar 20 21:32:03.658005 sshd-session[4392]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Mar 20 21:32:03.662313 systemd-logind[1487]: New session 27 of user core.
Mar 20 21:32:03.674734 systemd[1]: Started session-27.scope - Session 27 of User core.
Mar 20 21:32:03.726036 sshd[4395]: Connection closed by 10.0.0.1 port 48892
Mar 20 21:32:03.726355 sshd-session[4392]: pam_unix(sshd:session): session closed for user core
Mar 20 21:32:03.737418 systemd[1]: sshd@26-10.0.0.132:22-10.0.0.1:48892.service: Deactivated successfully.
Mar 20 21:32:03.739604 systemd[1]: session-27.scope: Deactivated successfully.
Mar 20 21:32:03.741243 systemd-logind[1487]: Session 27 logged out. Waiting for processes to exit.
Mar 20 21:32:03.742544 systemd[1]: Started sshd@27-10.0.0.132:22-10.0.0.1:48896.service - OpenSSH per-connection server daemon (10.0.0.1:48896).
Mar 20 21:32:03.743743 systemd-logind[1487]: Removed session 27.
Mar 20 21:32:03.793062 sshd[4401]: Accepted publickey for core from 10.0.0.1 port 48896 ssh2: RSA SHA256:KJ7ck8imsv1/sWVS7eR1M7V7NSskkAYjKibngyOtAC0
Mar 20 21:32:03.794714 sshd-session[4401]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Mar 20 21:32:03.798914 systemd-logind[1487]: New session 28 of user core.
Mar 20 21:32:03.808788 systemd[1]: Started session-28.scope - Session 28 of User core.
Mar 20 21:32:03.929603 containerd[1505]: time="2025-03-20T21:32:03.929486609Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-zhh5g,Uid:306848d8-bc51-4b4c-94a3-8f52ad2d110b,Namespace:kube-system,Attempt:0,}"
Mar 20 21:32:03.944335 containerd[1505]: time="2025-03-20T21:32:03.944290340Z" level=info msg="connecting to shim 220ae988f9f95c31190f456aa63f66df0e1a4364c8e5b6cb1ebfcf9663416cba" address="unix:///run/containerd/s/9a32d66b27678ffde4afb777ceb65a97a2a525c15882991ae929825e83e11bab" namespace=k8s.io protocol=ttrpc version=3
Mar 20 21:32:03.959031 kubelet[2605]: I0320 21:32:03.958978    2605 setters.go:600] "Node became not ready" node="localhost" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-03-20T21:32:03Z","lastTransitionTime":"2025-03-20T21:32:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"}
Mar 20 21:32:03.970848 systemd[1]: Started cri-containerd-220ae988f9f95c31190f456aa63f66df0e1a4364c8e5b6cb1ebfcf9663416cba.scope - libcontainer container 220ae988f9f95c31190f456aa63f66df0e1a4364c8e5b6cb1ebfcf9663416cba.
Mar 20 21:32:03.999519 containerd[1505]: time="2025-03-20T21:32:03.999466515Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-zhh5g,Uid:306848d8-bc51-4b4c-94a3-8f52ad2d110b,Namespace:kube-system,Attempt:0,} returns sandbox id \"220ae988f9f95c31190f456aa63f66df0e1a4364c8e5b6cb1ebfcf9663416cba\""
Mar 20 21:32:04.002046 containerd[1505]: time="2025-03-20T21:32:04.002020911Z" level=info msg="CreateContainer within sandbox \"220ae988f9f95c31190f456aa63f66df0e1a4364c8e5b6cb1ebfcf9663416cba\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}"
Mar 20 21:32:04.009391 containerd[1505]: time="2025-03-20T21:32:04.009349067Z" level=info msg="Container e7c3577d9d83b277a95daf03cb0e2f93e7539bd2290926b1bae77f143c580885: CDI devices from CRI Config.CDIDevices: []"
Mar 20 21:32:04.017026 containerd[1505]: time="2025-03-20T21:32:04.016987368Z" level=info msg="CreateContainer within sandbox \"220ae988f9f95c31190f456aa63f66df0e1a4364c8e5b6cb1ebfcf9663416cba\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"e7c3577d9d83b277a95daf03cb0e2f93e7539bd2290926b1bae77f143c580885\""
Mar 20 21:32:04.017506 containerd[1505]: time="2025-03-20T21:32:04.017475685Z" level=info msg="StartContainer for \"e7c3577d9d83b277a95daf03cb0e2f93e7539bd2290926b1bae77f143c580885\""
Mar 20 21:32:04.018460 containerd[1505]: time="2025-03-20T21:32:04.018436820Z" level=info msg="connecting to shim e7c3577d9d83b277a95daf03cb0e2f93e7539bd2290926b1bae77f143c580885" address="unix:///run/containerd/s/9a32d66b27678ffde4afb777ceb65a97a2a525c15882991ae929825e83e11bab" protocol=ttrpc version=3
Mar 20 21:32:04.043753 systemd[1]: Started cri-containerd-e7c3577d9d83b277a95daf03cb0e2f93e7539bd2290926b1bae77f143c580885.scope - libcontainer container e7c3577d9d83b277a95daf03cb0e2f93e7539bd2290926b1bae77f143c580885.
Mar 20 21:32:04.072446 containerd[1505]: time="2025-03-20T21:32:04.072404367Z" level=info msg="StartContainer for \"e7c3577d9d83b277a95daf03cb0e2f93e7539bd2290926b1bae77f143c580885\" returns successfully"
Mar 20 21:32:04.080238 systemd[1]: cri-containerd-e7c3577d9d83b277a95daf03cb0e2f93e7539bd2290926b1bae77f143c580885.scope: Deactivated successfully.
Mar 20 21:32:04.081533 containerd[1505]: time="2025-03-20T21:32:04.081484545Z" level=info msg="received exit event container_id:\"e7c3577d9d83b277a95daf03cb0e2f93e7539bd2290926b1bae77f143c580885\"  id:\"e7c3577d9d83b277a95daf03cb0e2f93e7539bd2290926b1bae77f143c580885\"  pid:4475  exited_at:{seconds:1742506324  nanos:81248893}"
Mar 20 21:32:04.081686 containerd[1505]: time="2025-03-20T21:32:04.081655965Z" level=info msg="TaskExit event in podsandbox handler container_id:\"e7c3577d9d83b277a95daf03cb0e2f93e7539bd2290926b1bae77f143c580885\"  id:\"e7c3577d9d83b277a95daf03cb0e2f93e7539bd2290926b1bae77f143c580885\"  pid:4475  exited_at:{seconds:1742506324  nanos:81248893}"
Mar 20 21:32:04.460897 containerd[1505]: time="2025-03-20T21:32:04.460856289Z" level=info msg="CreateContainer within sandbox \"220ae988f9f95c31190f456aa63f66df0e1a4364c8e5b6cb1ebfcf9663416cba\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}"
Mar 20 21:32:04.472506 containerd[1505]: time="2025-03-20T21:32:04.472251592Z" level=info msg="Container df2ed1ce50adcc797b7b512f0d4a3474c456cec133e4378e2e4155965179ba6e: CDI devices from CRI Config.CDIDevices: []"
Mar 20 21:32:04.478183 containerd[1505]: time="2025-03-20T21:32:04.478151116Z" level=info msg="CreateContainer within sandbox \"220ae988f9f95c31190f456aa63f66df0e1a4364c8e5b6cb1ebfcf9663416cba\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"df2ed1ce50adcc797b7b512f0d4a3474c456cec133e4378e2e4155965179ba6e\""
Mar 20 21:32:04.478661 containerd[1505]: time="2025-03-20T21:32:04.478592824Z" level=info msg="StartContainer for \"df2ed1ce50adcc797b7b512f0d4a3474c456cec133e4378e2e4155965179ba6e\""
Mar 20 21:32:04.479704 containerd[1505]: time="2025-03-20T21:32:04.479674360Z" level=info msg="connecting to shim df2ed1ce50adcc797b7b512f0d4a3474c456cec133e4378e2e4155965179ba6e" address="unix:///run/containerd/s/9a32d66b27678ffde4afb777ceb65a97a2a525c15882991ae929825e83e11bab" protocol=ttrpc version=3
Mar 20 21:32:04.501773 systemd[1]: Started cri-containerd-df2ed1ce50adcc797b7b512f0d4a3474c456cec133e4378e2e4155965179ba6e.scope - libcontainer container df2ed1ce50adcc797b7b512f0d4a3474c456cec133e4378e2e4155965179ba6e.
Mar 20 21:32:04.532313 containerd[1505]: time="2025-03-20T21:32:04.532257128Z" level=info msg="StartContainer for \"df2ed1ce50adcc797b7b512f0d4a3474c456cec133e4378e2e4155965179ba6e\" returns successfully"
Mar 20 21:32:04.537245 systemd[1]: cri-containerd-df2ed1ce50adcc797b7b512f0d4a3474c456cec133e4378e2e4155965179ba6e.scope: Deactivated successfully.
Mar 20 21:32:04.537729 containerd[1505]: time="2025-03-20T21:32:04.537679926Z" level=info msg="received exit event container_id:\"df2ed1ce50adcc797b7b512f0d4a3474c456cec133e4378e2e4155965179ba6e\"  id:\"df2ed1ce50adcc797b7b512f0d4a3474c456cec133e4378e2e4155965179ba6e\"  pid:4519  exited_at:{seconds:1742506324  nanos:537343982}"
Mar 20 21:32:04.537945 containerd[1505]: time="2025-03-20T21:32:04.537696228Z" level=info msg="TaskExit event in podsandbox handler container_id:\"df2ed1ce50adcc797b7b512f0d4a3474c456cec133e4378e2e4155965179ba6e\"  id:\"df2ed1ce50adcc797b7b512f0d4a3474c456cec133e4378e2e4155965179ba6e\"  pid:4519  exited_at:{seconds:1742506324  nanos:537343982}"
Mar 20 21:32:04.757161 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2652187531.mount: Deactivated successfully.
Mar 20 21:32:05.464980 containerd[1505]: time="2025-03-20T21:32:05.464665315Z" level=info msg="CreateContainer within sandbox \"220ae988f9f95c31190f456aa63f66df0e1a4364c8e5b6cb1ebfcf9663416cba\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}"
Mar 20 21:32:05.473728 containerd[1505]: time="2025-03-20T21:32:05.473671081Z" level=info msg="Container b60ed1cb0b54dfc506ee47e6d2362fe5f74539e1402459a1a36eced8718e5f0c: CDI devices from CRI Config.CDIDevices: []"
Mar 20 21:32:05.478757 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2941768751.mount: Deactivated successfully.
Mar 20 21:32:05.486032 containerd[1505]: time="2025-03-20T21:32:05.485983782Z" level=info msg="CreateContainer within sandbox \"220ae988f9f95c31190f456aa63f66df0e1a4364c8e5b6cb1ebfcf9663416cba\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"b60ed1cb0b54dfc506ee47e6d2362fe5f74539e1402459a1a36eced8718e5f0c\""
Mar 20 21:32:05.486520 containerd[1505]: time="2025-03-20T21:32:05.486482519Z" level=info msg="StartContainer for \"b60ed1cb0b54dfc506ee47e6d2362fe5f74539e1402459a1a36eced8718e5f0c\""
Mar 20 21:32:05.488003 containerd[1505]: time="2025-03-20T21:32:05.487972487Z" level=info msg="connecting to shim b60ed1cb0b54dfc506ee47e6d2362fe5f74539e1402459a1a36eced8718e5f0c" address="unix:///run/containerd/s/9a32d66b27678ffde4afb777ceb65a97a2a525c15882991ae929825e83e11bab" protocol=ttrpc version=3
Mar 20 21:32:05.515768 systemd[1]: Started cri-containerd-b60ed1cb0b54dfc506ee47e6d2362fe5f74539e1402459a1a36eced8718e5f0c.scope - libcontainer container b60ed1cb0b54dfc506ee47e6d2362fe5f74539e1402459a1a36eced8718e5f0c.
Mar 20 21:32:05.556375 systemd[1]: cri-containerd-b60ed1cb0b54dfc506ee47e6d2362fe5f74539e1402459a1a36eced8718e5f0c.scope: Deactivated successfully.
Mar 20 21:32:05.557643 containerd[1505]: time="2025-03-20T21:32:05.557590571Z" level=info msg="received exit event container_id:\"b60ed1cb0b54dfc506ee47e6d2362fe5f74539e1402459a1a36eced8718e5f0c\"  id:\"b60ed1cb0b54dfc506ee47e6d2362fe5f74539e1402459a1a36eced8718e5f0c\"  pid:4564  exited_at:{seconds:1742506325  nanos:556748254}"
Mar 20 21:32:05.557930 containerd[1505]: time="2025-03-20T21:32:05.557906116Z" level=info msg="TaskExit event in podsandbox handler container_id:\"b60ed1cb0b54dfc506ee47e6d2362fe5f74539e1402459a1a36eced8718e5f0c\"  id:\"b60ed1cb0b54dfc506ee47e6d2362fe5f74539e1402459a1a36eced8718e5f0c\"  pid:4564  exited_at:{seconds:1742506325  nanos:556748254}"
Mar 20 21:32:05.558252 containerd[1505]: time="2025-03-20T21:32:05.558232121Z" level=info msg="StartContainer for \"b60ed1cb0b54dfc506ee47e6d2362fe5f74539e1402459a1a36eced8718e5f0c\" returns successfully"
Mar 20 21:32:05.579286 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-b60ed1cb0b54dfc506ee47e6d2362fe5f74539e1402459a1a36eced8718e5f0c-rootfs.mount: Deactivated successfully.
Mar 20 21:32:06.472119 containerd[1505]: time="2025-03-20T21:32:06.472074029Z" level=info msg="CreateContainer within sandbox \"220ae988f9f95c31190f456aa63f66df0e1a4364c8e5b6cb1ebfcf9663416cba\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}"
Mar 20 21:32:06.622843 containerd[1505]: time="2025-03-20T21:32:06.622795015Z" level=info msg="Container 020f16734af0165ffbb0fd31671fb7052f790222958ca02224879014975df2fb: CDI devices from CRI Config.CDIDevices: []"
Mar 20 21:32:06.636699 containerd[1505]: time="2025-03-20T21:32:06.636648818Z" level=info msg="CreateContainer within sandbox \"220ae988f9f95c31190f456aa63f66df0e1a4364c8e5b6cb1ebfcf9663416cba\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"020f16734af0165ffbb0fd31671fb7052f790222958ca02224879014975df2fb\""
Mar 20 21:32:06.637355 containerd[1505]: time="2025-03-20T21:32:06.637317630Z" level=info msg="StartContainer for \"020f16734af0165ffbb0fd31671fb7052f790222958ca02224879014975df2fb\""
Mar 20 21:32:06.638144 containerd[1505]: time="2025-03-20T21:32:06.638111713Z" level=info msg="connecting to shim 020f16734af0165ffbb0fd31671fb7052f790222958ca02224879014975df2fb" address="unix:///run/containerd/s/9a32d66b27678ffde4afb777ceb65a97a2a525c15882991ae929825e83e11bab" protocol=ttrpc version=3
Mar 20 21:32:06.665828 systemd[1]: Started cri-containerd-020f16734af0165ffbb0fd31671fb7052f790222958ca02224879014975df2fb.scope - libcontainer container 020f16734af0165ffbb0fd31671fb7052f790222958ca02224879014975df2fb.
Mar 20 21:32:06.691749 systemd[1]: cri-containerd-020f16734af0165ffbb0fd31671fb7052f790222958ca02224879014975df2fb.scope: Deactivated successfully.
Mar 20 21:32:06.692498 containerd[1505]: time="2025-03-20T21:32:06.692309088Z" level=info msg="TaskExit event in podsandbox handler container_id:\"020f16734af0165ffbb0fd31671fb7052f790222958ca02224879014975df2fb\"  id:\"020f16734af0165ffbb0fd31671fb7052f790222958ca02224879014975df2fb\"  pid:4603  exited_at:{seconds:1742506326  nanos:691961341}"
Mar 20 21:32:06.693914 containerd[1505]: time="2025-03-20T21:32:06.693876513Z" level=info msg="received exit event container_id:\"020f16734af0165ffbb0fd31671fb7052f790222958ca02224879014975df2fb\"  id:\"020f16734af0165ffbb0fd31671fb7052f790222958ca02224879014975df2fb\"  pid:4603  exited_at:{seconds:1742506326  nanos:691961341}"
Mar 20 21:32:06.704012 containerd[1505]: time="2025-03-20T21:32:06.703971590Z" level=info msg="StartContainer for \"020f16734af0165ffbb0fd31671fb7052f790222958ca02224879014975df2fb\" returns successfully"
Mar 20 21:32:06.716122 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-020f16734af0165ffbb0fd31671fb7052f790222958ca02224879014975df2fb-rootfs.mount: Deactivated successfully.
Mar 20 21:32:06.857511 kubelet[2605]: E0320 21:32:06.857462    2605 kubelet.go:2901] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"
Mar 20 21:32:07.477383 containerd[1505]: time="2025-03-20T21:32:07.477326324Z" level=info msg="CreateContainer within sandbox \"220ae988f9f95c31190f456aa63f66df0e1a4364c8e5b6cb1ebfcf9663416cba\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}"
Mar 20 21:32:07.486900 containerd[1505]: time="2025-03-20T21:32:07.486195010Z" level=info msg="Container 2b92c5a455ecbd826f37688aa877265db9c0d0bb489c441a1241978d58a7f992: CDI devices from CRI Config.CDIDevices: []"
Mar 20 21:32:07.493834 containerd[1505]: time="2025-03-20T21:32:07.493679182Z" level=info msg="CreateContainer within sandbox \"220ae988f9f95c31190f456aa63f66df0e1a4364c8e5b6cb1ebfcf9663416cba\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"2b92c5a455ecbd826f37688aa877265db9c0d0bb489c441a1241978d58a7f992\""
Mar 20 21:32:07.494421 containerd[1505]: time="2025-03-20T21:32:07.494373854Z" level=info msg="StartContainer for \"2b92c5a455ecbd826f37688aa877265db9c0d0bb489c441a1241978d58a7f992\""
Mar 20 21:32:07.495204 containerd[1505]: time="2025-03-20T21:32:07.495174979Z" level=info msg="connecting to shim 2b92c5a455ecbd826f37688aa877265db9c0d0bb489c441a1241978d58a7f992" address="unix:///run/containerd/s/9a32d66b27678ffde4afb777ceb65a97a2a525c15882991ae929825e83e11bab" protocol=ttrpc version=3
Mar 20 21:32:07.519805 systemd[1]: Started cri-containerd-2b92c5a455ecbd826f37688aa877265db9c0d0bb489c441a1241978d58a7f992.scope - libcontainer container 2b92c5a455ecbd826f37688aa877265db9c0d0bb489c441a1241978d58a7f992.
Mar 20 21:32:07.553325 containerd[1505]: time="2025-03-20T21:32:07.553283105Z" level=info msg="StartContainer for \"2b92c5a455ecbd826f37688aa877265db9c0d0bb489c441a1241978d58a7f992\" returns successfully"
Mar 20 21:32:07.617600 containerd[1505]: time="2025-03-20T21:32:07.617546277Z" level=info msg="TaskExit event in podsandbox handler container_id:\"2b92c5a455ecbd826f37688aa877265db9c0d0bb489c441a1241978d58a7f992\"  id:\"49ead0d2f3232d17afc27c387e2815b4a2dfc0b3c4978847c763f78f08eeca2d\"  pid:4672  exited_at:{seconds:1742506327  nanos:616817390}"
Mar 20 21:32:07.973666 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aesni))
Mar 20 21:32:10.125464 containerd[1505]: time="2025-03-20T21:32:10.125424948Z" level=info msg="TaskExit event in podsandbox handler container_id:\"2b92c5a455ecbd826f37688aa877265db9c0d0bb489c441a1241978d58a7f992\"  id:\"55eda3c9d7dd734b0b1028c1d3ecbf5a8f19366f827d382a65af2127b84e8124\"  pid:4980  exit_status:1  exited_at:{seconds:1742506330  nanos:125074639}"
Mar 20 21:32:10.988046 systemd-networkd[1424]: lxc_health: Link UP
Mar 20 21:32:10.996261 systemd-networkd[1424]: lxc_health: Gained carrier
Mar 20 21:32:11.943646 kubelet[2605]: I0320 21:32:11.943017    2605 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-zhh5g" podStartSLOduration=8.942996053 podStartE2EDuration="8.942996053s" podCreationTimestamp="2025-03-20 21:32:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-20 21:32:08.494032153 +0000 UTC m=+86.761338701" watchObservedRunningTime="2025-03-20 21:32:11.942996053 +0000 UTC m=+90.210302601"
Mar 20 21:32:12.240012 containerd[1505]: time="2025-03-20T21:32:12.239881651Z" level=info msg="TaskExit event in podsandbox handler container_id:\"2b92c5a455ecbd826f37688aa877265db9c0d0bb489c441a1241978d58a7f992\"  id:\"160c5a4dbad7d73a2c0ff1b25a7512d0a852a5a5f8f6a62230d1e1602cffe771\"  pid:5240  exited_at:{seconds:1742506332  nanos:239438695}"
Mar 20 21:32:12.371835 systemd-networkd[1424]: lxc_health: Gained IPv6LL
Mar 20 21:32:14.348477 containerd[1505]: time="2025-03-20T21:32:14.348431970Z" level=info msg="TaskExit event in podsandbox handler container_id:\"2b92c5a455ecbd826f37688aa877265db9c0d0bb489c441a1241978d58a7f992\"  id:\"5407615d2325965b077b7c4b7d8f32452295aaad746719b6e8d0ecba60cc6507\"  pid:5274  exited_at:{seconds:1742506334  nanos:348129422}"
Mar 20 21:32:16.430759 containerd[1505]: time="2025-03-20T21:32:16.430711181Z" level=info msg="TaskExit event in podsandbox handler container_id:\"2b92c5a455ecbd826f37688aa877265db9c0d0bb489c441a1241978d58a7f992\"  id:\"6e0c386a6fc6eeec84a1629e9bd8528cc290c3bbfb6d1a69a7219527c4637872\"  pid:5299  exited_at:{seconds:1742506336  nanos:430353708}"
Mar 20 21:32:16.435940 sshd[4408]: Connection closed by 10.0.0.1 port 48896
Mar 20 21:32:16.436370 sshd-session[4401]: pam_unix(sshd:session): session closed for user core
Mar 20 21:32:16.440418 systemd[1]: sshd@27-10.0.0.132:22-10.0.0.1:48896.service: Deactivated successfully.
Mar 20 21:32:16.442467 systemd[1]: session-28.scope: Deactivated successfully.
Mar 20 21:32:16.443181 systemd-logind[1487]: Session 28 logged out. Waiting for processes to exit.
Mar 20 21:32:16.444201 systemd-logind[1487]: Removed session 28.