Mar 19 11:54:36.953254 kernel: Linux version 6.6.83-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.2.1_p20241221 p7) 14.2.1 20241221, GNU ld (Gentoo 2.43 p3) 2.43.1) #1 SMP PREEMPT_DYNAMIC Wed Mar 19 10:13:43 -00 2025 Mar 19 11:54:36.953775 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=hetzner verity.usrhash=08c32ef14ad6302a92b1d281c48443f5b56d59f0d37d38df628e5b6f012967bc Mar 19 11:54:36.953793 kernel: BIOS-provided physical RAM map: Mar 19 11:54:36.953801 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable Mar 19 11:54:36.953809 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved Mar 19 11:54:36.953817 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Mar 19 11:54:36.953826 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000007cfdbfff] usable Mar 19 11:54:36.953834 kernel: BIOS-e820: [mem 0x000000007cfdc000-0x000000007cffffff] reserved Mar 19 11:54:36.953845 kernel: BIOS-e820: [mem 0x00000000b0000000-0x00000000bfffffff] reserved Mar 19 11:54:36.953853 kernel: BIOS-e820: [mem 0x00000000fed1c000-0x00000000fed1ffff] reserved Mar 19 11:54:36.953862 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Mar 19 11:54:36.953870 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Mar 19 11:54:36.953878 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Mar 19 11:54:36.953887 kernel: NX (Execute Disable) protection: active Mar 19 11:54:36.953900 kernel: APIC: Static calls initialized Mar 19 11:54:36.953910 kernel: SMBIOS 3.0.0 present. Mar 19 11:54:36.953919 kernel: DMI: Hetzner vServer/Standard PC (Q35 + ICH9, 2009), BIOS 20171111 11/11/2017 Mar 19 11:54:36.953928 kernel: Hypervisor detected: KVM Mar 19 11:54:36.953937 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Mar 19 11:54:36.953946 kernel: kvm-clock: using sched offset of 3671211690 cycles Mar 19 11:54:36.953955 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Mar 19 11:54:36.953964 kernel: tsc: Detected 2495.312 MHz processor Mar 19 11:54:36.953973 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Mar 19 11:54:36.953983 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Mar 19 11:54:36.953995 kernel: last_pfn = 0x7cfdc max_arch_pfn = 0x400000000 Mar 19 11:54:36.954004 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs Mar 19 11:54:36.954013 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Mar 19 11:54:36.954022 kernel: Using GB pages for direct mapping Mar 19 11:54:36.954031 kernel: ACPI: Early table checksum verification disabled Mar 19 11:54:36.954040 kernel: ACPI: RSDP 0x00000000000F5270 000014 (v00 BOCHS ) Mar 19 11:54:36.954049 kernel: ACPI: RSDT 0x000000007CFE265D 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 19 11:54:36.954058 kernel: ACPI: FACP 0x000000007CFE244D 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Mar 19 11:54:36.954068 kernel: ACPI: DSDT 0x000000007CFE0040 00240D (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 19 11:54:36.954079 kernel: ACPI: FACS 0x000000007CFE0000 000040 Mar 19 11:54:36.954088 kernel: ACPI: APIC 0x000000007CFE2541 000080 (v03 BOCHS BXPC 00000001 BXPC 00000001) Mar 19 11:54:36.954097 kernel: ACPI: HPET 0x000000007CFE25C1 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 19 11:54:36.954106 kernel: ACPI: MCFG 0x000000007CFE25F9 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 19 11:54:36.954115 kernel: ACPI: WAET 0x000000007CFE2635 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 19 11:54:36.954124 kernel: ACPI: Reserving FACP table memory at [mem 0x7cfe244d-0x7cfe2540] Mar 19 11:54:36.954134 kernel: ACPI: Reserving DSDT table memory at [mem 0x7cfe0040-0x7cfe244c] Mar 19 11:54:36.954148 kernel: ACPI: Reserving FACS table memory at [mem 0x7cfe0000-0x7cfe003f] Mar 19 11:54:36.954157 kernel: ACPI: Reserving APIC table memory at [mem 0x7cfe2541-0x7cfe25c0] Mar 19 11:54:36.954166 kernel: ACPI: Reserving HPET table memory at [mem 0x7cfe25c1-0x7cfe25f8] Mar 19 11:54:36.954177 kernel: ACPI: Reserving MCFG table memory at [mem 0x7cfe25f9-0x7cfe2634] Mar 19 11:54:36.954186 kernel: ACPI: Reserving WAET table memory at [mem 0x7cfe2635-0x7cfe265c] Mar 19 11:54:36.954195 kernel: No NUMA configuration found Mar 19 11:54:36.954205 kernel: Faking a node at [mem 0x0000000000000000-0x000000007cfdbfff] Mar 19 11:54:36.954218 kernel: NODE_DATA(0) allocated [mem 0x7cfd6000-0x7cfdbfff] Mar 19 11:54:36.954228 kernel: Zone ranges: Mar 19 11:54:36.954238 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Mar 19 11:54:36.954247 kernel: DMA32 [mem 0x0000000001000000-0x000000007cfdbfff] Mar 19 11:54:36.954257 kernel: Normal empty Mar 19 11:54:36.954267 kernel: Movable zone start for each node Mar 19 11:54:36.954276 kernel: Early memory node ranges Mar 19 11:54:36.954314 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Mar 19 11:54:36.954324 kernel: node 0: [mem 0x0000000000100000-0x000000007cfdbfff] Mar 19 11:54:36.954336 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000007cfdbfff] Mar 19 11:54:36.954345 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Mar 19 11:54:36.954355 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Mar 19 11:54:36.954364 kernel: On node 0, zone DMA32: 12324 pages in unavailable ranges Mar 19 11:54:36.954373 kernel: ACPI: PM-Timer IO Port: 0x608 Mar 19 11:54:36.954383 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Mar 19 11:54:36.954392 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Mar 19 11:54:36.954402 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Mar 19 11:54:36.954411 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Mar 19 11:54:36.954423 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Mar 19 11:54:36.954432 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Mar 19 11:54:36.954442 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Mar 19 11:54:36.954451 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Mar 19 11:54:36.954461 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Mar 19 11:54:36.954470 kernel: smpboot: Allowing 2 CPUs, 0 hotplug CPUs Mar 19 11:54:36.954480 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Mar 19 11:54:36.954489 kernel: [mem 0xc0000000-0xfed1bfff] available for PCI devices Mar 19 11:54:36.954498 kernel: Booting paravirtualized kernel on KVM Mar 19 11:54:36.954508 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Mar 19 11:54:36.954520 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1 Mar 19 11:54:36.954530 kernel: percpu: Embedded 58 pages/cpu s197032 r8192 d32344 u1048576 Mar 19 11:54:36.954539 kernel: pcpu-alloc: s197032 r8192 d32344 u1048576 alloc=1*2097152 Mar 19 11:54:36.954549 kernel: pcpu-alloc: [0] 0 1 Mar 19 11:54:36.954558 kernel: kvm-guest: PV spinlocks disabled, no host support Mar 19 11:54:36.954569 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=hetzner verity.usrhash=08c32ef14ad6302a92b1d281c48443f5b56d59f0d37d38df628e5b6f012967bc Mar 19 11:54:36.954579 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Mar 19 11:54:36.954591 kernel: random: crng init done Mar 19 11:54:36.954602 kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Mar 19 11:54:36.954611 kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear) Mar 19 11:54:36.954621 kernel: Fallback order for Node 0: 0 Mar 19 11:54:36.954631 kernel: Built 1 zonelists, mobility grouping on. Total pages: 503708 Mar 19 11:54:36.954640 kernel: Policy zone: DMA32 Mar 19 11:54:36.954649 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Mar 19 11:54:36.954659 kernel: Memory: 1920004K/2047464K available (14336K kernel code, 2303K rwdata, 22860K rodata, 43480K init, 1592K bss, 127200K reserved, 0K cma-reserved) Mar 19 11:54:36.954669 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Mar 19 11:54:36.954681 kernel: ftrace: allocating 37910 entries in 149 pages Mar 19 11:54:36.954691 kernel: ftrace: allocated 149 pages with 4 groups Mar 19 11:54:36.954700 kernel: Dynamic Preempt: voluntary Mar 19 11:54:36.954710 kernel: rcu: Preemptible hierarchical RCU implementation. Mar 19 11:54:36.954720 kernel: rcu: RCU event tracing is enabled. Mar 19 11:54:36.954730 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Mar 19 11:54:36.954760 kernel: Trampoline variant of Tasks RCU enabled. Mar 19 11:54:36.954770 kernel: Rude variant of Tasks RCU enabled. Mar 19 11:54:36.954780 kernel: Tracing variant of Tasks RCU enabled. Mar 19 11:54:36.954791 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Mar 19 11:54:36.954804 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Mar 19 11:54:36.954813 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 Mar 19 11:54:36.954823 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Mar 19 11:54:36.954832 kernel: Console: colour VGA+ 80x25 Mar 19 11:54:36.954841 kernel: printk: console [tty0] enabled Mar 19 11:54:36.954851 kernel: printk: console [ttyS0] enabled Mar 19 11:54:36.954860 kernel: ACPI: Core revision 20230628 Mar 19 11:54:36.954870 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Mar 19 11:54:36.954879 kernel: APIC: Switch to symmetric I/O mode setup Mar 19 11:54:36.954892 kernel: x2apic enabled Mar 19 11:54:36.954901 kernel: APIC: Switched APIC routing to: physical x2apic Mar 19 11:54:36.954911 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Mar 19 11:54:36.954920 kernel: tsc: Marking TSC unstable due to TSCs unsynchronized Mar 19 11:54:36.954930 kernel: Calibrating delay loop (skipped) preset value.. 4990.62 BogoMIPS (lpj=2495312) Mar 19 11:54:36.954939 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Mar 19 11:54:36.954949 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Mar 19 11:54:36.954959 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Mar 19 11:54:36.954978 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Mar 19 11:54:36.954988 kernel: Spectre V2 : Mitigation: Retpolines Mar 19 11:54:36.954998 kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch Mar 19 11:54:36.955009 kernel: Spectre V2 : Spectre v2 / SpectreRSB : Filling RSB on VMEXIT Mar 19 11:54:36.955020 kernel: Spectre V2 : Enabling Speculation Barrier for firmware calls Mar 19 11:54:36.955029 kernel: RETBleed: Mitigation: untrained return thunk Mar 19 11:54:36.955040 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Mar 19 11:54:36.955050 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Mar 19 11:54:36.955061 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Mar 19 11:54:36.955073 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Mar 19 11:54:36.955083 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Mar 19 11:54:36.955093 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Mar 19 11:54:36.955103 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. Mar 19 11:54:36.955114 kernel: Freeing SMP alternatives memory: 32K Mar 19 11:54:36.955124 kernel: pid_max: default: 32768 minimum: 301 Mar 19 11:54:36.955134 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Mar 19 11:54:36.955144 kernel: landlock: Up and running. Mar 19 11:54:36.955157 kernel: SELinux: Initializing. Mar 19 11:54:36.955168 kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Mar 19 11:54:36.955178 kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Mar 19 11:54:36.955189 kernel: smpboot: CPU0: AMD EPYC-Rome Processor (family: 0x17, model: 0x31, stepping: 0x0) Mar 19 11:54:36.955199 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Mar 19 11:54:36.955209 kernel: RCU Tasks Rude: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Mar 19 11:54:36.955219 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Mar 19 11:54:36.955229 kernel: Performance Events: Fam17h+ core perfctr, AMD PMU driver. Mar 19 11:54:36.955239 kernel: ... version: 0 Mar 19 11:54:36.955251 kernel: ... bit width: 48 Mar 19 11:54:36.955261 kernel: ... generic registers: 6 Mar 19 11:54:36.955271 kernel: ... value mask: 0000ffffffffffff Mar 19 11:54:36.955295 kernel: ... max period: 00007fffffffffff Mar 19 11:54:36.955306 kernel: ... fixed-purpose events: 0 Mar 19 11:54:36.955316 kernel: ... event mask: 000000000000003f Mar 19 11:54:36.955326 kernel: signal: max sigframe size: 1776 Mar 19 11:54:36.955336 kernel: rcu: Hierarchical SRCU implementation. Mar 19 11:54:36.955346 kernel: rcu: Max phase no-delay instances is 400. Mar 19 11:54:36.955359 kernel: smp: Bringing up secondary CPUs ... Mar 19 11:54:36.955369 kernel: smpboot: x86: Booting SMP configuration: Mar 19 11:54:36.955379 kernel: .... node #0, CPUs: #1 Mar 19 11:54:36.955389 kernel: smp: Brought up 1 node, 2 CPUs Mar 19 11:54:36.955399 kernel: smpboot: Max logical packages: 1 Mar 19 11:54:36.955409 kernel: smpboot: Total of 2 processors activated (9981.24 BogoMIPS) Mar 19 11:54:36.955419 kernel: devtmpfs: initialized Mar 19 11:54:36.955428 kernel: x86/mm: Memory block size: 128MB Mar 19 11:54:36.955438 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Mar 19 11:54:36.955451 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Mar 19 11:54:36.955461 kernel: pinctrl core: initialized pinctrl subsystem Mar 19 11:54:36.955471 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Mar 19 11:54:36.955481 kernel: audit: initializing netlink subsys (disabled) Mar 19 11:54:36.955492 kernel: audit: type=2000 audit(1742385276.443:1): state=initialized audit_enabled=0 res=1 Mar 19 11:54:36.955502 kernel: thermal_sys: Registered thermal governor 'step_wise' Mar 19 11:54:36.955513 kernel: thermal_sys: Registered thermal governor 'user_space' Mar 19 11:54:36.955524 kernel: cpuidle: using governor menu Mar 19 11:54:36.955534 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Mar 19 11:54:36.955547 kernel: dca service started, version 1.12.1 Mar 19 11:54:36.955557 kernel: PCI: MMCONFIG for domain 0000 [bus 00-ff] at [mem 0xb0000000-0xbfffffff] (base 0xb0000000) Mar 19 11:54:36.955567 kernel: PCI: Using configuration type 1 for base access Mar 19 11:54:36.955577 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Mar 19 11:54:36.955587 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Mar 19 11:54:36.955597 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Mar 19 11:54:36.955607 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Mar 19 11:54:36.955617 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Mar 19 11:54:36.955627 kernel: ACPI: Added _OSI(Module Device) Mar 19 11:54:36.955639 kernel: ACPI: Added _OSI(Processor Device) Mar 19 11:54:36.955648 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Mar 19 11:54:36.955658 kernel: ACPI: Added _OSI(Processor Aggregator Device) Mar 19 11:54:36.955668 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Mar 19 11:54:36.955678 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Mar 19 11:54:36.955688 kernel: ACPI: Interpreter enabled Mar 19 11:54:36.955698 kernel: ACPI: PM: (supports S0 S5) Mar 19 11:54:36.955707 kernel: ACPI: Using IOAPIC for interrupt routing Mar 19 11:54:36.955717 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Mar 19 11:54:36.955729 kernel: PCI: Using E820 reservations for host bridge windows Mar 19 11:54:36.955751 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Mar 19 11:54:36.955760 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Mar 19 11:54:36.955929 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Mar 19 11:54:36.956040 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Mar 19 11:54:36.956147 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Mar 19 11:54:36.956162 kernel: PCI host bridge to bus 0000:00 Mar 19 11:54:36.956292 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Mar 19 11:54:36.956420 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Mar 19 11:54:36.956515 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Mar 19 11:54:36.956609 kernel: pci_bus 0000:00: root bus resource [mem 0x7d000000-0xafffffff window] Mar 19 11:54:36.956709 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window] Mar 19 11:54:36.956811 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x8ffffffff window] Mar 19 11:54:36.956907 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Mar 19 11:54:36.957033 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 Mar 19 11:54:36.957163 kernel: pci 0000:00:01.0: [1af4:1050] type 00 class 0x030000 Mar 19 11:54:36.957276 kernel: pci 0000:00:01.0: reg 0x10: [mem 0xfb800000-0xfbffffff pref] Mar 19 11:54:36.957411 kernel: pci 0000:00:01.0: reg 0x18: [mem 0xfd200000-0xfd203fff 64bit pref] Mar 19 11:54:36.957517 kernel: pci 0000:00:01.0: reg 0x20: [mem 0xfea10000-0xfea10fff] Mar 19 11:54:36.957645 kernel: pci 0000:00:01.0: reg 0x30: [mem 0xfea00000-0xfea0ffff pref] Mar 19 11:54:36.957768 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Mar 19 11:54:36.957891 kernel: pci 0000:00:02.0: [1b36:000c] type 01 class 0x060400 Mar 19 11:54:36.957997 kernel: pci 0000:00:02.0: reg 0x10: [mem 0xfea11000-0xfea11fff] Mar 19 11:54:36.958136 kernel: pci 0000:00:02.1: [1b36:000c] type 01 class 0x060400 Mar 19 11:54:36.958247 kernel: pci 0000:00:02.1: reg 0x10: [mem 0xfea12000-0xfea12fff] Mar 19 11:54:36.960420 kernel: pci 0000:00:02.2: [1b36:000c] type 01 class 0x060400 Mar 19 11:54:36.960535 kernel: pci 0000:00:02.2: reg 0x10: [mem 0xfea13000-0xfea13fff] Mar 19 11:54:36.960658 kernel: pci 0000:00:02.3: [1b36:000c] type 01 class 0x060400 Mar 19 11:54:36.960778 kernel: pci 0000:00:02.3: reg 0x10: [mem 0xfea14000-0xfea14fff] Mar 19 11:54:36.960889 kernel: pci 0000:00:02.4: [1b36:000c] type 01 class 0x060400 Mar 19 11:54:36.960991 kernel: pci 0000:00:02.4: reg 0x10: [mem 0xfea15000-0xfea15fff] Mar 19 11:54:36.961102 kernel: pci 0000:00:02.5: [1b36:000c] type 01 class 0x060400 Mar 19 11:54:36.961207 kernel: pci 0000:00:02.5: reg 0x10: [mem 0xfea16000-0xfea16fff] Mar 19 11:54:36.962412 kernel: pci 0000:00:02.6: [1b36:000c] type 01 class 0x060400 Mar 19 11:54:36.962528 kernel: pci 0000:00:02.6: reg 0x10: [mem 0xfea17000-0xfea17fff] Mar 19 11:54:36.962652 kernel: pci 0000:00:02.7: [1b36:000c] type 01 class 0x060400 Mar 19 11:54:36.962776 kernel: pci 0000:00:02.7: reg 0x10: [mem 0xfea18000-0xfea18fff] Mar 19 11:54:36.962891 kernel: pci 0000:00:03.0: [1b36:000c] type 01 class 0x060400 Mar 19 11:54:36.962997 kernel: pci 0000:00:03.0: reg 0x10: [mem 0xfea19000-0xfea19fff] Mar 19 11:54:36.963117 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 Mar 19 11:54:36.963221 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Mar 19 11:54:36.964346 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 Mar 19 11:54:36.964460 kernel: pci 0000:00:1f.2: reg 0x20: [io 0xc040-0xc05f] Mar 19 11:54:36.964561 kernel: pci 0000:00:1f.2: reg 0x24: [mem 0xfea1a000-0xfea1afff] Mar 19 11:54:36.964672 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 Mar 19 11:54:36.964823 kernel: pci 0000:00:1f.3: reg 0x20: [io 0x0700-0x073f] Mar 19 11:54:36.964943 kernel: pci 0000:01:00.0: [1af4:1041] type 00 class 0x020000 Mar 19 11:54:36.965053 kernel: pci 0000:01:00.0: reg 0x14: [mem 0xfe880000-0xfe880fff] Mar 19 11:54:36.965165 kernel: pci 0000:01:00.0: reg 0x20: [mem 0xfd000000-0xfd003fff 64bit pref] Mar 19 11:54:36.965275 kernel: pci 0000:01:00.0: reg 0x30: [mem 0xfe800000-0xfe87ffff pref] Mar 19 11:54:36.966453 kernel: pci 0000:00:02.0: PCI bridge to [bus 01] Mar 19 11:54:36.966564 kernel: pci 0000:00:02.0: bridge window [mem 0xfe800000-0xfe9fffff] Mar 19 11:54:36.966669 kernel: pci 0000:00:02.0: bridge window [mem 0xfd000000-0xfd1fffff 64bit pref] Mar 19 11:54:36.966800 kernel: pci 0000:02:00.0: [1b36:000d] type 00 class 0x0c0330 Mar 19 11:54:36.966908 kernel: pci 0000:02:00.0: reg 0x10: [mem 0xfe600000-0xfe603fff 64bit] Mar 19 11:54:36.967016 kernel: pci 0000:00:02.1: PCI bridge to [bus 02] Mar 19 11:54:36.967124 kernel: pci 0000:00:02.1: bridge window [mem 0xfe600000-0xfe7fffff] Mar 19 11:54:36.967229 kernel: pci 0000:00:02.1: bridge window [mem 0xfce00000-0xfcffffff 64bit pref] Mar 19 11:54:36.968397 kernel: pci 0000:03:00.0: [1af4:1043] type 00 class 0x078000 Mar 19 11:54:36.968520 kernel: pci 0000:03:00.0: reg 0x14: [mem 0xfe400000-0xfe400fff] Mar 19 11:54:36.968632 kernel: pci 0000:03:00.0: reg 0x20: [mem 0xfcc00000-0xfcc03fff 64bit pref] Mar 19 11:54:36.968757 kernel: pci 0000:00:02.2: PCI bridge to [bus 03] Mar 19 11:54:36.968866 kernel: pci 0000:00:02.2: bridge window [mem 0xfe400000-0xfe5fffff] Mar 19 11:54:36.968975 kernel: pci 0000:00:02.2: bridge window [mem 0xfcc00000-0xfcdfffff 64bit pref] Mar 19 11:54:36.969101 kernel: pci 0000:04:00.0: [1af4:1045] type 00 class 0x00ff00 Mar 19 11:54:36.969210 kernel: pci 0000:04:00.0: reg 0x20: [mem 0xfca00000-0xfca03fff 64bit pref] Mar 19 11:54:36.970366 kernel: pci 0000:00:02.3: PCI bridge to [bus 04] Mar 19 11:54:36.970477 kernel: pci 0000:00:02.3: bridge window [mem 0xfe200000-0xfe3fffff] Mar 19 11:54:36.970584 kernel: pci 0000:00:02.3: bridge window [mem 0xfca00000-0xfcbfffff 64bit pref] Mar 19 11:54:36.970700 kernel: pci 0000:05:00.0: [1af4:1044] type 00 class 0x00ff00 Mar 19 11:54:36.970823 kernel: pci 0000:05:00.0: reg 0x14: [mem 0xfe000000-0xfe000fff] Mar 19 11:54:36.970933 kernel: pci 0000:05:00.0: reg 0x20: [mem 0xfc800000-0xfc803fff 64bit pref] Mar 19 11:54:36.971037 kernel: pci 0000:00:02.4: PCI bridge to [bus 05] Mar 19 11:54:36.971147 kernel: pci 0000:00:02.4: bridge window [mem 0xfe000000-0xfe1fffff] Mar 19 11:54:36.971252 kernel: pci 0000:00:02.4: bridge window [mem 0xfc800000-0xfc9fffff 64bit pref] Mar 19 11:54:36.971385 kernel: pci 0000:06:00.0: [1af4:1048] type 00 class 0x010000 Mar 19 11:54:36.971496 kernel: pci 0000:06:00.0: reg 0x14: [mem 0xfde00000-0xfde00fff] Mar 19 11:54:36.972384 kernel: pci 0000:06:00.0: reg 0x20: [mem 0xfc600000-0xfc603fff 64bit pref] Mar 19 11:54:36.972496 kernel: pci 0000:00:02.5: PCI bridge to [bus 06] Mar 19 11:54:36.972599 kernel: pci 0000:00:02.5: bridge window [mem 0xfde00000-0xfdffffff] Mar 19 11:54:36.972710 kernel: pci 0000:00:02.5: bridge window [mem 0xfc600000-0xfc7fffff 64bit pref] Mar 19 11:54:36.972731 kernel: acpiphp: Slot [0] registered Mar 19 11:54:36.972862 kernel: pci 0000:07:00.0: [1af4:1041] type 00 class 0x020000 Mar 19 11:54:36.972973 kernel: pci 0000:07:00.0: reg 0x14: [mem 0xfdc80000-0xfdc80fff] Mar 19 11:54:36.973081 kernel: pci 0000:07:00.0: reg 0x20: [mem 0xfc400000-0xfc403fff 64bit pref] Mar 19 11:54:36.973184 kernel: pci 0000:07:00.0: reg 0x30: [mem 0xfdc00000-0xfdc7ffff pref] Mar 19 11:54:36.975373 kernel: pci 0000:00:02.6: PCI bridge to [bus 07] Mar 19 11:54:36.975499 kernel: pci 0000:00:02.6: bridge window [mem 0xfdc00000-0xfddfffff] Mar 19 11:54:36.975611 kernel: pci 0000:00:02.6: bridge window [mem 0xfc400000-0xfc5fffff 64bit pref] Mar 19 11:54:36.975627 kernel: acpiphp: Slot [0-2] registered Mar 19 11:54:36.975729 kernel: pci 0000:00:02.7: PCI bridge to [bus 08] Mar 19 11:54:36.975891 kernel: pci 0000:00:02.7: bridge window [mem 0xfda00000-0xfdbfffff] Mar 19 11:54:36.975990 kernel: pci 0000:00:02.7: bridge window [mem 0xfc200000-0xfc3fffff 64bit pref] Mar 19 11:54:36.976005 kernel: acpiphp: Slot [0-3] registered Mar 19 11:54:36.976110 kernel: pci 0000:00:03.0: PCI bridge to [bus 09] Mar 19 11:54:36.976211 kernel: pci 0000:00:03.0: bridge window [mem 0xfd800000-0xfd9fffff] Mar 19 11:54:36.976332 kernel: pci 0000:00:03.0: bridge window [mem 0xfc000000-0xfc1fffff 64bit pref] Mar 19 11:54:36.976353 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Mar 19 11:54:36.976363 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Mar 19 11:54:36.976374 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Mar 19 11:54:36.976384 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Mar 19 11:54:36.976395 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Mar 19 11:54:36.976406 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Mar 19 11:54:36.976416 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Mar 19 11:54:36.976426 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Mar 19 11:54:36.976440 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Mar 19 11:54:36.976450 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Mar 19 11:54:36.976460 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Mar 19 11:54:36.976470 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Mar 19 11:54:36.976480 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Mar 19 11:54:36.976490 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Mar 19 11:54:36.976500 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Mar 19 11:54:36.976511 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Mar 19 11:54:36.976521 kernel: iommu: Default domain type: Translated Mar 19 11:54:36.976534 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Mar 19 11:54:36.976544 kernel: PCI: Using ACPI for IRQ routing Mar 19 11:54:36.976555 kernel: PCI: pci_cache_line_size set to 64 bytes Mar 19 11:54:36.976565 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] Mar 19 11:54:36.976575 kernel: e820: reserve RAM buffer [mem 0x7cfdc000-0x7fffffff] Mar 19 11:54:36.976684 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Mar 19 11:54:36.976805 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Mar 19 11:54:36.976910 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Mar 19 11:54:36.976925 kernel: vgaarb: loaded Mar 19 11:54:36.976939 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Mar 19 11:54:36.976950 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Mar 19 11:54:36.976960 kernel: clocksource: Switched to clocksource kvm-clock Mar 19 11:54:36.976971 kernel: VFS: Disk quotas dquot_6.6.0 Mar 19 11:54:36.976982 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Mar 19 11:54:36.976994 kernel: pnp: PnP ACPI init Mar 19 11:54:36.977109 kernel: system 00:04: [mem 0xb0000000-0xbfffffff window] has been reserved Mar 19 11:54:36.977126 kernel: pnp: PnP ACPI: found 5 devices Mar 19 11:54:36.977140 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Mar 19 11:54:36.977151 kernel: NET: Registered PF_INET protocol family Mar 19 11:54:36.977161 kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear) Mar 19 11:54:36.977172 kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear) Mar 19 11:54:36.977183 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Mar 19 11:54:36.977193 kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear) Mar 19 11:54:36.977203 kernel: TCP bind hash table entries: 16384 (order: 7, 524288 bytes, linear) Mar 19 11:54:36.977213 kernel: TCP: Hash tables configured (established 16384 bind 16384) Mar 19 11:54:36.977223 kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear) Mar 19 11:54:36.977236 kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear) Mar 19 11:54:36.977246 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Mar 19 11:54:36.977256 kernel: NET: Registered PF_XDP protocol family Mar 19 11:54:36.979446 kernel: pci 0000:00:02.6: bridge window [io 0x1000-0x0fff] to [bus 07] add_size 1000 Mar 19 11:54:36.979611 kernel: pci 0000:00:02.7: bridge window [io 0x1000-0x0fff] to [bus 08] add_size 1000 Mar 19 11:54:36.979720 kernel: pci 0000:00:03.0: bridge window [io 0x1000-0x0fff] to [bus 09] add_size 1000 Mar 19 11:54:36.979839 kernel: pci 0000:00:02.6: BAR 13: assigned [io 0x1000-0x1fff] Mar 19 11:54:36.979948 kernel: pci 0000:00:02.7: BAR 13: assigned [io 0x2000-0x2fff] Mar 19 11:54:36.980059 kernel: pci 0000:00:03.0: BAR 13: assigned [io 0x3000-0x3fff] Mar 19 11:54:36.980172 kernel: pci 0000:00:02.0: PCI bridge to [bus 01] Mar 19 11:54:36.980275 kernel: pci 0000:00:02.0: bridge window [mem 0xfe800000-0xfe9fffff] Mar 19 11:54:36.981449 kernel: pci 0000:00:02.0: bridge window [mem 0xfd000000-0xfd1fffff 64bit pref] Mar 19 11:54:36.981556 kernel: pci 0000:00:02.1: PCI bridge to [bus 02] Mar 19 11:54:36.981663 kernel: pci 0000:00:02.1: bridge window [mem 0xfe600000-0xfe7fffff] Mar 19 11:54:36.981786 kernel: pci 0000:00:02.1: bridge window [mem 0xfce00000-0xfcffffff 64bit pref] Mar 19 11:54:36.981894 kernel: pci 0000:00:02.2: PCI bridge to [bus 03] Mar 19 11:54:36.982005 kernel: pci 0000:00:02.2: bridge window [mem 0xfe400000-0xfe5fffff] Mar 19 11:54:36.982175 kernel: pci 0000:00:02.2: bridge window [mem 0xfcc00000-0xfcdfffff 64bit pref] Mar 19 11:54:36.982307 kernel: pci 0000:00:02.3: PCI bridge to [bus 04] Mar 19 11:54:36.982414 kernel: pci 0000:00:02.3: bridge window [mem 0xfe200000-0xfe3fffff] Mar 19 11:54:36.982520 kernel: pci 0000:00:02.3: bridge window [mem 0xfca00000-0xfcbfffff 64bit pref] Mar 19 11:54:36.982633 kernel: pci 0000:00:02.4: PCI bridge to [bus 05] Mar 19 11:54:36.982757 kernel: pci 0000:00:02.4: bridge window [mem 0xfe000000-0xfe1fffff] Mar 19 11:54:36.982880 kernel: pci 0000:00:02.4: bridge window [mem 0xfc800000-0xfc9fffff 64bit pref] Mar 19 11:54:36.982991 kernel: pci 0000:00:02.5: PCI bridge to [bus 06] Mar 19 11:54:36.983101 kernel: pci 0000:00:02.5: bridge window [mem 0xfde00000-0xfdffffff] Mar 19 11:54:36.983208 kernel: pci 0000:00:02.5: bridge window [mem 0xfc600000-0xfc7fffff 64bit pref] Mar 19 11:54:36.985376 kernel: pci 0000:00:02.6: PCI bridge to [bus 07] Mar 19 11:54:36.985492 kernel: pci 0000:00:02.6: bridge window [io 0x1000-0x1fff] Mar 19 11:54:36.985596 kernel: pci 0000:00:02.6: bridge window [mem 0xfdc00000-0xfddfffff] Mar 19 11:54:36.985699 kernel: pci 0000:00:02.6: bridge window [mem 0xfc400000-0xfc5fffff 64bit pref] Mar 19 11:54:36.985817 kernel: pci 0000:00:02.7: PCI bridge to [bus 08] Mar 19 11:54:36.985922 kernel: pci 0000:00:02.7: bridge window [io 0x2000-0x2fff] Mar 19 11:54:36.986033 kernel: pci 0000:00:02.7: bridge window [mem 0xfda00000-0xfdbfffff] Mar 19 11:54:36.986138 kernel: pci 0000:00:02.7: bridge window [mem 0xfc200000-0xfc3fffff 64bit pref] Mar 19 11:54:36.986249 kernel: pci 0000:00:03.0: PCI bridge to [bus 09] Mar 19 11:54:36.987410 kernel: pci 0000:00:03.0: bridge window [io 0x3000-0x3fff] Mar 19 11:54:36.987518 kernel: pci 0000:00:03.0: bridge window [mem 0xfd800000-0xfd9fffff] Mar 19 11:54:36.987631 kernel: pci 0000:00:03.0: bridge window [mem 0xfc000000-0xfc1fffff 64bit pref] Mar 19 11:54:36.987749 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Mar 19 11:54:36.987844 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Mar 19 11:54:36.987943 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Mar 19 11:54:36.988033 kernel: pci_bus 0000:00: resource 7 [mem 0x7d000000-0xafffffff window] Mar 19 11:54:36.988129 kernel: pci_bus 0000:00: resource 8 [mem 0xc0000000-0xfebfffff window] Mar 19 11:54:36.988223 kernel: pci_bus 0000:00: resource 9 [mem 0x100000000-0x8ffffffff window] Mar 19 11:54:36.988389 kernel: pci_bus 0000:01: resource 1 [mem 0xfe800000-0xfe9fffff] Mar 19 11:54:36.988490 kernel: pci_bus 0000:01: resource 2 [mem 0xfd000000-0xfd1fffff 64bit pref] Mar 19 11:54:36.988593 kernel: pci_bus 0000:02: resource 1 [mem 0xfe600000-0xfe7fffff] Mar 19 11:54:36.988695 kernel: pci_bus 0000:02: resource 2 [mem 0xfce00000-0xfcffffff 64bit pref] Mar 19 11:54:36.988826 kernel: pci_bus 0000:03: resource 1 [mem 0xfe400000-0xfe5fffff] Mar 19 11:54:36.988925 kernel: pci_bus 0000:03: resource 2 [mem 0xfcc00000-0xfcdfffff 64bit pref] Mar 19 11:54:36.989031 kernel: pci_bus 0000:04: resource 1 [mem 0xfe200000-0xfe3fffff] Mar 19 11:54:36.989131 kernel: pci_bus 0000:04: resource 2 [mem 0xfca00000-0xfcbfffff 64bit pref] Mar 19 11:54:36.989241 kernel: pci_bus 0000:05: resource 1 [mem 0xfe000000-0xfe1fffff] Mar 19 11:54:36.991493 kernel: pci_bus 0000:05: resource 2 [mem 0xfc800000-0xfc9fffff 64bit pref] Mar 19 11:54:36.991688 kernel: pci_bus 0000:06: resource 1 [mem 0xfde00000-0xfdffffff] Mar 19 11:54:36.991860 kernel: pci_bus 0000:06: resource 2 [mem 0xfc600000-0xfc7fffff 64bit pref] Mar 19 11:54:36.992019 kernel: pci_bus 0000:07: resource 0 [io 0x1000-0x1fff] Mar 19 11:54:36.992167 kernel: pci_bus 0000:07: resource 1 [mem 0xfdc00000-0xfddfffff] Mar 19 11:54:36.994316 kernel: pci_bus 0000:07: resource 2 [mem 0xfc400000-0xfc5fffff 64bit pref] Mar 19 11:54:36.994395 kernel: pci_bus 0000:08: resource 0 [io 0x2000-0x2fff] Mar 19 11:54:36.994463 kernel: pci_bus 0000:08: resource 1 [mem 0xfda00000-0xfdbfffff] Mar 19 11:54:36.994534 kernel: pci_bus 0000:08: resource 2 [mem 0xfc200000-0xfc3fffff 64bit pref] Mar 19 11:54:36.994605 kernel: pci_bus 0000:09: resource 0 [io 0x3000-0x3fff] Mar 19 11:54:36.994672 kernel: pci_bus 0000:09: resource 1 [mem 0xfd800000-0xfd9fffff] Mar 19 11:54:36.994760 kernel: pci_bus 0000:09: resource 2 [mem 0xfc000000-0xfc1fffff 64bit pref] Mar 19 11:54:36.994772 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Mar 19 11:54:36.994782 kernel: PCI: CLS 0 bytes, default 64 Mar 19 11:54:36.994791 kernel: Initialise system trusted keyrings Mar 19 11:54:36.994801 kernel: workingset: timestamp_bits=39 max_order=19 bucket_order=0 Mar 19 11:54:36.994812 kernel: Key type asymmetric registered Mar 19 11:54:36.994820 kernel: Asymmetric key parser 'x509' registered Mar 19 11:54:36.994828 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Mar 19 11:54:36.994836 kernel: io scheduler mq-deadline registered Mar 19 11:54:36.994844 kernel: io scheduler kyber registered Mar 19 11:54:36.994851 kernel: io scheduler bfq registered Mar 19 11:54:36.994934 kernel: pcieport 0000:00:02.0: PME: Signaling with IRQ 24 Mar 19 11:54:36.995009 kernel: pcieport 0000:00:02.0: AER: enabled with IRQ 24 Mar 19 11:54:36.995092 kernel: pcieport 0000:00:02.1: PME: Signaling with IRQ 25 Mar 19 11:54:36.995169 kernel: pcieport 0000:00:02.1: AER: enabled with IRQ 25 Mar 19 11:54:36.995243 kernel: pcieport 0000:00:02.2: PME: Signaling with IRQ 26 Mar 19 11:54:36.995338 kernel: pcieport 0000:00:02.2: AER: enabled with IRQ 26 Mar 19 11:54:36.995414 kernel: pcieport 0000:00:02.3: PME: Signaling with IRQ 27 Mar 19 11:54:36.995487 kernel: pcieport 0000:00:02.3: AER: enabled with IRQ 27 Mar 19 11:54:36.995564 kernel: pcieport 0000:00:02.4: PME: Signaling with IRQ 28 Mar 19 11:54:36.995639 kernel: pcieport 0000:00:02.4: AER: enabled with IRQ 28 Mar 19 11:54:36.995714 kernel: pcieport 0000:00:02.5: PME: Signaling with IRQ 29 Mar 19 11:54:36.995803 kernel: pcieport 0000:00:02.5: AER: enabled with IRQ 29 Mar 19 11:54:36.995878 kernel: pcieport 0000:00:02.6: PME: Signaling with IRQ 30 Mar 19 11:54:36.995951 kernel: pcieport 0000:00:02.6: AER: enabled with IRQ 30 Mar 19 11:54:36.996026 kernel: pcieport 0000:00:02.7: PME: Signaling with IRQ 31 Mar 19 11:54:36.996100 kernel: pcieport 0000:00:02.7: AER: enabled with IRQ 31 Mar 19 11:54:36.996110 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Mar 19 11:54:36.996185 kernel: pcieport 0000:00:03.0: PME: Signaling with IRQ 32 Mar 19 11:54:36.996258 kernel: pcieport 0000:00:03.0: AER: enabled with IRQ 32 Mar 19 11:54:36.996271 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Mar 19 11:54:36.997321 kernel: ACPI: \_SB_.GSIF: Enabled at IRQ 21 Mar 19 11:54:36.997330 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Mar 19 11:54:36.997338 kernel: 00:00: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Mar 19 11:54:36.997346 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Mar 19 11:54:36.997354 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Mar 19 11:54:36.997362 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Mar 19 11:54:36.997448 kernel: rtc_cmos 00:03: RTC can wake from S4 Mar 19 11:54:36.997466 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Mar 19 11:54:36.997554 kernel: rtc_cmos 00:03: registered as rtc0 Mar 19 11:54:36.997640 kernel: rtc_cmos 00:03: setting system clock to 2025-03-19T11:54:36 UTC (1742385276) Mar 19 11:54:36.997725 kernel: rtc_cmos 00:03: alarms up to one day, y3k, 242 bytes nvram, hpet irqs Mar 19 11:54:36.997754 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Mar 19 11:54:36.997768 kernel: NET: Registered PF_INET6 protocol family Mar 19 11:54:36.997778 kernel: Segment Routing with IPv6 Mar 19 11:54:36.997788 kernel: In-situ OAM (IOAM) with IPv6 Mar 19 11:54:36.997798 kernel: NET: Registered PF_PACKET protocol family Mar 19 11:54:36.997812 kernel: Key type dns_resolver registered Mar 19 11:54:36.997821 kernel: IPI shorthand broadcast: enabled Mar 19 11:54:36.997831 kernel: sched_clock: Marking stable (1325009647, 147821181)->(1484112116, -11281288) Mar 19 11:54:36.997841 kernel: registered taskstats version 1 Mar 19 11:54:36.997852 kernel: Loading compiled-in X.509 certificates Mar 19 11:54:36.997862 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.83-flatcar: ea8d6696bd19c98b32173a761210456cdad6b56b' Mar 19 11:54:36.997870 kernel: Key type .fscrypt registered Mar 19 11:54:36.997878 kernel: Key type fscrypt-provisioning registered Mar 19 11:54:36.997885 kernel: ima: No TPM chip found, activating TPM-bypass! Mar 19 11:54:36.997896 kernel: ima: Allocated hash algorithm: sha1 Mar 19 11:54:36.997904 kernel: ima: No architecture policies found Mar 19 11:54:36.997911 kernel: clk: Disabling unused clocks Mar 19 11:54:36.997919 kernel: Freeing unused kernel image (initmem) memory: 43480K Mar 19 11:54:36.997927 kernel: Write protecting the kernel read-only data: 38912k Mar 19 11:54:36.997934 kernel: Freeing unused kernel image (rodata/data gap) memory: 1716K Mar 19 11:54:36.997942 kernel: Run /init as init process Mar 19 11:54:36.997949 kernel: with arguments: Mar 19 11:54:36.997958 kernel: /init Mar 19 11:54:36.997966 kernel: with environment: Mar 19 11:54:36.997973 kernel: HOME=/ Mar 19 11:54:36.997980 kernel: TERM=linux Mar 19 11:54:36.997990 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Mar 19 11:54:36.998002 systemd[1]: Successfully made /usr/ read-only. Mar 19 11:54:36.998017 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Mar 19 11:54:36.998029 systemd[1]: Detected virtualization kvm. Mar 19 11:54:36.998041 systemd[1]: Detected architecture x86-64. Mar 19 11:54:36.998050 systemd[1]: Running in initrd. Mar 19 11:54:36.998058 systemd[1]: No hostname configured, using default hostname. Mar 19 11:54:36.998067 systemd[1]: Hostname set to . Mar 19 11:54:36.998074 systemd[1]: Initializing machine ID from VM UUID. Mar 19 11:54:36.998082 systemd[1]: Queued start job for default target initrd.target. Mar 19 11:54:36.998090 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 19 11:54:36.998098 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 19 11:54:36.998108 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Mar 19 11:54:36.998117 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Mar 19 11:54:36.998125 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Mar 19 11:54:36.998134 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Mar 19 11:54:36.998142 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Mar 19 11:54:36.998151 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Mar 19 11:54:36.998159 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 19 11:54:36.998169 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Mar 19 11:54:36.998177 systemd[1]: Reached target paths.target - Path Units. Mar 19 11:54:36.998185 systemd[1]: Reached target slices.target - Slice Units. Mar 19 11:54:36.998193 systemd[1]: Reached target swap.target - Swaps. Mar 19 11:54:36.998201 systemd[1]: Reached target timers.target - Timer Units. Mar 19 11:54:36.998209 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Mar 19 11:54:36.998217 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Mar 19 11:54:36.998226 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Mar 19 11:54:36.998234 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Mar 19 11:54:36.998244 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Mar 19 11:54:36.998252 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Mar 19 11:54:36.998260 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Mar 19 11:54:36.998268 systemd[1]: Reached target sockets.target - Socket Units. Mar 19 11:54:36.999289 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Mar 19 11:54:36.999301 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Mar 19 11:54:36.999319 systemd[1]: Finished network-cleanup.service - Network Cleanup. Mar 19 11:54:36.999327 systemd[1]: Starting systemd-fsck-usr.service... Mar 19 11:54:36.999338 systemd[1]: Starting systemd-journald.service - Journal Service... Mar 19 11:54:36.999346 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Mar 19 11:54:36.999385 systemd-journald[189]: Collecting audit messages is disabled. Mar 19 11:54:36.999407 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 19 11:54:36.999417 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Mar 19 11:54:36.999425 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Mar 19 11:54:36.999435 systemd[1]: Finished systemd-fsck-usr.service. Mar 19 11:54:36.999444 systemd-journald[189]: Journal started Mar 19 11:54:36.999468 systemd-journald[189]: Runtime Journal (/run/log/journal/8f6e0c84addd436cb2eccb830195f8d0) is 4.8M, max 38.3M, 33.5M free. Mar 19 11:54:36.979630 systemd-modules-load[191]: Inserted module 'overlay' Mar 19 11:54:37.029609 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Mar 19 11:54:37.029637 kernel: Bridge firewalling registered Mar 19 11:54:37.029649 systemd[1]: Started systemd-journald.service - Journal Service. Mar 19 11:54:37.009335 systemd-modules-load[191]: Inserted module 'br_netfilter' Mar 19 11:54:37.030511 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Mar 19 11:54:37.031433 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 19 11:54:37.038437 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 19 11:54:37.040172 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Mar 19 11:54:37.041605 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Mar 19 11:54:37.044654 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Mar 19 11:54:37.053753 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Mar 19 11:54:37.059914 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Mar 19 11:54:37.061606 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Mar 19 11:54:37.064648 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 19 11:54:37.067590 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 19 11:54:37.069719 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Mar 19 11:54:37.072411 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Mar 19 11:54:37.075157 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 19 11:54:37.086854 dracut-cmdline[222]: dracut-dracut-053 Mar 19 11:54:37.089500 dracut-cmdline[222]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=hetzner verity.usrhash=08c32ef14ad6302a92b1d281c48443f5b56d59f0d37d38df628e5b6f012967bc Mar 19 11:54:37.106850 systemd-resolved[223]: Positive Trust Anchors: Mar 19 11:54:37.107553 systemd-resolved[223]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Mar 19 11:54:37.107585 systemd-resolved[223]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Mar 19 11:54:37.116559 systemd-resolved[223]: Defaulting to hostname 'linux'. Mar 19 11:54:37.117521 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Mar 19 11:54:37.118263 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Mar 19 11:54:37.147307 kernel: SCSI subsystem initialized Mar 19 11:54:37.156308 kernel: Loading iSCSI transport class v2.0-870. Mar 19 11:54:37.173340 kernel: iscsi: registered transport (tcp) Mar 19 11:54:37.193713 kernel: iscsi: registered transport (qla4xxx) Mar 19 11:54:37.193822 kernel: QLogic iSCSI HBA Driver Mar 19 11:54:37.237775 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Mar 19 11:54:37.244696 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Mar 19 11:54:37.281206 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Mar 19 11:54:37.281318 kernel: device-mapper: uevent: version 1.0.3 Mar 19 11:54:37.281338 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Mar 19 11:54:37.338392 kernel: raid6: avx2x4 gen() 17491 MB/s Mar 19 11:54:37.355372 kernel: raid6: avx2x2 gen() 29811 MB/s Mar 19 11:54:37.372572 kernel: raid6: avx2x1 gen() 25803 MB/s Mar 19 11:54:37.372682 kernel: raid6: using algorithm avx2x2 gen() 29811 MB/s Mar 19 11:54:37.391326 kernel: raid6: .... xor() 19969 MB/s, rmw enabled Mar 19 11:54:37.391408 kernel: raid6: using avx2x2 recovery algorithm Mar 19 11:54:37.411345 kernel: xor: automatically using best checksumming function avx Mar 19 11:54:37.593343 kernel: Btrfs loaded, zoned=no, fsverity=no Mar 19 11:54:37.604417 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Mar 19 11:54:37.609481 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 19 11:54:37.623337 systemd-udevd[407]: Using default interface naming scheme 'v255'. Mar 19 11:54:37.627729 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 19 11:54:37.634874 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Mar 19 11:54:37.647263 dracut-pre-trigger[416]: rd.md=0: removing MD RAID activation Mar 19 11:54:37.675937 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Mar 19 11:54:37.687642 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Mar 19 11:54:37.730576 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Mar 19 11:54:37.740494 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Mar 19 11:54:37.758752 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Mar 19 11:54:37.760931 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Mar 19 11:54:37.763204 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 19 11:54:37.764634 systemd[1]: Reached target remote-fs.target - Remote File Systems. Mar 19 11:54:37.770643 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Mar 19 11:54:37.783533 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Mar 19 11:54:37.811393 kernel: scsi host0: Virtio SCSI HBA Mar 19 11:54:37.834657 kernel: scsi 0:0:0:0: Direct-Access QEMU QEMU HARDDISK 2.5+ PQ: 0 ANSI: 5 Mar 19 11:54:37.841350 kernel: libata version 3.00 loaded. Mar 19 11:54:37.871316 kernel: cryptd: max_cpu_qlen set to 1000 Mar 19 11:54:37.858010 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Mar 19 11:54:37.858128 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 19 11:54:37.875138 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 19 11:54:37.876613 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 19 11:54:37.877222 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 19 11:54:37.883814 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Mar 19 11:54:37.891624 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 19 11:54:37.897506 kernel: ahci 0000:00:1f.2: version 3.0 Mar 19 11:54:37.949226 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Mar 19 11:54:37.949252 kernel: AVX2 version of gcm_enc/dec engaged. Mar 19 11:54:37.949304 kernel: AES CTR mode by8 optimization enabled Mar 19 11:54:37.949320 kernel: ahci 0000:00:1f.2: AHCI 0001.0000 32 slots 6 ports 1.5 Gbps 0x3f impl SATA mode Mar 19 11:54:37.949493 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Mar 19 11:54:37.949635 kernel: scsi host1: ahci Mar 19 11:54:37.949804 kernel: ACPI: bus type USB registered Mar 19 11:54:37.949821 kernel: scsi host2: ahci Mar 19 11:54:37.949959 kernel: usbcore: registered new interface driver usbfs Mar 19 11:54:37.949976 kernel: usbcore: registered new interface driver hub Mar 19 11:54:37.949994 kernel: usbcore: registered new device driver usb Mar 19 11:54:37.950008 kernel: scsi host3: ahci Mar 19 11:54:37.950149 kernel: scsi host4: ahci Mar 19 11:54:37.950322 kernel: scsi host5: ahci Mar 19 11:54:37.950469 kernel: scsi host6: ahci Mar 19 11:54:37.950609 kernel: ata1: SATA max UDMA/133 abar m4096@0xfea1a000 port 0xfea1a100 irq 48 Mar 19 11:54:37.950626 kernel: ata2: SATA max UDMA/133 abar m4096@0xfea1a000 port 0xfea1a180 irq 48 Mar 19 11:54:37.950641 kernel: ata3: SATA max UDMA/133 abar m4096@0xfea1a000 port 0xfea1a200 irq 48 Mar 19 11:54:37.950654 kernel: ata4: SATA max UDMA/133 abar m4096@0xfea1a000 port 0xfea1a280 irq 48 Mar 19 11:54:37.950667 kernel: ata5: SATA max UDMA/133 abar m4096@0xfea1a000 port 0xfea1a300 irq 48 Mar 19 11:54:37.950680 kernel: ata6: SATA max UDMA/133 abar m4096@0xfea1a000 port 0xfea1a380 irq 48 Mar 19 11:54:37.987309 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 19 11:54:37.991524 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 19 11:54:38.008127 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 19 11:54:38.253311 kernel: ata1: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Mar 19 11:54:38.253408 kernel: ata1.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Mar 19 11:54:38.253423 kernel: ata1.00: applying bridge limits Mar 19 11:54:38.253437 kernel: ata5: SATA link down (SStatus 0 SControl 300) Mar 19 11:54:38.259441 kernel: ata4: SATA link down (SStatus 0 SControl 300) Mar 19 11:54:38.259544 kernel: ata2: SATA link down (SStatus 0 SControl 300) Mar 19 11:54:38.260299 kernel: ata3: SATA link down (SStatus 0 SControl 300) Mar 19 11:54:38.262338 kernel: ata6: SATA link down (SStatus 0 SControl 300) Mar 19 11:54:38.264329 kernel: ata1.00: configured for UDMA/100 Mar 19 11:54:38.266602 kernel: scsi 1:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Mar 19 11:54:38.293350 kernel: xhci_hcd 0000:02:00.0: xHCI Host Controller Mar 19 11:54:38.323058 kernel: xhci_hcd 0000:02:00.0: new USB bus registered, assigned bus number 1 Mar 19 11:54:38.323233 kernel: sd 0:0:0:0: Power-on or device reset occurred Mar 19 11:54:38.341419 kernel: xhci_hcd 0000:02:00.0: hcc params 0x00087001 hci version 0x100 quirks 0x0000000000000010 Mar 19 11:54:38.341613 kernel: sd 0:0:0:0: [sda] 80003072 512-byte logical blocks: (41.0 GB/38.1 GiB) Mar 19 11:54:38.341765 kernel: xhci_hcd 0000:02:00.0: xHCI Host Controller Mar 19 11:54:38.341894 kernel: xhci_hcd 0000:02:00.0: new USB bus registered, assigned bus number 2 Mar 19 11:54:38.342013 kernel: sd 0:0:0:0: [sda] Write Protect is off Mar 19 11:54:38.342135 kernel: sd 0:0:0:0: [sda] Mode Sense: 63 00 00 08 Mar 19 11:54:38.342259 kernel: xhci_hcd 0000:02:00.0: Host supports USB 3.0 SuperSpeed Mar 19 11:54:38.342417 kernel: sd 0:0:0:0: [sda] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA Mar 19 11:54:38.342542 kernel: hub 1-0:1.0: USB hub found Mar 19 11:54:38.342683 kernel: hub 1-0:1.0: 4 ports detected Mar 19 11:54:38.342827 kernel: usb usb2: We don't know the algorithms for LPM for this host, disabling LPM. Mar 19 11:54:38.342976 kernel: hub 2-0:1.0: USB hub found Mar 19 11:54:38.343126 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Mar 19 11:54:38.343146 kernel: hub 2-0:1.0: 4 ports detected Mar 19 11:54:38.343276 kernel: GPT:17805311 != 80003071 Mar 19 11:54:38.344548 kernel: GPT:Alternate GPT header not at the end of the disk. Mar 19 11:54:38.344561 kernel: GPT:17805311 != 80003071 Mar 19 11:54:38.344573 kernel: GPT: Use GNU Parted to correct GPT errors. Mar 19 11:54:38.344584 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Mar 19 11:54:38.344596 kernel: sd 0:0:0:0: [sda] Attached SCSI disk Mar 19 11:54:38.355570 kernel: sr 1:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Mar 19 11:54:38.371645 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Mar 19 11:54:38.371669 kernel: sr 1:0:0:0: Attached scsi CD-ROM sr0 Mar 19 11:54:38.395778 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/sda6 scanned by (udev-worker) (461) Mar 19 11:54:38.399319 kernel: BTRFS: device fsid 8d57424d-5abc-4888-810f-658d040a58e4 devid 1 transid 36 /dev/sda3 scanned by (udev-worker) (464) Mar 19 11:54:38.420690 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - QEMU_HARDDISK ROOT. Mar 19 11:54:38.436925 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - QEMU_HARDDISK EFI-SYSTEM. Mar 19 11:54:38.444674 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - QEMU_HARDDISK USR-A. Mar 19 11:54:38.445269 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - QEMU_HARDDISK USR-A. Mar 19 11:54:38.455057 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - QEMU_HARDDISK OEM. Mar 19 11:54:38.460459 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Mar 19 11:54:38.465907 disk-uuid[581]: Primary Header is updated. Mar 19 11:54:38.465907 disk-uuid[581]: Secondary Entries is updated. Mar 19 11:54:38.465907 disk-uuid[581]: Secondary Header is updated. Mar 19 11:54:38.480310 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Mar 19 11:54:38.555343 kernel: usb 1-1: new high-speed USB device number 2 using xhci_hcd Mar 19 11:54:38.692317 kernel: hid: raw HID events driver (C) Jiri Kosina Mar 19 11:54:38.697420 kernel: usbcore: registered new interface driver usbhid Mar 19 11:54:38.697497 kernel: usbhid: USB HID core driver Mar 19 11:54:38.703326 kernel: input: QEMU QEMU USB Tablet as /devices/pci0000:00/0000:00:02.1/0000:02:00.0/usb1/1-1/1-1:1.0/0003:0627:0001.0001/input/input2 Mar 19 11:54:38.706306 kernel: hid-generic 0003:0627:0001.0001: input,hidraw0: USB HID v0.01 Mouse [QEMU QEMU USB Tablet] on usb-0000:02:00.0-1/input0 Mar 19 11:54:39.490516 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Mar 19 11:54:39.491418 disk-uuid[582]: The operation has completed successfully. Mar 19 11:54:39.545023 systemd[1]: disk-uuid.service: Deactivated successfully. Mar 19 11:54:39.545134 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Mar 19 11:54:39.612520 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Mar 19 11:54:39.615767 sh[598]: Success Mar 19 11:54:39.630310 kernel: device-mapper: verity: sha256 using implementation "sha256-ni" Mar 19 11:54:39.688095 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Mar 19 11:54:39.695377 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Mar 19 11:54:39.699350 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Mar 19 11:54:39.716425 kernel: BTRFS info (device dm-0): first mount of filesystem 8d57424d-5abc-4888-810f-658d040a58e4 Mar 19 11:54:39.716522 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Mar 19 11:54:39.718379 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Mar 19 11:54:39.720389 kernel: BTRFS info (device dm-0): disabling log replay at mount time Mar 19 11:54:39.722895 kernel: BTRFS info (device dm-0): using free space tree Mar 19 11:54:39.731334 kernel: BTRFS info (device dm-0): enabling ssd optimizations Mar 19 11:54:39.733949 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Mar 19 11:54:39.735181 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Mar 19 11:54:39.740487 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Mar 19 11:54:39.742492 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Mar 19 11:54:39.758321 kernel: BTRFS info (device sda6): first mount of filesystem 3c2c2d54-a06e-4f36-8d13-ab30a5d0eab5 Mar 19 11:54:39.761587 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Mar 19 11:54:39.761659 kernel: BTRFS info (device sda6): using free space tree Mar 19 11:54:39.768627 kernel: BTRFS info (device sda6): enabling ssd optimizations Mar 19 11:54:39.768703 kernel: BTRFS info (device sda6): auto enabling async discard Mar 19 11:54:39.781862 systemd[1]: mnt-oem.mount: Deactivated successfully. Mar 19 11:54:39.782483 kernel: BTRFS info (device sda6): last unmount of filesystem 3c2c2d54-a06e-4f36-8d13-ab30a5d0eab5 Mar 19 11:54:39.788050 systemd[1]: Finished ignition-setup.service - Ignition (setup). Mar 19 11:54:39.793514 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Mar 19 11:54:39.831973 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Mar 19 11:54:39.841578 systemd[1]: Starting systemd-networkd.service - Network Configuration... Mar 19 11:54:39.874505 systemd-networkd[780]: lo: Link UP Mar 19 11:54:39.874516 systemd-networkd[780]: lo: Gained carrier Mar 19 11:54:39.880692 systemd-networkd[780]: Enumeration completed Mar 19 11:54:39.880954 systemd[1]: Started systemd-networkd.service - Network Configuration. Mar 19 11:54:39.881635 systemd[1]: Reached target network.target - Network. Mar 19 11:54:39.882926 systemd-networkd[780]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 19 11:54:39.882930 systemd-networkd[780]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 19 11:54:39.884396 systemd-networkd[780]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 19 11:54:39.884401 systemd-networkd[780]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 19 11:54:39.892559 systemd-networkd[780]: eth0: Link UP Mar 19 11:54:39.892566 systemd-networkd[780]: eth0: Gained carrier Mar 19 11:54:39.892577 systemd-networkd[780]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 19 11:54:39.896702 systemd-networkd[780]: eth1: Link UP Mar 19 11:54:39.896711 systemd-networkd[780]: eth1: Gained carrier Mar 19 11:54:39.896724 systemd-networkd[780]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 19 11:54:39.903933 ignition[730]: Ignition 2.20.0 Mar 19 11:54:39.903946 ignition[730]: Stage: fetch-offline Mar 19 11:54:39.903986 ignition[730]: no configs at "/usr/lib/ignition/base.d" Mar 19 11:54:39.905708 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Mar 19 11:54:39.903995 ignition[730]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Mar 19 11:54:39.904099 ignition[730]: parsed url from cmdline: "" Mar 19 11:54:39.904102 ignition[730]: no config URL provided Mar 19 11:54:39.904107 ignition[730]: reading system config file "/usr/lib/ignition/user.ign" Mar 19 11:54:39.904116 ignition[730]: no config at "/usr/lib/ignition/user.ign" Mar 19 11:54:39.904121 ignition[730]: failed to fetch config: resource requires networking Mar 19 11:54:39.904477 ignition[730]: Ignition finished successfully Mar 19 11:54:39.916926 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Mar 19 11:54:39.927391 systemd-networkd[780]: eth1: DHCPv4 address 10.0.0.3/32, gateway 10.0.0.1 acquired from 10.0.0.1 Mar 19 11:54:39.929772 ignition[790]: Ignition 2.20.0 Mar 19 11:54:39.929785 ignition[790]: Stage: fetch Mar 19 11:54:39.929976 ignition[790]: no configs at "/usr/lib/ignition/base.d" Mar 19 11:54:39.929987 ignition[790]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Mar 19 11:54:39.930073 ignition[790]: parsed url from cmdline: "" Mar 19 11:54:39.930077 ignition[790]: no config URL provided Mar 19 11:54:39.930082 ignition[790]: reading system config file "/usr/lib/ignition/user.ign" Mar 19 11:54:39.930091 ignition[790]: no config at "/usr/lib/ignition/user.ign" Mar 19 11:54:39.930116 ignition[790]: GET http://169.254.169.254/hetzner/v1/userdata: attempt #1 Mar 19 11:54:39.930338 ignition[790]: GET error: Get "http://169.254.169.254/hetzner/v1/userdata": dial tcp 169.254.169.254:80: connect: network is unreachable Mar 19 11:54:39.957405 systemd-networkd[780]: eth0: DHCPv4 address 37.27.246.226/32, gateway 172.31.1.1 acquired from 172.31.1.1 Mar 19 11:54:40.130582 ignition[790]: GET http://169.254.169.254/hetzner/v1/userdata: attempt #2 Mar 19 11:54:40.134759 ignition[790]: GET result: OK Mar 19 11:54:40.134834 ignition[790]: parsing config with SHA512: 8cff0bc9cf4b58dae1b5a4e6936558c2cbec8097c47b4d3b17823e289b4b59c02d57d35a11b55c04d3716bbfe3092fd6bda4617547368b0dad55a7f38628b7ae Mar 19 11:54:40.139809 unknown[790]: fetched base config from "system" Mar 19 11:54:40.139822 unknown[790]: fetched base config from "system" Mar 19 11:54:40.140263 ignition[790]: fetch: fetch complete Mar 19 11:54:40.139827 unknown[790]: fetched user config from "hetzner" Mar 19 11:54:40.140269 ignition[790]: fetch: fetch passed Mar 19 11:54:40.142242 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Mar 19 11:54:40.140570 ignition[790]: Ignition finished successfully Mar 19 11:54:40.148509 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Mar 19 11:54:40.163562 ignition[799]: Ignition 2.20.0 Mar 19 11:54:40.164464 ignition[799]: Stage: kargs Mar 19 11:54:40.165181 ignition[799]: no configs at "/usr/lib/ignition/base.d" Mar 19 11:54:40.165783 ignition[799]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Mar 19 11:54:40.167621 ignition[799]: kargs: kargs passed Mar 19 11:54:40.168225 ignition[799]: Ignition finished successfully Mar 19 11:54:40.169875 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Mar 19 11:54:40.177535 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Mar 19 11:54:40.190862 ignition[806]: Ignition 2.20.0 Mar 19 11:54:40.190877 ignition[806]: Stage: disks Mar 19 11:54:40.191126 ignition[806]: no configs at "/usr/lib/ignition/base.d" Mar 19 11:54:40.191143 ignition[806]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Mar 19 11:54:40.192787 ignition[806]: disks: disks passed Mar 19 11:54:40.194665 systemd[1]: Finished ignition-disks.service - Ignition (disks). Mar 19 11:54:40.192841 ignition[806]: Ignition finished successfully Mar 19 11:54:40.195778 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Mar 19 11:54:40.196621 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Mar 19 11:54:40.197727 systemd[1]: Reached target local-fs.target - Local File Systems. Mar 19 11:54:40.198991 systemd[1]: Reached target sysinit.target - System Initialization. Mar 19 11:54:40.200366 systemd[1]: Reached target basic.target - Basic System. Mar 19 11:54:40.207537 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Mar 19 11:54:40.223003 systemd-fsck[814]: ROOT: clean, 14/1628000 files, 120691/1617920 blocks Mar 19 11:54:40.226034 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Mar 19 11:54:40.724498 systemd[1]: Mounting sysroot.mount - /sysroot... Mar 19 11:54:40.813309 kernel: EXT4-fs (sda9): mounted filesystem 303a73dd-e104-408b-9302-bf91b04ba1ca r/w with ordered data mode. Quota mode: none. Mar 19 11:54:40.814148 systemd[1]: Mounted sysroot.mount - /sysroot. Mar 19 11:54:40.815239 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Mar 19 11:54:40.823441 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Mar 19 11:54:40.825929 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Mar 19 11:54:40.828977 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Mar 19 11:54:40.832271 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Mar 19 11:54:40.832442 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Mar 19 11:54:40.845975 kernel: BTRFS: device label OEM devid 1 transid 15 /dev/sda6 scanned by mount (822) Mar 19 11:54:40.846006 kernel: BTRFS info (device sda6): first mount of filesystem 3c2c2d54-a06e-4f36-8d13-ab30a5d0eab5 Mar 19 11:54:40.846020 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Mar 19 11:54:40.846032 kernel: BTRFS info (device sda6): using free space tree Mar 19 11:54:40.836136 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Mar 19 11:54:40.848671 kernel: BTRFS info (device sda6): enabling ssd optimizations Mar 19 11:54:40.848721 kernel: BTRFS info (device sda6): auto enabling async discard Mar 19 11:54:40.847179 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Mar 19 11:54:40.854645 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Mar 19 11:54:40.910675 initrd-setup-root[850]: cut: /sysroot/etc/passwd: No such file or directory Mar 19 11:54:40.912290 coreos-metadata[824]: Mar 19 11:54:40.911 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/hostname: Attempt #1 Mar 19 11:54:40.914313 coreos-metadata[824]: Mar 19 11:54:40.914 INFO Fetch successful Mar 19 11:54:40.915202 coreos-metadata[824]: Mar 19 11:54:40.915 INFO wrote hostname ci-4230-1-0-d-e284590f49 to /sysroot/etc/hostname Mar 19 11:54:40.916409 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Mar 19 11:54:40.919362 initrd-setup-root[857]: cut: /sysroot/etc/group: No such file or directory Mar 19 11:54:40.922902 initrd-setup-root[865]: cut: /sysroot/etc/shadow: No such file or directory Mar 19 11:54:40.926382 initrd-setup-root[872]: cut: /sysroot/etc/gshadow: No such file or directory Mar 19 11:54:41.010873 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Mar 19 11:54:41.015406 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Mar 19 11:54:41.018483 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Mar 19 11:54:41.026334 kernel: BTRFS info (device sda6): last unmount of filesystem 3c2c2d54-a06e-4f36-8d13-ab30a5d0eab5 Mar 19 11:54:41.049776 ignition[939]: INFO : Ignition 2.20.0 Mar 19 11:54:41.049776 ignition[939]: INFO : Stage: mount Mar 19 11:54:41.052795 ignition[939]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 19 11:54:41.052795 ignition[939]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Mar 19 11:54:41.052795 ignition[939]: INFO : mount: mount passed Mar 19 11:54:41.052795 ignition[939]: INFO : Ignition finished successfully Mar 19 11:54:41.053097 systemd[1]: Finished ignition-mount.service - Ignition (mount). Mar 19 11:54:41.060494 systemd[1]: Starting ignition-files.service - Ignition (files)... Mar 19 11:54:41.061350 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Mar 19 11:54:41.150523 systemd-networkd[780]: eth1: Gained IPv6LL Mar 19 11:54:41.470511 systemd-networkd[780]: eth0: Gained IPv6LL Mar 19 11:54:41.714304 systemd[1]: sysroot-oem.mount: Deactivated successfully. Mar 19 11:54:41.719563 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Mar 19 11:54:41.732420 kernel: BTRFS: device label OEM devid 1 transid 16 /dev/sda6 scanned by mount (951) Mar 19 11:54:41.732490 kernel: BTRFS info (device sda6): first mount of filesystem 3c2c2d54-a06e-4f36-8d13-ab30a5d0eab5 Mar 19 11:54:41.736184 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Mar 19 11:54:41.736263 kernel: BTRFS info (device sda6): using free space tree Mar 19 11:54:41.742332 kernel: BTRFS info (device sda6): enabling ssd optimizations Mar 19 11:54:41.742411 kernel: BTRFS info (device sda6): auto enabling async discard Mar 19 11:54:41.746697 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Mar 19 11:54:41.767575 ignition[967]: INFO : Ignition 2.20.0 Mar 19 11:54:41.768509 ignition[967]: INFO : Stage: files Mar 19 11:54:41.769253 ignition[967]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 19 11:54:41.770780 ignition[967]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Mar 19 11:54:41.771749 ignition[967]: DEBUG : files: compiled without relabeling support, skipping Mar 19 11:54:41.772531 ignition[967]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Mar 19 11:54:41.772531 ignition[967]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Mar 19 11:54:41.775082 ignition[967]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Mar 19 11:54:41.776060 ignition[967]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Mar 19 11:54:41.776060 ignition[967]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Mar 19 11:54:41.775550 unknown[967]: wrote ssh authorized keys file for user: core Mar 19 11:54:41.778600 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Mar 19 11:54:41.778600 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-amd64.tar.gz: attempt #1 Mar 19 11:54:41.970555 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Mar 19 11:54:42.210752 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Mar 19 11:54:42.212074 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Mar 19 11:54:42.212074 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-amd64.tar.gz: attempt #1 Mar 19 11:54:42.977965 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Mar 19 11:54:43.243255 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Mar 19 11:54:43.243255 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Mar 19 11:54:43.245333 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Mar 19 11:54:43.245333 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Mar 19 11:54:43.245333 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Mar 19 11:54:43.245333 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Mar 19 11:54:43.245333 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Mar 19 11:54:43.245333 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Mar 19 11:54:43.245333 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Mar 19 11:54:43.245333 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Mar 19 11:54:43.245333 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Mar 19 11:54:43.245333 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.0-x86-64.raw" Mar 19 11:54:43.245333 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.0-x86-64.raw" Mar 19 11:54:43.245333 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.0-x86-64.raw" Mar 19 11:54:43.245333 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.31.0-x86-64.raw: attempt #1 Mar 19 11:54:43.953076 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Mar 19 11:54:45.490879 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.0-x86-64.raw" Mar 19 11:54:45.490879 ignition[967]: INFO : files: op(c): [started] processing unit "prepare-helm.service" Mar 19 11:54:45.495930 ignition[967]: INFO : files: op(c): op(d): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Mar 19 11:54:45.495930 ignition[967]: INFO : files: op(c): op(d): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Mar 19 11:54:45.495930 ignition[967]: INFO : files: op(c): [finished] processing unit "prepare-helm.service" Mar 19 11:54:45.495930 ignition[967]: INFO : files: op(e): [started] processing unit "coreos-metadata.service" Mar 19 11:54:45.495930 ignition[967]: INFO : files: op(e): op(f): [started] writing systemd drop-in "00-custom-metadata.conf" at "/sysroot/etc/systemd/system/coreos-metadata.service.d/00-custom-metadata.conf" Mar 19 11:54:45.495930 ignition[967]: INFO : files: op(e): op(f): [finished] writing systemd drop-in "00-custom-metadata.conf" at "/sysroot/etc/systemd/system/coreos-metadata.service.d/00-custom-metadata.conf" Mar 19 11:54:45.495930 ignition[967]: INFO : files: op(e): [finished] processing unit "coreos-metadata.service" Mar 19 11:54:45.495930 ignition[967]: INFO : files: op(10): [started] setting preset to enabled for "prepare-helm.service" Mar 19 11:54:45.495930 ignition[967]: INFO : files: op(10): [finished] setting preset to enabled for "prepare-helm.service" Mar 19 11:54:45.495930 ignition[967]: INFO : files: createResultFile: createFiles: op(11): [started] writing file "/sysroot/etc/.ignition-result.json" Mar 19 11:54:45.495930 ignition[967]: INFO : files: createResultFile: createFiles: op(11): [finished] writing file "/sysroot/etc/.ignition-result.json" Mar 19 11:54:45.495930 ignition[967]: INFO : files: files passed Mar 19 11:54:45.495930 ignition[967]: INFO : Ignition finished successfully Mar 19 11:54:45.496910 systemd[1]: Finished ignition-files.service - Ignition (files). Mar 19 11:54:45.510701 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Mar 19 11:54:45.524654 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Mar 19 11:54:45.528900 systemd[1]: ignition-quench.service: Deactivated successfully. Mar 19 11:54:45.529037 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Mar 19 11:54:45.542369 initrd-setup-root-after-ignition[1001]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Mar 19 11:54:45.547248 initrd-setup-root-after-ignition[997]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Mar 19 11:54:45.547248 initrd-setup-root-after-ignition[997]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Mar 19 11:54:45.545077 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Mar 19 11:54:45.546882 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Mar 19 11:54:45.553612 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Mar 19 11:54:45.582197 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Mar 19 11:54:45.582384 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Mar 19 11:54:45.584542 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Mar 19 11:54:45.585949 systemd[1]: Reached target initrd.target - Initrd Default Target. Mar 19 11:54:45.586842 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Mar 19 11:54:45.591552 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Mar 19 11:54:45.606464 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Mar 19 11:54:45.614466 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Mar 19 11:54:45.630018 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Mar 19 11:54:45.631265 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 19 11:54:45.633262 systemd[1]: Stopped target timers.target - Timer Units. Mar 19 11:54:45.635306 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Mar 19 11:54:45.635560 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Mar 19 11:54:45.637390 systemd[1]: Stopped target initrd.target - Initrd Default Target. Mar 19 11:54:45.638768 systemd[1]: Stopped target basic.target - Basic System. Mar 19 11:54:45.640540 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Mar 19 11:54:45.642235 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Mar 19 11:54:45.643888 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Mar 19 11:54:45.645651 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Mar 19 11:54:45.647639 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Mar 19 11:54:45.649628 systemd[1]: Stopped target sysinit.target - System Initialization. Mar 19 11:54:45.651516 systemd[1]: Stopped target local-fs.target - Local File Systems. Mar 19 11:54:45.653457 systemd[1]: Stopped target swap.target - Swaps. Mar 19 11:54:45.655152 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Mar 19 11:54:45.655433 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Mar 19 11:54:45.657329 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Mar 19 11:54:45.658651 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 19 11:54:45.660420 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Mar 19 11:54:45.662385 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 19 11:54:45.664625 systemd[1]: dracut-initqueue.service: Deactivated successfully. Mar 19 11:54:45.664919 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Mar 19 11:54:45.667308 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Mar 19 11:54:45.667509 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Mar 19 11:54:45.669642 systemd[1]: ignition-files.service: Deactivated successfully. Mar 19 11:54:45.669890 systemd[1]: Stopped ignition-files.service - Ignition (files). Mar 19 11:54:45.671566 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Mar 19 11:54:45.671885 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Mar 19 11:54:45.682023 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Mar 19 11:54:45.682874 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Mar 19 11:54:45.683105 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Mar 19 11:54:45.685597 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Mar 19 11:54:45.689859 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Mar 19 11:54:45.690150 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Mar 19 11:54:45.693389 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Mar 19 11:54:45.693550 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Mar 19 11:54:45.703896 systemd[1]: initrd-cleanup.service: Deactivated successfully. Mar 19 11:54:45.704138 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Mar 19 11:54:45.714670 ignition[1021]: INFO : Ignition 2.20.0 Mar 19 11:54:45.714670 ignition[1021]: INFO : Stage: umount Mar 19 11:54:45.716324 ignition[1021]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 19 11:54:45.716324 ignition[1021]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Mar 19 11:54:45.718801 ignition[1021]: INFO : umount: umount passed Mar 19 11:54:45.718801 ignition[1021]: INFO : Ignition finished successfully Mar 19 11:54:45.721230 systemd[1]: ignition-mount.service: Deactivated successfully. Mar 19 11:54:45.722386 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Mar 19 11:54:45.724688 systemd[1]: ignition-disks.service: Deactivated successfully. Mar 19 11:54:45.724795 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Mar 19 11:54:45.726537 systemd[1]: ignition-kargs.service: Deactivated successfully. Mar 19 11:54:45.726594 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Mar 19 11:54:45.727662 systemd[1]: ignition-fetch.service: Deactivated successfully. Mar 19 11:54:45.727715 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Mar 19 11:54:45.728465 systemd[1]: Stopped target network.target - Network. Mar 19 11:54:45.729116 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Mar 19 11:54:45.729170 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Mar 19 11:54:45.732471 systemd[1]: Stopped target paths.target - Path Units. Mar 19 11:54:45.733722 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Mar 19 11:54:45.739399 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 19 11:54:45.741167 systemd[1]: Stopped target slices.target - Slice Units. Mar 19 11:54:45.741644 systemd[1]: Stopped target sockets.target - Socket Units. Mar 19 11:54:45.742898 systemd[1]: iscsid.socket: Deactivated successfully. Mar 19 11:54:45.742940 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Mar 19 11:54:45.744118 systemd[1]: iscsiuio.socket: Deactivated successfully. Mar 19 11:54:45.744148 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Mar 19 11:54:45.745306 systemd[1]: ignition-setup.service: Deactivated successfully. Mar 19 11:54:45.745357 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Mar 19 11:54:45.746530 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Mar 19 11:54:45.746565 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Mar 19 11:54:45.747837 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Mar 19 11:54:45.749154 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Mar 19 11:54:45.751727 systemd[1]: sysroot-boot.mount: Deactivated successfully. Mar 19 11:54:45.752261 systemd[1]: sysroot-boot.service: Deactivated successfully. Mar 19 11:54:45.752373 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Mar 19 11:54:45.754468 systemd[1]: initrd-setup-root.service: Deactivated successfully. Mar 19 11:54:45.754533 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Mar 19 11:54:45.756976 systemd[1]: systemd-resolved.service: Deactivated successfully. Mar 19 11:54:45.757063 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Mar 19 11:54:45.761617 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. Mar 19 11:54:45.761900 systemd[1]: systemd-networkd.service: Deactivated successfully. Mar 19 11:54:45.761987 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Mar 19 11:54:45.764133 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. Mar 19 11:54:45.764921 systemd[1]: systemd-networkd.socket: Deactivated successfully. Mar 19 11:54:45.764967 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Mar 19 11:54:45.773428 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Mar 19 11:54:45.773922 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Mar 19 11:54:45.773984 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Mar 19 11:54:45.776239 systemd[1]: systemd-sysctl.service: Deactivated successfully. Mar 19 11:54:45.776310 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Mar 19 11:54:45.778437 systemd[1]: systemd-modules-load.service: Deactivated successfully. Mar 19 11:54:45.778480 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Mar 19 11:54:45.779344 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Mar 19 11:54:45.779384 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 19 11:54:45.781320 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 19 11:54:45.783069 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Mar 19 11:54:45.783129 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Mar 19 11:54:45.790561 systemd[1]: network-cleanup.service: Deactivated successfully. Mar 19 11:54:45.790647 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Mar 19 11:54:45.793841 systemd[1]: systemd-udevd.service: Deactivated successfully. Mar 19 11:54:45.794018 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 19 11:54:45.795930 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Mar 19 11:54:45.795998 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Mar 19 11:54:45.797417 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Mar 19 11:54:45.797456 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Mar 19 11:54:45.798696 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Mar 19 11:54:45.798751 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Mar 19 11:54:45.800534 systemd[1]: dracut-cmdline.service: Deactivated successfully. Mar 19 11:54:45.800589 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Mar 19 11:54:45.801790 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Mar 19 11:54:45.801847 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 19 11:54:45.813599 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Mar 19 11:54:45.814193 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Mar 19 11:54:45.814271 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 19 11:54:45.817260 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 19 11:54:45.817352 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 19 11:54:45.818764 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. Mar 19 11:54:45.818831 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Mar 19 11:54:45.820555 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Mar 19 11:54:45.820642 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Mar 19 11:54:45.822904 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Mar 19 11:54:45.829722 systemd[1]: Starting initrd-switch-root.service - Switch Root... Mar 19 11:54:45.838347 systemd[1]: Switching root. Mar 19 11:54:45.875558 systemd-journald[189]: Journal stopped Mar 19 11:54:47.040560 systemd-journald[189]: Received SIGTERM from PID 1 (systemd). Mar 19 11:54:47.040606 kernel: SELinux: policy capability network_peer_controls=1 Mar 19 11:54:47.040623 kernel: SELinux: policy capability open_perms=1 Mar 19 11:54:47.040632 kernel: SELinux: policy capability extended_socket_class=1 Mar 19 11:54:47.040641 kernel: SELinux: policy capability always_check_network=0 Mar 19 11:54:47.040653 kernel: SELinux: policy capability cgroup_seclabel=1 Mar 19 11:54:47.040662 kernel: SELinux: policy capability nnp_nosuid_transition=1 Mar 19 11:54:47.040671 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Mar 19 11:54:47.040680 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Mar 19 11:54:47.040691 kernel: audit: type=1403 audit(1742385286.078:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Mar 19 11:54:47.040704 systemd[1]: Successfully loaded SELinux policy in 62.114ms. Mar 19 11:54:47.040716 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 15.978ms. Mar 19 11:54:47.040727 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Mar 19 11:54:47.040737 systemd[1]: Detected virtualization kvm. Mar 19 11:54:47.040749 systemd[1]: Detected architecture x86-64. Mar 19 11:54:47.040761 systemd[1]: Detected first boot. Mar 19 11:54:47.040771 systemd[1]: Hostname set to . Mar 19 11:54:47.040794 systemd[1]: Initializing machine ID from VM UUID. Mar 19 11:54:47.040805 zram_generator::config[1066]: No configuration found. Mar 19 11:54:47.040817 kernel: Guest personality initialized and is inactive Mar 19 11:54:47.040827 kernel: VMCI host device registered (name=vmci, major=10, minor=125) Mar 19 11:54:47.040837 kernel: Initialized host personality Mar 19 11:54:47.040847 kernel: NET: Registered PF_VSOCK protocol family Mar 19 11:54:47.040856 systemd[1]: Populated /etc with preset unit settings. Mar 19 11:54:47.040867 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. Mar 19 11:54:47.040877 systemd[1]: initrd-switch-root.service: Deactivated successfully. Mar 19 11:54:47.040887 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Mar 19 11:54:47.040897 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Mar 19 11:54:47.040907 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Mar 19 11:54:47.040917 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Mar 19 11:54:47.040929 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Mar 19 11:54:47.040938 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Mar 19 11:54:47.040949 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Mar 19 11:54:47.040959 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Mar 19 11:54:47.040969 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Mar 19 11:54:47.040979 systemd[1]: Created slice user.slice - User and Session Slice. Mar 19 11:54:47.040989 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 19 11:54:47.040999 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 19 11:54:47.041009 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Mar 19 11:54:47.041021 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Mar 19 11:54:47.041031 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Mar 19 11:54:47.041042 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Mar 19 11:54:47.041052 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Mar 19 11:54:47.041062 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 19 11:54:47.041072 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Mar 19 11:54:47.041083 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Mar 19 11:54:47.041093 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Mar 19 11:54:47.041103 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Mar 19 11:54:47.041113 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 19 11:54:47.041123 systemd[1]: Reached target remote-fs.target - Remote File Systems. Mar 19 11:54:47.041134 systemd[1]: Reached target slices.target - Slice Units. Mar 19 11:54:47.041147 systemd[1]: Reached target swap.target - Swaps. Mar 19 11:54:47.041157 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Mar 19 11:54:47.041169 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Mar 19 11:54:47.041179 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Mar 19 11:54:47.041189 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Mar 19 11:54:47.041199 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Mar 19 11:54:47.041209 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Mar 19 11:54:47.041219 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Mar 19 11:54:47.041228 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Mar 19 11:54:47.041240 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Mar 19 11:54:47.041250 systemd[1]: Mounting media.mount - External Media Directory... Mar 19 11:54:47.041261 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 19 11:54:47.041271 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Mar 19 11:54:47.063382 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Mar 19 11:54:47.063404 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Mar 19 11:54:47.063419 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Mar 19 11:54:47.063430 systemd[1]: Reached target machines.target - Containers. Mar 19 11:54:47.063443 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Mar 19 11:54:47.063454 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 19 11:54:47.063464 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Mar 19 11:54:47.063475 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Mar 19 11:54:47.063485 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 19 11:54:47.063495 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Mar 19 11:54:47.063505 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 19 11:54:47.063516 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Mar 19 11:54:47.063526 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 19 11:54:47.063539 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Mar 19 11:54:47.063550 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Mar 19 11:54:47.063560 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Mar 19 11:54:47.063570 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Mar 19 11:54:47.063580 systemd[1]: Stopped systemd-fsck-usr.service. Mar 19 11:54:47.063590 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Mar 19 11:54:47.063600 kernel: loop: module loaded Mar 19 11:54:47.063610 systemd[1]: Starting systemd-journald.service - Journal Service... Mar 19 11:54:47.063622 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Mar 19 11:54:47.063632 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Mar 19 11:54:47.063642 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Mar 19 11:54:47.063653 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Mar 19 11:54:47.063663 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Mar 19 11:54:47.063673 kernel: fuse: init (API version 7.39) Mar 19 11:54:47.063683 systemd[1]: verity-setup.service: Deactivated successfully. Mar 19 11:54:47.063693 systemd[1]: Stopped verity-setup.service. Mar 19 11:54:47.063707 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 19 11:54:47.063720 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Mar 19 11:54:47.063731 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Mar 19 11:54:47.063742 systemd[1]: Mounted media.mount - External Media Directory. Mar 19 11:54:47.063752 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Mar 19 11:54:47.063763 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Mar 19 11:54:47.063773 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Mar 19 11:54:47.063793 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Mar 19 11:54:47.063804 systemd[1]: modprobe@configfs.service: Deactivated successfully. Mar 19 11:54:47.063814 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Mar 19 11:54:47.063848 systemd-journald[1143]: Collecting audit messages is disabled. Mar 19 11:54:47.063872 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 19 11:54:47.063882 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 19 11:54:47.063892 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 19 11:54:47.063903 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 19 11:54:47.063913 systemd[1]: modprobe@fuse.service: Deactivated successfully. Mar 19 11:54:47.063923 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Mar 19 11:54:47.063933 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 19 11:54:47.063945 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 19 11:54:47.063957 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Mar 19 11:54:47.063967 systemd-journald[1143]: Journal started Mar 19 11:54:47.063989 systemd-journald[1143]: Runtime Journal (/run/log/journal/8f6e0c84addd436cb2eccb830195f8d0) is 4.8M, max 38.3M, 33.5M free. Mar 19 11:54:46.711859 systemd[1]: Queued start job for default target multi-user.target. Mar 19 11:54:46.722115 systemd[1]: Unnecessary job was removed for dev-sda6.device - /dev/sda6. Mar 19 11:54:46.723023 systemd[1]: systemd-journald.service: Deactivated successfully. Mar 19 11:54:47.077342 systemd[1]: Started systemd-journald.service - Journal Service. Mar 19 11:54:47.070015 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Mar 19 11:54:47.070771 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Mar 19 11:54:47.072975 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Mar 19 11:54:47.084238 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Mar 19 11:54:47.096612 systemd[1]: Reached target network-pre.target - Preparation for Network. Mar 19 11:54:47.111301 kernel: ACPI: bus type drm_connector registered Mar 19 11:54:47.112573 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Mar 19 11:54:47.115628 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Mar 19 11:54:47.116194 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Mar 19 11:54:47.116229 systemd[1]: Reached target local-fs.target - Local File Systems. Mar 19 11:54:47.117814 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Mar 19 11:54:47.120070 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Mar 19 11:54:47.123119 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Mar 19 11:54:47.124081 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 19 11:54:47.126032 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Mar 19 11:54:47.128661 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Mar 19 11:54:47.131446 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 19 11:54:47.134371 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Mar 19 11:54:47.134915 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Mar 19 11:54:47.136436 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Mar 19 11:54:47.140657 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Mar 19 11:54:47.142806 systemd[1]: Starting systemd-sysusers.service - Create System Users... Mar 19 11:54:47.146248 systemd[1]: modprobe@drm.service: Deactivated successfully. Mar 19 11:54:47.149850 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Mar 19 11:54:47.151662 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Mar 19 11:54:47.152264 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Mar 19 11:54:47.152802 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Mar 19 11:54:47.156325 systemd-journald[1143]: Time spent on flushing to /var/log/journal/8f6e0c84addd436cb2eccb830195f8d0 is 21.260ms for 1149 entries. Mar 19 11:54:47.156325 systemd-journald[1143]: System Journal (/var/log/journal/8f6e0c84addd436cb2eccb830195f8d0) is 8M, max 584.8M, 576.8M free. Mar 19 11:54:47.206039 systemd-journald[1143]: Received client request to flush runtime journal. Mar 19 11:54:47.206075 kernel: loop0: detected capacity change from 0 to 205544 Mar 19 11:54:47.154626 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Mar 19 11:54:47.172421 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Mar 19 11:54:47.179387 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Mar 19 11:54:47.182368 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Mar 19 11:54:47.195679 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Mar 19 11:54:47.196531 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Mar 19 11:54:47.207923 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Mar 19 11:54:47.225014 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Mar 19 11:54:47.228477 udevadm[1199]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Mar 19 11:54:47.240453 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Mar 19 11:54:47.262539 systemd[1]: Finished systemd-sysusers.service - Create System Users. Mar 19 11:54:47.271529 kernel: loop1: detected capacity change from 0 to 147912 Mar 19 11:54:47.271601 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Mar 19 11:54:47.297768 systemd-tmpfiles[1211]: ACLs are not supported, ignoring. Mar 19 11:54:47.297808 systemd-tmpfiles[1211]: ACLs are not supported, ignoring. Mar 19 11:54:47.304843 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 19 11:54:47.318629 kernel: loop2: detected capacity change from 0 to 138176 Mar 19 11:54:47.365322 kernel: loop3: detected capacity change from 0 to 8 Mar 19 11:54:47.390330 kernel: loop4: detected capacity change from 0 to 205544 Mar 19 11:54:47.424554 kernel: loop5: detected capacity change from 0 to 147912 Mar 19 11:54:47.450353 kernel: loop6: detected capacity change from 0 to 138176 Mar 19 11:54:47.477267 kernel: loop7: detected capacity change from 0 to 8 Mar 19 11:54:47.478817 (sd-merge)[1217]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-hetzner'. Mar 19 11:54:47.479930 (sd-merge)[1217]: Merged extensions into '/usr'. Mar 19 11:54:47.488470 systemd[1]: Reload requested from client PID 1191 ('systemd-sysext') (unit systemd-sysext.service)... Mar 19 11:54:47.488631 systemd[1]: Reloading... Mar 19 11:54:47.581338 zram_generator::config[1248]: No configuration found. Mar 19 11:54:47.740158 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 19 11:54:47.780607 ldconfig[1186]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Mar 19 11:54:47.826452 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Mar 19 11:54:47.826752 systemd[1]: Reloading finished in 335 ms. Mar 19 11:54:47.842162 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Mar 19 11:54:47.843033 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Mar 19 11:54:47.854444 systemd[1]: Starting ensure-sysext.service... Mar 19 11:54:47.858410 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Mar 19 11:54:47.875367 systemd[1]: Reload requested from client PID 1288 ('systemctl') (unit ensure-sysext.service)... Mar 19 11:54:47.875377 systemd[1]: Reloading... Mar 19 11:54:47.894217 systemd-tmpfiles[1289]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Mar 19 11:54:47.894741 systemd-tmpfiles[1289]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Mar 19 11:54:47.895494 systemd-tmpfiles[1289]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Mar 19 11:54:47.896353 systemd-tmpfiles[1289]: ACLs are not supported, ignoring. Mar 19 11:54:47.896403 systemd-tmpfiles[1289]: ACLs are not supported, ignoring. Mar 19 11:54:47.903633 systemd-tmpfiles[1289]: Detected autofs mount point /boot during canonicalization of boot. Mar 19 11:54:47.903761 systemd-tmpfiles[1289]: Skipping /boot Mar 19 11:54:47.916554 systemd-tmpfiles[1289]: Detected autofs mount point /boot during canonicalization of boot. Mar 19 11:54:47.916565 systemd-tmpfiles[1289]: Skipping /boot Mar 19 11:54:47.974311 zram_generator::config[1326]: No configuration found. Mar 19 11:54:48.068510 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 19 11:54:48.139192 systemd[1]: Reloading finished in 263 ms. Mar 19 11:54:48.149737 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Mar 19 11:54:48.158870 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 19 11:54:48.169476 systemd[1]: Starting audit-rules.service - Load Audit Rules... Mar 19 11:54:48.177348 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Mar 19 11:54:48.182468 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Mar 19 11:54:48.186444 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Mar 19 11:54:48.190712 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 19 11:54:48.197562 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Mar 19 11:54:48.216397 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 19 11:54:48.216707 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 19 11:54:48.224079 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 19 11:54:48.227908 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 19 11:54:48.235552 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 19 11:54:48.236614 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 19 11:54:48.236845 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Mar 19 11:54:48.246570 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Mar 19 11:54:48.247650 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 19 11:54:48.250977 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Mar 19 11:54:48.258182 systemd-udevd[1367]: Using default interface naming scheme 'v255'. Mar 19 11:54:48.260351 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 19 11:54:48.260544 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 19 11:54:48.261967 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 19 11:54:48.262105 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Mar 19 11:54:48.262237 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 19 11:54:48.265777 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 19 11:54:48.266171 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 19 11:54:48.267771 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 19 11:54:48.268461 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 19 11:54:48.276385 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Mar 19 11:54:48.283753 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 19 11:54:48.284274 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 19 11:54:48.287004 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 19 11:54:48.287254 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 19 11:54:48.292534 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 19 11:54:48.299960 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Mar 19 11:54:48.305874 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 19 11:54:48.306423 augenrules[1399]: No rules Mar 19 11:54:48.307559 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 19 11:54:48.307710 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Mar 19 11:54:48.307880 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 19 11:54:48.311685 systemd[1]: Starting systemd-update-done.service - Update is Completed... Mar 19 11:54:48.313400 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 19 11:54:48.316227 systemd[1]: audit-rules.service: Deactivated successfully. Mar 19 11:54:48.316489 systemd[1]: Finished audit-rules.service - Load Audit Rules. Mar 19 11:54:48.320644 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 19 11:54:48.320849 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 19 11:54:48.322049 systemd[1]: modprobe@drm.service: Deactivated successfully. Mar 19 11:54:48.322228 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Mar 19 11:54:48.324124 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 19 11:54:48.333049 systemd[1]: Finished ensure-sysext.service. Mar 19 11:54:48.338729 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 19 11:54:48.338940 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 19 11:54:48.351908 systemd[1]: Starting systemd-networkd.service - Network Configuration... Mar 19 11:54:48.352424 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Mar 19 11:54:48.356096 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Mar 19 11:54:48.356831 systemd[1]: Started systemd-userdbd.service - User Database Manager. Mar 19 11:54:48.358520 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Mar 19 11:54:48.359563 systemd[1]: Finished systemd-update-done.service - Update is Completed. Mar 19 11:54:48.370586 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Mar 19 11:54:48.408114 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Mar 19 11:54:48.486475 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Mar 19 11:54:48.521306 kernel: mousedev: PS/2 mouse device common for all mice Mar 19 11:54:48.526346 kernel: ACPI: button: Power Button [PWRF] Mar 19 11:54:48.529658 systemd-networkd[1415]: lo: Link UP Mar 19 11:54:48.529671 systemd-networkd[1415]: lo: Gained carrier Mar 19 11:54:48.531614 systemd-networkd[1415]: Enumeration completed Mar 19 11:54:48.531695 systemd[1]: Started systemd-networkd.service - Network Configuration. Mar 19 11:54:48.543462 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Mar 19 11:54:48.547018 systemd-resolved[1366]: Positive Trust Anchors: Mar 19 11:54:48.547040 systemd-resolved[1366]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Mar 19 11:54:48.547072 systemd-resolved[1366]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Mar 19 11:54:48.552467 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Mar 19 11:54:48.553122 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Mar 19 11:54:48.553823 systemd[1]: Reached target time-set.target - System Time Set. Mar 19 11:54:48.556567 systemd-resolved[1366]: Using system hostname 'ci-4230-1-0-d-e284590f49'. Mar 19 11:54:48.557981 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Mar 19 11:54:48.558672 systemd[1]: Reached target network.target - Network. Mar 19 11:54:48.559174 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Mar 19 11:54:48.567468 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Mar 19 11:54:48.567893 systemd-networkd[1415]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 19 11:54:48.567898 systemd-networkd[1415]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 19 11:54:48.568463 systemd-networkd[1415]: eth1: Link UP Mar 19 11:54:48.568466 systemd-networkd[1415]: eth1: Gained carrier Mar 19 11:54:48.568483 systemd-networkd[1415]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 19 11:54:48.574095 systemd[1]: Condition check resulted in dev-virtio\x2dports-org.qemu.guest_agent.0.device - /dev/virtio-ports/org.qemu.guest_agent.0 being skipped. Mar 19 11:54:48.574153 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 19 11:54:48.574245 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 19 11:54:48.582053 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 19 11:54:48.585486 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 19 11:54:48.591663 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 19 11:54:48.592235 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 19 11:54:48.592264 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Mar 19 11:54:48.592306 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Mar 19 11:54:48.592324 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 19 11:54:48.592663 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 19 11:54:48.593156 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 19 11:54:48.604407 systemd-networkd[1415]: eth1: DHCPv4 address 10.0.0.3/32, gateway 10.0.0.1 acquired from 10.0.0.1 Mar 19 11:54:48.605057 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 19 11:54:48.605260 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 19 11:54:48.606538 systemd-timesyncd[1416]: Network configuration changed, trying to establish connection. Mar 19 11:54:48.606888 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 19 11:54:48.609623 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 19 11:54:48.609913 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 19 11:54:48.612212 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Mar 19 11:54:48.614146 systemd-networkd[1415]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 19 11:54:48.614160 systemd-networkd[1415]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 19 11:54:48.616067 systemd-networkd[1415]: eth0: Link UP Mar 19 11:54:48.616078 systemd-networkd[1415]: eth0: Gained carrier Mar 19 11:54:48.616094 systemd-networkd[1415]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 19 11:54:48.616263 systemd-timesyncd[1416]: Network configuration changed, trying to establish connection. Mar 19 11:54:48.619740 systemd-timesyncd[1416]: Network configuration changed, trying to establish connection. Mar 19 11:54:48.641344 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Mar 19 11:54:48.647981 kernel: i2c i2c-0: 1/1 memory slots populated (from DMI) Mar 19 11:54:48.648138 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Mar 19 11:54:48.652317 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 36 scanned by (udev-worker) (1407) Mar 19 11:54:48.660679 systemd-networkd[1415]: eth0: DHCPv4 address 37.27.246.226/32, gateway 172.31.1.1 acquired from 172.31.1.1 Mar 19 11:54:48.661268 systemd-timesyncd[1416]: Network configuration changed, trying to establish connection. Mar 19 11:54:48.677435 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input4 Mar 19 11:54:48.677503 kernel: EDAC MC: Ver: 3.0.0 Mar 19 11:54:48.700371 kernel: [drm] pci: virtio-vga detected at 0000:00:01.0 Mar 19 11:54:48.712309 kernel: virtio-pci 0000:00:01.0: vgaarb: deactivate vga console Mar 19 11:54:48.718303 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - QEMU_HARDDISK OEM. Mar 19 11:54:48.720394 kernel: Console: switching to colour dummy device 80x25 Mar 19 11:54:48.720446 kernel: [drm] features: -virgl +edid -resource_blob -host_visible Mar 19 11:54:48.720461 kernel: [drm] features: -context_init Mar 19 11:54:48.723636 kernel: [drm] number of scanouts: 1 Mar 19 11:54:48.723685 kernel: [drm] number of cap sets: 0 Mar 19 11:54:48.729340 kernel: [drm] Initialized virtio_gpu 0.1.0 0 for 0000:00:01.0 on minor 0 Mar 19 11:54:48.729690 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Mar 19 11:54:48.734257 kernel: fbcon: virtio_gpudrmfb (fb0) is primary device Mar 19 11:54:48.734359 kernel: Console: switching to colour frame buffer device 160x50 Mar 19 11:54:48.739307 kernel: virtio-pci 0000:00:01.0: [drm] fb0: virtio_gpudrmfb frame buffer device Mar 19 11:54:48.752492 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 19 11:54:48.752837 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Mar 19 11:54:48.763417 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 19 11:54:48.763675 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 19 11:54:48.769475 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 19 11:54:48.831700 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 19 11:54:48.879032 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Mar 19 11:54:48.886580 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Mar 19 11:54:48.899895 lvm[1481]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Mar 19 11:54:48.932890 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Mar 19 11:54:48.934214 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Mar 19 11:54:48.935375 systemd[1]: Reached target sysinit.target - System Initialization. Mar 19 11:54:48.935641 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Mar 19 11:54:48.935759 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Mar 19 11:54:48.936134 systemd[1]: Started logrotate.timer - Daily rotation of log files. Mar 19 11:54:48.936741 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Mar 19 11:54:48.938049 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Mar 19 11:54:48.938183 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Mar 19 11:54:48.938220 systemd[1]: Reached target paths.target - Path Units. Mar 19 11:54:48.938470 systemd[1]: Reached target timers.target - Timer Units. Mar 19 11:54:48.942461 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Mar 19 11:54:48.945195 systemd[1]: Starting docker.socket - Docker Socket for the API... Mar 19 11:54:48.951344 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Mar 19 11:54:48.951957 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Mar 19 11:54:48.952129 systemd[1]: Reached target ssh-access.target - SSH Access Available. Mar 19 11:54:48.956307 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Mar 19 11:54:48.960149 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Mar 19 11:54:48.968590 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Mar 19 11:54:48.971485 systemd[1]: Listening on docker.socket - Docker Socket for the API. Mar 19 11:54:48.973909 systemd[1]: Reached target sockets.target - Socket Units. Mar 19 11:54:48.976138 systemd[1]: Reached target basic.target - Basic System. Mar 19 11:54:48.978198 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Mar 19 11:54:48.978857 lvm[1485]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Mar 19 11:54:48.979273 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Mar 19 11:54:48.987519 systemd[1]: Starting containerd.service - containerd container runtime... Mar 19 11:54:48.992511 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Mar 19 11:54:48.998000 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Mar 19 11:54:49.008021 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Mar 19 11:54:49.011440 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Mar 19 11:54:49.015558 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Mar 19 11:54:49.023621 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Mar 19 11:54:49.031222 jq[1489]: false Mar 19 11:54:49.028478 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Mar 19 11:54:49.032648 systemd[1]: Started qemu-guest-agent.service - QEMU Guest Agent. Mar 19 11:54:49.047275 extend-filesystems[1490]: Found loop4 Mar 19 11:54:49.047275 extend-filesystems[1490]: Found loop5 Mar 19 11:54:49.047275 extend-filesystems[1490]: Found loop6 Mar 19 11:54:49.047275 extend-filesystems[1490]: Found loop7 Mar 19 11:54:49.047275 extend-filesystems[1490]: Found sda Mar 19 11:54:49.047275 extend-filesystems[1490]: Found sda1 Mar 19 11:54:49.047275 extend-filesystems[1490]: Found sda2 Mar 19 11:54:49.047275 extend-filesystems[1490]: Found sda3 Mar 19 11:54:49.047275 extend-filesystems[1490]: Found usr Mar 19 11:54:49.047275 extend-filesystems[1490]: Found sda4 Mar 19 11:54:49.047275 extend-filesystems[1490]: Found sda6 Mar 19 11:54:49.047275 extend-filesystems[1490]: Found sda7 Mar 19 11:54:49.047275 extend-filesystems[1490]: Found sda9 Mar 19 11:54:49.047275 extend-filesystems[1490]: Checking size of /dev/sda9 Mar 19 11:54:49.044351 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Mar 19 11:54:49.049157 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Mar 19 11:54:49.072558 systemd[1]: Starting systemd-logind.service - User Login Management... Mar 19 11:54:49.083743 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Mar 19 11:54:49.086707 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Mar 19 11:54:49.095563 systemd[1]: Starting update-engine.service - Update Engine... Mar 19 11:54:49.109474 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Mar 19 11:54:49.114484 extend-filesystems[1490]: Resized partition /dev/sda9 Mar 19 11:54:49.114636 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Mar 19 11:54:49.126347 jq[1509]: true Mar 19 11:54:49.126664 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Mar 19 11:54:49.127344 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Mar 19 11:54:49.131037 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Mar 19 11:54:49.132188 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Mar 19 11:54:49.134772 extend-filesystems[1519]: resize2fs 1.47.1 (20-May-2024) Mar 19 11:54:49.143754 kernel: EXT4-fs (sda9): resizing filesystem from 1617920 to 9393147 blocks Mar 19 11:54:49.134864 dbus-daemon[1488]: [system] SELinux support is enabled Mar 19 11:54:49.146766 systemd[1]: Started dbus.service - D-Bus System Message Bus. Mar 19 11:54:49.148540 coreos-metadata[1487]: Mar 19 11:54:49.148 INFO Fetching http://169.254.169.254/hetzner/v1/metadata: Attempt #1 Mar 19 11:54:49.162028 coreos-metadata[1487]: Mar 19 11:54:49.161 INFO Fetch successful Mar 19 11:54:49.162028 coreos-metadata[1487]: Mar 19 11:54:49.161 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/private-networks: Attempt #1 Mar 19 11:54:49.162028 coreos-metadata[1487]: Mar 19 11:54:49.161 INFO Fetch successful Mar 19 11:54:49.176000 jq[1521]: true Mar 19 11:54:49.175972 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Mar 19 11:54:49.176001 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Mar 19 11:54:49.178412 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Mar 19 11:54:49.178440 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Mar 19 11:54:49.186482 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 36 scanned by (udev-worker) (1430) Mar 19 11:54:49.186600 tar[1518]: linux-amd64/helm Mar 19 11:54:49.191718 update_engine[1506]: I20250319 11:54:49.191390 1506 main.cc:92] Flatcar Update Engine starting Mar 19 11:54:49.194945 systemd[1]: motdgen.service: Deactivated successfully. Mar 19 11:54:49.195195 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Mar 19 11:54:49.204479 update_engine[1506]: I20250319 11:54:49.204402 1506 update_check_scheduler.cc:74] Next update check in 9m4s Mar 19 11:54:49.210994 systemd[1]: Started update-engine.service - Update Engine. Mar 19 11:54:49.219722 (ntainerd)[1531]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Mar 19 11:54:49.220909 systemd[1]: Started locksmithd.service - Cluster reboot manager. Mar 19 11:54:49.333866 bash[1553]: Updated "/home/core/.ssh/authorized_keys" Mar 19 11:54:49.338949 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Mar 19 11:54:49.347542 kernel: EXT4-fs (sda9): resized filesystem to 9393147 Mar 19 11:54:49.356460 systemd[1]: Starting sshkeys.service... Mar 19 11:54:49.378327 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Mar 19 11:54:49.388741 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Mar 19 11:54:49.396541 systemd-logind[1502]: New seat seat0. Mar 19 11:54:49.398668 systemd-logind[1502]: Watching system buttons on /dev/input/event2 (Power Button) Mar 19 11:54:49.398770 systemd-logind[1502]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Mar 19 11:54:49.399059 systemd[1]: Started systemd-logind.service - User Login Management. Mar 19 11:54:49.402844 extend-filesystems[1519]: Filesystem at /dev/sda9 is mounted on /; on-line resizing required Mar 19 11:54:49.402844 extend-filesystems[1519]: old_desc_blocks = 1, new_desc_blocks = 5 Mar 19 11:54:49.402844 extend-filesystems[1519]: The filesystem on /dev/sda9 is now 9393147 (4k) blocks long. Mar 19 11:54:49.406867 extend-filesystems[1490]: Resized filesystem in /dev/sda9 Mar 19 11:54:49.406867 extend-filesystems[1490]: Found sr0 Mar 19 11:54:49.410063 systemd[1]: extend-filesystems.service: Deactivated successfully. Mar 19 11:54:49.410492 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Mar 19 11:54:49.424047 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Mar 19 11:54:49.432126 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Mar 19 11:54:49.520543 coreos-metadata[1564]: Mar 19 11:54:49.519 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/public-keys: Attempt #1 Mar 19 11:54:49.522600 coreos-metadata[1564]: Mar 19 11:54:49.521 INFO Fetch successful Mar 19 11:54:49.526120 unknown[1564]: wrote ssh authorized keys file for user: core Mar 19 11:54:49.558711 update-ssh-keys[1574]: Updated "/home/core/.ssh/authorized_keys" Mar 19 11:54:49.554935 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Mar 19 11:54:49.560613 systemd[1]: Finished sshkeys.service. Mar 19 11:54:49.564189 sshd_keygen[1520]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Mar 19 11:54:49.576212 locksmithd[1539]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Mar 19 11:54:49.612010 containerd[1531]: time="2025-03-19T11:54:49.611904610Z" level=info msg="starting containerd" revision=9b2ad7760328148397346d10c7b2004271249db4 version=v1.7.23 Mar 19 11:54:49.630885 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Mar 19 11:54:49.643081 systemd[1]: Starting issuegen.service - Generate /run/issue... Mar 19 11:54:49.667978 containerd[1531]: time="2025-03-19T11:54:49.667921694Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Mar 19 11:54:49.669836 systemd[1]: issuegen.service: Deactivated successfully. Mar 19 11:54:49.670024 systemd[1]: Finished issuegen.service - Generate /run/issue. Mar 19 11:54:49.670568 containerd[1531]: time="2025-03-19T11:54:49.670525476Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.83-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Mar 19 11:54:49.670634 containerd[1531]: time="2025-03-19T11:54:49.670622047Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Mar 19 11:54:49.670678 containerd[1531]: time="2025-03-19T11:54:49.670670087Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Mar 19 11:54:49.670879 containerd[1531]: time="2025-03-19T11:54:49.670861135Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Mar 19 11:54:49.670988 containerd[1531]: time="2025-03-19T11:54:49.670975099Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Mar 19 11:54:49.671099 containerd[1531]: time="2025-03-19T11:54:49.671082821Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Mar 19 11:54:49.671154 containerd[1531]: time="2025-03-19T11:54:49.671141752Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Mar 19 11:54:49.671447 containerd[1531]: time="2025-03-19T11:54:49.671424983Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Mar 19 11:54:49.671508 containerd[1531]: time="2025-03-19T11:54:49.671495996Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Mar 19 11:54:49.671556 containerd[1531]: time="2025-03-19T11:54:49.671545930Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Mar 19 11:54:49.671610 containerd[1531]: time="2025-03-19T11:54:49.671599019Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Mar 19 11:54:49.671752 containerd[1531]: time="2025-03-19T11:54:49.671738491Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Mar 19 11:54:49.671985 containerd[1531]: time="2025-03-19T11:54:49.671970956Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Mar 19 11:54:49.672145 containerd[1531]: time="2025-03-19T11:54:49.672132309Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Mar 19 11:54:49.672189 containerd[1531]: time="2025-03-19T11:54:49.672180870Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Mar 19 11:54:49.672312 containerd[1531]: time="2025-03-19T11:54:49.672299423Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Mar 19 11:54:49.672399 containerd[1531]: time="2025-03-19T11:54:49.672387417Z" level=info msg="metadata content store policy set" policy=shared Mar 19 11:54:49.681710 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Mar 19 11:54:49.687091 containerd[1531]: time="2025-03-19T11:54:49.684011926Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Mar 19 11:54:49.687091 containerd[1531]: time="2025-03-19T11:54:49.684268877Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Mar 19 11:54:49.687091 containerd[1531]: time="2025-03-19T11:54:49.684311918Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Mar 19 11:54:49.687091 containerd[1531]: time="2025-03-19T11:54:49.684328660Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Mar 19 11:54:49.687091 containerd[1531]: time="2025-03-19T11:54:49.684342105Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Mar 19 11:54:49.687091 containerd[1531]: time="2025-03-19T11:54:49.684493569Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Mar 19 11:54:49.687091 containerd[1531]: time="2025-03-19T11:54:49.684700658Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Mar 19 11:54:49.687091 containerd[1531]: time="2025-03-19T11:54:49.684773684Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Mar 19 11:54:49.687091 containerd[1531]: time="2025-03-19T11:54:49.684797459Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Mar 19 11:54:49.687091 containerd[1531]: time="2025-03-19T11:54:49.684809381Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Mar 19 11:54:49.687091 containerd[1531]: time="2025-03-19T11:54:49.684820783Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Mar 19 11:54:49.687091 containerd[1531]: time="2025-03-19T11:54:49.684832194Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Mar 19 11:54:49.687091 containerd[1531]: time="2025-03-19T11:54:49.684843405Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Mar 19 11:54:49.687091 containerd[1531]: time="2025-03-19T11:54:49.684854876Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Mar 19 11:54:49.687506 containerd[1531]: time="2025-03-19T11:54:49.684868011Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Mar 19 11:54:49.687506 containerd[1531]: time="2025-03-19T11:54:49.684882127Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Mar 19 11:54:49.687506 containerd[1531]: time="2025-03-19T11:54:49.684892778Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Mar 19 11:54:49.687506 containerd[1531]: time="2025-03-19T11:54:49.684902506Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Mar 19 11:54:49.687506 containerd[1531]: time="2025-03-19T11:54:49.684920600Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Mar 19 11:54:49.687506 containerd[1531]: time="2025-03-19T11:54:49.684939996Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Mar 19 11:54:49.687506 containerd[1531]: time="2025-03-19T11:54:49.684950226Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Mar 19 11:54:49.687506 containerd[1531]: time="2025-03-19T11:54:49.684961286Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Mar 19 11:54:49.687506 containerd[1531]: time="2025-03-19T11:54:49.684972127Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Mar 19 11:54:49.687506 containerd[1531]: time="2025-03-19T11:54:49.684983267Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Mar 19 11:54:49.687506 containerd[1531]: time="2025-03-19T11:54:49.684994057Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Mar 19 11:54:49.687506 containerd[1531]: time="2025-03-19T11:54:49.685009286Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Mar 19 11:54:49.687506 containerd[1531]: time="2025-03-19T11:54:49.685020597Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Mar 19 11:54:49.687506 containerd[1531]: time="2025-03-19T11:54:49.685033141Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Mar 19 11:54:49.687738 containerd[1531]: time="2025-03-19T11:54:49.685042989Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Mar 19 11:54:49.687738 containerd[1531]: time="2025-03-19T11:54:49.685053018Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Mar 19 11:54:49.687738 containerd[1531]: time="2025-03-19T11:54:49.685063027Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Mar 19 11:54:49.687738 containerd[1531]: time="2025-03-19T11:54:49.685076703Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Mar 19 11:54:49.687738 containerd[1531]: time="2025-03-19T11:54:49.685094957Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Mar 19 11:54:49.687738 containerd[1531]: time="2025-03-19T11:54:49.685105597Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Mar 19 11:54:49.687738 containerd[1531]: time="2025-03-19T11:54:49.685117179Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Mar 19 11:54:49.687738 containerd[1531]: time="2025-03-19T11:54:49.685227715Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Mar 19 11:54:49.687738 containerd[1531]: time="2025-03-19T11:54:49.685243285Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Mar 19 11:54:49.687738 containerd[1531]: time="2025-03-19T11:54:49.685251671Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Mar 19 11:54:49.687738 containerd[1531]: time="2025-03-19T11:54:49.685263002Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Mar 19 11:54:49.687738 containerd[1531]: time="2025-03-19T11:54:49.685270506Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Mar 19 11:54:49.687738 containerd[1531]: time="2025-03-19T11:54:49.685425938Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Mar 19 11:54:49.687738 containerd[1531]: time="2025-03-19T11:54:49.685439332Z" level=info msg="NRI interface is disabled by configuration." Mar 19 11:54:49.687975 containerd[1531]: time="2025-03-19T11:54:49.685449642Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Mar 19 11:54:49.687994 containerd[1531]: time="2025-03-19T11:54:49.685832520Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Mar 19 11:54:49.687994 containerd[1531]: time="2025-03-19T11:54:49.685885960Z" level=info msg="Connect containerd service" Mar 19 11:54:49.687994 containerd[1531]: time="2025-03-19T11:54:49.685909213Z" level=info msg="using legacy CRI server" Mar 19 11:54:49.687994 containerd[1531]: time="2025-03-19T11:54:49.685914363Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Mar 19 11:54:49.687994 containerd[1531]: time="2025-03-19T11:54:49.686017617Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Mar 19 11:54:49.687994 containerd[1531]: time="2025-03-19T11:54:49.686885815Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Mar 19 11:54:49.687994 containerd[1531]: time="2025-03-19T11:54:49.687009056Z" level=info msg="Start subscribing containerd event" Mar 19 11:54:49.687994 containerd[1531]: time="2025-03-19T11:54:49.687048160Z" level=info msg="Start recovering state" Mar 19 11:54:49.687994 containerd[1531]: time="2025-03-19T11:54:49.687096219Z" level=info msg="Start event monitor" Mar 19 11:54:49.687994 containerd[1531]: time="2025-03-19T11:54:49.687111528Z" level=info msg="Start snapshots syncer" Mar 19 11:54:49.687994 containerd[1531]: time="2025-03-19T11:54:49.687121056Z" level=info msg="Start cni network conf syncer for default" Mar 19 11:54:49.687994 containerd[1531]: time="2025-03-19T11:54:49.687127538Z" level=info msg="Start streaming server" Mar 19 11:54:49.687994 containerd[1531]: time="2025-03-19T11:54:49.687513762Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Mar 19 11:54:49.687994 containerd[1531]: time="2025-03-19T11:54:49.687550631Z" level=info msg=serving... address=/run/containerd/containerd.sock Mar 19 11:54:49.688980 systemd[1]: Started containerd.service - containerd container runtime. Mar 19 11:54:49.692866 containerd[1531]: time="2025-03-19T11:54:49.692489914Z" level=info msg="containerd successfully booted in 0.083676s" Mar 19 11:54:49.700646 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Mar 19 11:54:49.720735 systemd[1]: Started getty@tty1.service - Getty on tty1. Mar 19 11:54:49.730085 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Mar 19 11:54:49.732227 systemd[1]: Reached target getty.target - Login Prompts. Mar 19 11:54:49.837722 tar[1518]: linux-amd64/LICENSE Mar 19 11:54:49.837722 tar[1518]: linux-amd64/README.md Mar 19 11:54:49.849681 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Mar 19 11:54:50.366537 systemd-networkd[1415]: eth0: Gained IPv6LL Mar 19 11:54:50.367666 systemd-timesyncd[1416]: Network configuration changed, trying to establish connection. Mar 19 11:54:50.369831 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Mar 19 11:54:50.371990 systemd[1]: Reached target network-online.target - Network is Online. Mar 19 11:54:50.383594 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 19 11:54:50.386737 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Mar 19 11:54:50.411359 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Mar 19 11:54:50.432441 systemd-networkd[1415]: eth1: Gained IPv6LL Mar 19 11:54:50.434012 systemd-timesyncd[1416]: Network configuration changed, trying to establish connection. Mar 19 11:54:51.506971 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 19 11:54:51.509729 systemd[1]: Reached target multi-user.target - Multi-User System. Mar 19 11:54:51.513659 systemd[1]: Startup finished in 1.483s (kernel) + 9.339s (initrd) + 5.488s (userspace) = 16.311s. Mar 19 11:54:51.513883 (kubelet)[1619]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 19 11:54:52.469558 kubelet[1619]: E0319 11:54:52.469407 1619 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 19 11:54:52.471507 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 19 11:54:52.471774 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 19 11:54:52.472330 systemd[1]: kubelet.service: Consumed 1.397s CPU time, 237.1M memory peak. Mar 19 11:55:02.630485 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Mar 19 11:55:02.635798 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 19 11:55:02.767271 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 19 11:55:02.787712 (kubelet)[1638]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 19 11:55:02.833313 kubelet[1638]: E0319 11:55:02.831883 1638 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 19 11:55:02.836021 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 19 11:55:02.836183 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 19 11:55:02.836679 systemd[1]: kubelet.service: Consumed 154ms CPU time, 96M memory peak. Mar 19 11:55:12.880384 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Mar 19 11:55:12.885588 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 19 11:55:12.988651 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 19 11:55:13.001789 (kubelet)[1653]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 19 11:55:13.043849 kubelet[1653]: E0319 11:55:13.043773 1653 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 19 11:55:13.046582 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 19 11:55:13.046735 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 19 11:55:13.047154 systemd[1]: kubelet.service: Consumed 141ms CPU time, 95.8M memory peak. Mar 19 11:55:22.072098 systemd-timesyncd[1416]: Contacted time server 5.9.193.27:123 (2.flatcar.pool.ntp.org). Mar 19 11:55:22.072169 systemd-timesyncd[1416]: Initial clock synchronization to Wed 2025-03-19 11:55:22.071864 UTC. Mar 19 11:55:22.072234 systemd-resolved[1366]: Clock change detected. Flushing caches. Mar 19 11:55:24.316020 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Mar 19 11:55:24.321150 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 19 11:55:24.423783 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 19 11:55:24.432236 (kubelet)[1669]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 19 11:55:24.470904 kubelet[1669]: E0319 11:55:24.469795 1669 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 19 11:55:24.472126 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 19 11:55:24.472247 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 19 11:55:24.472503 systemd[1]: kubelet.service: Consumed 135ms CPU time, 97.5M memory peak. Mar 19 11:55:34.566109 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 4. Mar 19 11:55:34.573208 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 19 11:55:34.680477 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 19 11:55:34.685428 (kubelet)[1684]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 19 11:55:34.728987 kubelet[1684]: E0319 11:55:34.728907 1684 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 19 11:55:34.732513 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 19 11:55:34.732674 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 19 11:55:34.733007 systemd[1]: kubelet.service: Consumed 149ms CPU time, 97.7M memory peak. Mar 19 11:55:35.432940 update_engine[1506]: I20250319 11:55:35.432739 1506 update_attempter.cc:509] Updating boot flags... Mar 19 11:55:35.474956 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 36 scanned by (udev-worker) (1701) Mar 19 11:55:35.561209 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 36 scanned by (udev-worker) (1700) Mar 19 11:55:35.614921 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 36 scanned by (udev-worker) (1700) Mar 19 11:55:44.815940 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 5. Mar 19 11:55:44.821162 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 19 11:55:44.975565 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 19 11:55:44.992438 (kubelet)[1721]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 19 11:55:45.037806 kubelet[1721]: E0319 11:55:45.037717 1721 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 19 11:55:45.040355 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 19 11:55:45.040745 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 19 11:55:45.041113 systemd[1]: kubelet.service: Consumed 154ms CPU time, 93.8M memory peak. Mar 19 11:55:55.065899 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 6. Mar 19 11:55:55.071361 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 19 11:55:55.170176 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 19 11:55:55.183373 (kubelet)[1737]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 19 11:55:55.227844 kubelet[1737]: E0319 11:55:55.227776 1737 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 19 11:55:55.230613 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 19 11:55:55.230790 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 19 11:55:55.231120 systemd[1]: kubelet.service: Consumed 142ms CPU time, 97.5M memory peak. Mar 19 11:56:05.316384 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 7. Mar 19 11:56:05.323115 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 19 11:56:05.440230 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 19 11:56:05.444995 (kubelet)[1752]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 19 11:56:05.482373 kubelet[1752]: E0319 11:56:05.482243 1752 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 19 11:56:05.484668 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 19 11:56:05.484914 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 19 11:56:05.485175 systemd[1]: kubelet.service: Consumed 136ms CPU time, 97.6M memory peak. Mar 19 11:56:15.566089 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 8. Mar 19 11:56:15.571163 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 19 11:56:15.778543 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 19 11:56:15.783176 (kubelet)[1768]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 19 11:56:15.835733 kubelet[1768]: E0319 11:56:15.835560 1768 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 19 11:56:15.839473 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 19 11:56:15.839663 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 19 11:56:15.840036 systemd[1]: kubelet.service: Consumed 187ms CPU time, 97.5M memory peak. Mar 19 11:56:26.065997 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 9. Mar 19 11:56:26.073254 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 19 11:56:26.216951 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 19 11:56:26.221547 (kubelet)[1783]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 19 11:56:26.265006 kubelet[1783]: E0319 11:56:26.264838 1783 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 19 11:56:26.266807 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 19 11:56:26.266992 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 19 11:56:26.267304 systemd[1]: kubelet.service: Consumed 176ms CPU time, 97.4M memory peak. Mar 19 11:56:31.733264 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Mar 19 11:56:31.738212 systemd[1]: Started sshd@0-37.27.246.226:22-139.178.68.195:55060.service - OpenSSH per-connection server daemon (139.178.68.195:55060). Mar 19 11:56:32.751800 sshd[1791]: Accepted publickey for core from 139.178.68.195 port 55060 ssh2: RSA SHA256:mZbisEv4EcLuWBRjDimak4H4E9nhJ1zN2E80H6lgS54 Mar 19 11:56:32.753952 sshd-session[1791]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 19 11:56:32.762070 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Mar 19 11:56:32.767266 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Mar 19 11:56:32.777184 systemd-logind[1502]: New session 1 of user core. Mar 19 11:56:32.784239 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Mar 19 11:56:32.791434 systemd[1]: Starting user@500.service - User Manager for UID 500... Mar 19 11:56:32.796790 (systemd)[1795]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Mar 19 11:56:32.800185 systemd-logind[1502]: New session c1 of user core. Mar 19 11:56:33.001375 systemd[1795]: Queued start job for default target default.target. Mar 19 11:56:33.013219 systemd[1795]: Created slice app.slice - User Application Slice. Mar 19 11:56:33.013257 systemd[1795]: Reached target paths.target - Paths. Mar 19 11:56:33.013302 systemd[1795]: Reached target timers.target - Timers. Mar 19 11:56:33.015036 systemd[1795]: Starting dbus.socket - D-Bus User Message Bus Socket... Mar 19 11:56:33.030446 systemd[1795]: Listening on dbus.socket - D-Bus User Message Bus Socket. Mar 19 11:56:33.030589 systemd[1795]: Reached target sockets.target - Sockets. Mar 19 11:56:33.030642 systemd[1795]: Reached target basic.target - Basic System. Mar 19 11:56:33.030700 systemd[1795]: Reached target default.target - Main User Target. Mar 19 11:56:33.030731 systemd[1795]: Startup finished in 222ms. Mar 19 11:56:33.030971 systemd[1]: Started user@500.service - User Manager for UID 500. Mar 19 11:56:33.043185 systemd[1]: Started session-1.scope - Session 1 of User core. Mar 19 11:56:33.743345 systemd[1]: Started sshd@1-37.27.246.226:22-139.178.68.195:55066.service - OpenSSH per-connection server daemon (139.178.68.195:55066). Mar 19 11:56:34.722368 sshd[1807]: Accepted publickey for core from 139.178.68.195 port 55066 ssh2: RSA SHA256:mZbisEv4EcLuWBRjDimak4H4E9nhJ1zN2E80H6lgS54 Mar 19 11:56:34.724409 sshd-session[1807]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 19 11:56:34.731682 systemd-logind[1502]: New session 2 of user core. Mar 19 11:56:34.744280 systemd[1]: Started session-2.scope - Session 2 of User core. Mar 19 11:56:35.400244 sshd[1809]: Connection closed by 139.178.68.195 port 55066 Mar 19 11:56:35.401272 sshd-session[1807]: pam_unix(sshd:session): session closed for user core Mar 19 11:56:35.405553 systemd-logind[1502]: Session 2 logged out. Waiting for processes to exit. Mar 19 11:56:35.405704 systemd[1]: sshd@1-37.27.246.226:22-139.178.68.195:55066.service: Deactivated successfully. Mar 19 11:56:35.407740 systemd[1]: session-2.scope: Deactivated successfully. Mar 19 11:56:35.409334 systemd-logind[1502]: Removed session 2. Mar 19 11:56:35.570198 systemd[1]: Started sshd@2-37.27.246.226:22-139.178.68.195:55072.service - OpenSSH per-connection server daemon (139.178.68.195:55072). Mar 19 11:56:36.315936 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 10. Mar 19 11:56:36.321549 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 19 11:56:36.436207 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 19 11:56:36.437681 (kubelet)[1825]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 19 11:56:36.482701 kubelet[1825]: E0319 11:56:36.482626 1825 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 19 11:56:36.486022 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 19 11:56:36.486201 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 19 11:56:36.486710 systemd[1]: kubelet.service: Consumed 153ms CPU time, 97.5M memory peak. Mar 19 11:56:36.558857 sshd[1815]: Accepted publickey for core from 139.178.68.195 port 55072 ssh2: RSA SHA256:mZbisEv4EcLuWBRjDimak4H4E9nhJ1zN2E80H6lgS54 Mar 19 11:56:36.560535 sshd-session[1815]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 19 11:56:36.565747 systemd-logind[1502]: New session 3 of user core. Mar 19 11:56:36.575151 systemd[1]: Started session-3.scope - Session 3 of User core. Mar 19 11:56:37.230087 sshd[1832]: Connection closed by 139.178.68.195 port 55072 Mar 19 11:56:37.230747 sshd-session[1815]: pam_unix(sshd:session): session closed for user core Mar 19 11:56:37.233528 systemd[1]: sshd@2-37.27.246.226:22-139.178.68.195:55072.service: Deactivated successfully. Mar 19 11:56:37.235305 systemd[1]: session-3.scope: Deactivated successfully. Mar 19 11:56:37.236528 systemd-logind[1502]: Session 3 logged out. Waiting for processes to exit. Mar 19 11:56:37.237701 systemd-logind[1502]: Removed session 3. Mar 19 11:56:37.439548 systemd[1]: Started sshd@3-37.27.246.226:22-139.178.68.195:48466.service - OpenSSH per-connection server daemon (139.178.68.195:48466). Mar 19 11:56:38.508470 sshd[1838]: Accepted publickey for core from 139.178.68.195 port 48466 ssh2: RSA SHA256:mZbisEv4EcLuWBRjDimak4H4E9nhJ1zN2E80H6lgS54 Mar 19 11:56:38.509897 sshd-session[1838]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 19 11:56:38.514942 systemd-logind[1502]: New session 4 of user core. Mar 19 11:56:38.523151 systemd[1]: Started session-4.scope - Session 4 of User core. Mar 19 11:56:39.247041 sshd[1840]: Connection closed by 139.178.68.195 port 48466 Mar 19 11:56:39.247913 sshd-session[1838]: pam_unix(sshd:session): session closed for user core Mar 19 11:56:39.251197 systemd[1]: sshd@3-37.27.246.226:22-139.178.68.195:48466.service: Deactivated successfully. Mar 19 11:56:39.253125 systemd[1]: session-4.scope: Deactivated successfully. Mar 19 11:56:39.254493 systemd-logind[1502]: Session 4 logged out. Waiting for processes to exit. Mar 19 11:56:39.255732 systemd-logind[1502]: Removed session 4. Mar 19 11:56:39.404233 systemd[1]: Started sshd@4-37.27.246.226:22-139.178.68.195:48474.service - OpenSSH per-connection server daemon (139.178.68.195:48474). Mar 19 11:56:40.381394 sshd[1846]: Accepted publickey for core from 139.178.68.195 port 48474 ssh2: RSA SHA256:mZbisEv4EcLuWBRjDimak4H4E9nhJ1zN2E80H6lgS54 Mar 19 11:56:40.382835 sshd-session[1846]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 19 11:56:40.389898 systemd-logind[1502]: New session 5 of user core. Mar 19 11:56:40.402305 systemd[1]: Started session-5.scope - Session 5 of User core. Mar 19 11:56:40.909048 sudo[1849]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Mar 19 11:56:40.909321 sudo[1849]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 19 11:56:40.925570 sudo[1849]: pam_unix(sudo:session): session closed for user root Mar 19 11:56:41.083918 sshd[1848]: Connection closed by 139.178.68.195 port 48474 Mar 19 11:56:41.084891 sshd-session[1846]: pam_unix(sshd:session): session closed for user core Mar 19 11:56:41.089519 systemd[1]: sshd@4-37.27.246.226:22-139.178.68.195:48474.service: Deactivated successfully. Mar 19 11:56:41.092373 systemd[1]: session-5.scope: Deactivated successfully. Mar 19 11:56:41.093255 systemd-logind[1502]: Session 5 logged out. Waiting for processes to exit. Mar 19 11:56:41.094440 systemd-logind[1502]: Removed session 5. Mar 19 11:56:41.261370 systemd[1]: Started sshd@5-37.27.246.226:22-139.178.68.195:48480.service - OpenSSH per-connection server daemon (139.178.68.195:48480). Mar 19 11:56:42.243051 sshd[1855]: Accepted publickey for core from 139.178.68.195 port 48480 ssh2: RSA SHA256:mZbisEv4EcLuWBRjDimak4H4E9nhJ1zN2E80H6lgS54 Mar 19 11:56:42.244932 sshd-session[1855]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 19 11:56:42.252038 systemd-logind[1502]: New session 6 of user core. Mar 19 11:56:42.261199 systemd[1]: Started session-6.scope - Session 6 of User core. Mar 19 11:56:42.766037 sudo[1859]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Mar 19 11:56:42.766610 sudo[1859]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 19 11:56:42.771682 sudo[1859]: pam_unix(sudo:session): session closed for user root Mar 19 11:56:42.778462 sudo[1858]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Mar 19 11:56:42.778810 sudo[1858]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 19 11:56:42.794389 systemd[1]: Starting audit-rules.service - Load Audit Rules... Mar 19 11:56:42.827032 augenrules[1881]: No rules Mar 19 11:56:42.829392 systemd[1]: audit-rules.service: Deactivated successfully. Mar 19 11:56:42.829662 systemd[1]: Finished audit-rules.service - Load Audit Rules. Mar 19 11:56:42.831188 sudo[1858]: pam_unix(sudo:session): session closed for user root Mar 19 11:56:42.989384 sshd[1857]: Connection closed by 139.178.68.195 port 48480 Mar 19 11:56:42.989965 sshd-session[1855]: pam_unix(sshd:session): session closed for user core Mar 19 11:56:42.993739 systemd[1]: sshd@5-37.27.246.226:22-139.178.68.195:48480.service: Deactivated successfully. Mar 19 11:56:42.996505 systemd[1]: session-6.scope: Deactivated successfully. Mar 19 11:56:42.997817 systemd-logind[1502]: Session 6 logged out. Waiting for processes to exit. Mar 19 11:56:42.999565 systemd-logind[1502]: Removed session 6. Mar 19 11:56:43.174459 systemd[1]: Started sshd@6-37.27.246.226:22-139.178.68.195:48482.service - OpenSSH per-connection server daemon (139.178.68.195:48482). Mar 19 11:56:44.184789 sshd[1890]: Accepted publickey for core from 139.178.68.195 port 48482 ssh2: RSA SHA256:mZbisEv4EcLuWBRjDimak4H4E9nhJ1zN2E80H6lgS54 Mar 19 11:56:44.186566 sshd-session[1890]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 19 11:56:44.192109 systemd-logind[1502]: New session 7 of user core. Mar 19 11:56:44.202251 systemd[1]: Started session-7.scope - Session 7 of User core. Mar 19 11:56:44.717781 sudo[1893]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Mar 19 11:56:44.718344 sudo[1893]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 19 11:56:45.123386 systemd[1]: Starting docker.service - Docker Application Container Engine... Mar 19 11:56:45.125743 (dockerd)[1909]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Mar 19 11:56:45.498102 dockerd[1909]: time="2025-03-19T11:56:45.497857834Z" level=info msg="Starting up" Mar 19 11:56:45.636685 dockerd[1909]: time="2025-03-19T11:56:45.636449859Z" level=info msg="Loading containers: start." Mar 19 11:56:45.781904 kernel: Initializing XFRM netlink socket Mar 19 11:56:45.880044 systemd-networkd[1415]: docker0: Link UP Mar 19 11:56:45.919984 dockerd[1909]: time="2025-03-19T11:56:45.919904667Z" level=info msg="Loading containers: done." Mar 19 11:56:45.938616 dockerd[1909]: time="2025-03-19T11:56:45.938529785Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Mar 19 11:56:45.938850 dockerd[1909]: time="2025-03-19T11:56:45.938668065Z" level=info msg="Docker daemon" commit=41ca978a0a5400cc24b274137efa9f25517fcc0b containerd-snapshotter=false storage-driver=overlay2 version=27.3.1 Mar 19 11:56:45.938850 dockerd[1909]: time="2025-03-19T11:56:45.938827855Z" level=info msg="Daemon has completed initialization" Mar 19 11:56:45.980112 dockerd[1909]: time="2025-03-19T11:56:45.979577733Z" level=info msg="API listen on /run/docker.sock" Mar 19 11:56:45.980362 systemd[1]: Started docker.service - Docker Application Container Engine. Mar 19 11:56:46.566006 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 11. Mar 19 11:56:46.573167 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 19 11:56:46.659360 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 19 11:56:46.663822 (kubelet)[2106]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 19 11:56:46.721456 kubelet[2106]: E0319 11:56:46.721379 2106 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 19 11:56:46.724518 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 19 11:56:46.724729 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 19 11:56:46.725392 systemd[1]: kubelet.service: Consumed 150ms CPU time, 97.6M memory peak. Mar 19 11:56:47.195596 containerd[1531]: time="2025-03-19T11:56:47.195480219Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.7\"" Mar 19 11:56:47.717277 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3085760141.mount: Deactivated successfully. Mar 19 11:56:48.847545 containerd[1531]: time="2025-03-19T11:56:48.847473488Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.31.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 19 11:56:48.848769 containerd[1531]: time="2025-03-19T11:56:48.848695298Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.31.7: active requests=0, bytes read=27959362" Mar 19 11:56:48.850215 containerd[1531]: time="2025-03-19T11:56:48.850145007Z" level=info msg="ImageCreate event name:\"sha256:f084bc047a8cf7c8484d47c51e70e646dde3977d916f282feb99207b7b9241af\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 19 11:56:48.853481 containerd[1531]: time="2025-03-19T11:56:48.853403716Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:22c19cc70fe5806d0a2cb28a6b6b33fd34e6f9e50616bdf6d53649bcfafbc277\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 19 11:56:48.854955 containerd[1531]: time="2025-03-19T11:56:48.854755390Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.31.7\" with image id \"sha256:f084bc047a8cf7c8484d47c51e70e646dde3977d916f282feb99207b7b9241af\", repo tag \"registry.k8s.io/kube-apiserver:v1.31.7\", repo digest \"registry.k8s.io/kube-apiserver@sha256:22c19cc70fe5806d0a2cb28a6b6b33fd34e6f9e50616bdf6d53649bcfafbc277\", size \"27956068\" in 1.659203676s" Mar 19 11:56:48.854955 containerd[1531]: time="2025-03-19T11:56:48.854796357Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.7\" returns image reference \"sha256:f084bc047a8cf7c8484d47c51e70e646dde3977d916f282feb99207b7b9241af\"" Mar 19 11:56:48.857313 containerd[1531]: time="2025-03-19T11:56:48.857280706Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.7\"" Mar 19 11:56:50.055445 containerd[1531]: time="2025-03-19T11:56:50.055379390Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.31.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 19 11:56:50.056482 containerd[1531]: time="2025-03-19T11:56:50.056427144Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.31.7: active requests=0, bytes read=24713798" Mar 19 11:56:50.057531 containerd[1531]: time="2025-03-19T11:56:50.057480890Z" level=info msg="ImageCreate event name:\"sha256:652dcad615a9a0c252c253860d5b5b7bfebd3efe159dc033a8555bc15a6d1985\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 19 11:56:50.060258 containerd[1531]: time="2025-03-19T11:56:50.060184279Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:6abe7a0accecf29db6ebab18a10f844678ffed693d79e2e51a18a6f2b4530cbb\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 19 11:56:50.061162 containerd[1531]: time="2025-03-19T11:56:50.061027089Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.31.7\" with image id \"sha256:652dcad615a9a0c252c253860d5b5b7bfebd3efe159dc033a8555bc15a6d1985\", repo tag \"registry.k8s.io/kube-controller-manager:v1.31.7\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:6abe7a0accecf29db6ebab18a10f844678ffed693d79e2e51a18a6f2b4530cbb\", size \"26201384\" in 1.203577106s" Mar 19 11:56:50.061162 containerd[1531]: time="2025-03-19T11:56:50.061069629Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.7\" returns image reference \"sha256:652dcad615a9a0c252c253860d5b5b7bfebd3efe159dc033a8555bc15a6d1985\"" Mar 19 11:56:50.061776 containerd[1531]: time="2025-03-19T11:56:50.061740607Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.7\"" Mar 19 11:56:51.107491 containerd[1531]: time="2025-03-19T11:56:51.107426665Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.31.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 19 11:56:51.108638 containerd[1531]: time="2025-03-19T11:56:51.108600005Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.31.7: active requests=0, bytes read=18780390" Mar 19 11:56:51.109772 containerd[1531]: time="2025-03-19T11:56:51.109734742Z" level=info msg="ImageCreate event name:\"sha256:7f1f6a63d8aa14cf61d0045e912ad312b4ade24637cecccc933b163582eae68c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 19 11:56:51.112669 containerd[1531]: time="2025-03-19T11:56:51.112615594Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:fb80249bcb77ee72b1c9fa5b70bc28a83ed107c9ca71957841ad91db379963bf\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 19 11:56:51.113662 containerd[1531]: time="2025-03-19T11:56:51.113525640Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.31.7\" with image id \"sha256:7f1f6a63d8aa14cf61d0045e912ad312b4ade24637cecccc933b163582eae68c\", repo tag \"registry.k8s.io/kube-scheduler:v1.31.7\", repo digest \"registry.k8s.io/kube-scheduler@sha256:fb80249bcb77ee72b1c9fa5b70bc28a83ed107c9ca71957841ad91db379963bf\", size \"20267994\" in 1.051749657s" Mar 19 11:56:51.113662 containerd[1531]: time="2025-03-19T11:56:51.113551578Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.7\" returns image reference \"sha256:7f1f6a63d8aa14cf61d0045e912ad312b4ade24637cecccc933b163582eae68c\"" Mar 19 11:56:51.114183 containerd[1531]: time="2025-03-19T11:56:51.114043631Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.7\"" Mar 19 11:56:52.191934 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount78656961.mount: Deactivated successfully. Mar 19 11:56:52.572298 containerd[1531]: time="2025-03-19T11:56:52.572207165Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.31.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 19 11:56:52.573189 containerd[1531]: time="2025-03-19T11:56:52.573135637Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.31.7: active requests=0, bytes read=30354658" Mar 19 11:56:52.574211 containerd[1531]: time="2025-03-19T11:56:52.574139720Z" level=info msg="ImageCreate event name:\"sha256:dcfc039c372ea285997a302d60e58a75b80905b4c4dba969993b9b22e8ac66d1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 19 11:56:52.576609 containerd[1531]: time="2025-03-19T11:56:52.576535220Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:e5839270c96c3ad1bea1dce4935126d3281297527f3655408d2970aa4b5cf178\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 19 11:56:52.577336 containerd[1531]: time="2025-03-19T11:56:52.577139754Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.31.7\" with image id \"sha256:dcfc039c372ea285997a302d60e58a75b80905b4c4dba969993b9b22e8ac66d1\", repo tag \"registry.k8s.io/kube-proxy:v1.31.7\", repo digest \"registry.k8s.io/kube-proxy@sha256:e5839270c96c3ad1bea1dce4935126d3281297527f3655408d2970aa4b5cf178\", size \"30353649\" in 1.462883234s" Mar 19 11:56:52.577336 containerd[1531]: time="2025-03-19T11:56:52.577179829Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.7\" returns image reference \"sha256:dcfc039c372ea285997a302d60e58a75b80905b4c4dba969993b9b22e8ac66d1\"" Mar 19 11:56:52.577644 containerd[1531]: time="2025-03-19T11:56:52.577619865Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\"" Mar 19 11:56:53.120045 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount431828506.mount: Deactivated successfully. Mar 19 11:56:53.933193 containerd[1531]: time="2025-03-19T11:56:53.933134464Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 19 11:56:53.934281 containerd[1531]: time="2025-03-19T11:56:53.934233315Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.1: active requests=0, bytes read=18185843" Mar 19 11:56:53.935400 containerd[1531]: time="2025-03-19T11:56:53.935355729Z" level=info msg="ImageCreate event name:\"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 19 11:56:53.938165 containerd[1531]: time="2025-03-19T11:56:53.938105215Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 19 11:56:53.939150 containerd[1531]: time="2025-03-19T11:56:53.939029267Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.1\" with image id \"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\", size \"18182961\" in 1.361382442s" Mar 19 11:56:53.939150 containerd[1531]: time="2025-03-19T11:56:53.939058772Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\" returns image reference \"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\"" Mar 19 11:56:53.939584 containerd[1531]: time="2025-03-19T11:56:53.939571103Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Mar 19 11:56:54.412668 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1132973286.mount: Deactivated successfully. Mar 19 11:56:54.418987 containerd[1531]: time="2025-03-19T11:56:54.418900761Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 19 11:56:54.419919 containerd[1531]: time="2025-03-19T11:56:54.419836374Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321160" Mar 19 11:56:54.420901 containerd[1531]: time="2025-03-19T11:56:54.420834155Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 19 11:56:54.423753 containerd[1531]: time="2025-03-19T11:56:54.423687596Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 19 11:56:54.424892 containerd[1531]: time="2025-03-19T11:56:54.424691918Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 485.051846ms" Mar 19 11:56:54.424892 containerd[1531]: time="2025-03-19T11:56:54.424753403Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Mar 19 11:56:54.425413 containerd[1531]: time="2025-03-19T11:56:54.425298556Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\"" Mar 19 11:56:54.957426 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2073264350.mount: Deactivated successfully. Mar 19 11:56:56.445427 containerd[1531]: time="2025-03-19T11:56:56.445336260Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.15-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 19 11:56:56.447214 containerd[1531]: time="2025-03-19T11:56:56.447135703Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.15-0: active requests=0, bytes read=56780037" Mar 19 11:56:56.448915 containerd[1531]: time="2025-03-19T11:56:56.448837584Z" level=info msg="ImageCreate event name:\"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 19 11:56:56.452951 containerd[1531]: time="2025-03-19T11:56:56.452895112Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 19 11:56:56.454627 containerd[1531]: time="2025-03-19T11:56:56.454520740Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.15-0\" with image id \"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\", repo tag \"registry.k8s.io/etcd:3.5.15-0\", repo digest \"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\", size \"56909194\" in 2.029195685s" Mar 19 11:56:56.454627 containerd[1531]: time="2025-03-19T11:56:56.454565084Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\" returns image reference \"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\"" Mar 19 11:56:56.815842 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 12. Mar 19 11:56:56.821217 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 19 11:56:56.992703 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 19 11:56:56.996486 (kubelet)[2301]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 19 11:56:57.075925 kubelet[2301]: E0319 11:56:57.074467 2301 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 19 11:56:57.076570 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 19 11:56:57.076726 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 19 11:56:57.077369 systemd[1]: kubelet.service: Consumed 181ms CPU time, 96.8M memory peak. Mar 19 11:57:00.364238 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 19 11:57:00.364938 systemd[1]: kubelet.service: Consumed 181ms CPU time, 96.8M memory peak. Mar 19 11:57:00.374351 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 19 11:57:00.417352 systemd[1]: Reload requested from client PID 2328 ('systemctl') (unit session-7.scope)... Mar 19 11:57:00.417539 systemd[1]: Reloading... Mar 19 11:57:00.566909 zram_generator::config[2376]: No configuration found. Mar 19 11:57:00.704819 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 19 11:57:00.847494 systemd[1]: Reloading finished in 429 ms. Mar 19 11:57:00.889031 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 19 11:57:00.893353 (kubelet)[2417]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Mar 19 11:57:00.897208 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Mar 19 11:57:00.897947 systemd[1]: kubelet.service: Deactivated successfully. Mar 19 11:57:00.898170 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 19 11:57:00.898219 systemd[1]: kubelet.service: Consumed 98ms CPU time, 85M memory peak. Mar 19 11:57:00.906446 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 19 11:57:01.012159 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 19 11:57:01.020242 (kubelet)[2434]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Mar 19 11:57:01.062271 kubelet[2434]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 19 11:57:01.062893 kubelet[2434]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Mar 19 11:57:01.062893 kubelet[2434]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 19 11:57:01.064939 kubelet[2434]: I0319 11:57:01.064530 2434 server.go:206] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 19 11:57:01.364896 kubelet[2434]: I0319 11:57:01.363765 2434 server.go:486] "Kubelet version" kubeletVersion="v1.31.0" Mar 19 11:57:01.364896 kubelet[2434]: I0319 11:57:01.363813 2434 server.go:488] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 19 11:57:01.364896 kubelet[2434]: I0319 11:57:01.364308 2434 server.go:929] "Client rotation is on, will bootstrap in background" Mar 19 11:57:01.407321 kubelet[2434]: E0319 11:57:01.407234 2434 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://37.27.246.226:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 37.27.246.226:6443: connect: connection refused" logger="UnhandledError" Mar 19 11:57:01.407711 kubelet[2434]: I0319 11:57:01.407515 2434 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Mar 19 11:57:01.418380 kubelet[2434]: E0319 11:57:01.418302 2434 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Mar 19 11:57:01.418380 kubelet[2434]: I0319 11:57:01.418355 2434 server.go:1403] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Mar 19 11:57:01.425671 kubelet[2434]: I0319 11:57:01.425605 2434 server.go:744] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Mar 19 11:57:01.429781 kubelet[2434]: I0319 11:57:01.429692 2434 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Mar 19 11:57:01.430185 kubelet[2434]: I0319 11:57:01.430130 2434 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 19 11:57:01.430445 kubelet[2434]: I0319 11:57:01.430184 2434 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4230-1-0-d-e284590f49","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 19 11:57:01.430445 kubelet[2434]: I0319 11:57:01.430443 2434 topology_manager.go:138] "Creating topology manager with none policy" Mar 19 11:57:01.430662 kubelet[2434]: I0319 11:57:01.430458 2434 container_manager_linux.go:300] "Creating device plugin manager" Mar 19 11:57:01.430662 kubelet[2434]: I0319 11:57:01.430609 2434 state_mem.go:36] "Initialized new in-memory state store" Mar 19 11:57:01.433610 kubelet[2434]: I0319 11:57:01.433276 2434 kubelet.go:408] "Attempting to sync node with API server" Mar 19 11:57:01.433610 kubelet[2434]: I0319 11:57:01.433319 2434 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 19 11:57:01.433610 kubelet[2434]: I0319 11:57:01.433370 2434 kubelet.go:314] "Adding apiserver pod source" Mar 19 11:57:01.433610 kubelet[2434]: I0319 11:57:01.433390 2434 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 19 11:57:01.447984 kubelet[2434]: W0319 11:57:01.447511 2434 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://37.27.246.226:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4230-1-0-d-e284590f49&limit=500&resourceVersion=0": dial tcp 37.27.246.226:6443: connect: connection refused Mar 19 11:57:01.447984 kubelet[2434]: E0319 11:57:01.447620 2434 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://37.27.246.226:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4230-1-0-d-e284590f49&limit=500&resourceVersion=0\": dial tcp 37.27.246.226:6443: connect: connection refused" logger="UnhandledError" Mar 19 11:57:01.449012 kubelet[2434]: I0319 11:57:01.448856 2434 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v1.7.23" apiVersion="v1" Mar 19 11:57:01.452132 kubelet[2434]: W0319 11:57:01.452044 2434 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://37.27.246.226:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 37.27.246.226:6443: connect: connection refused Mar 19 11:57:01.452999 kubelet[2434]: E0319 11:57:01.452144 2434 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://37.27.246.226:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 37.27.246.226:6443: connect: connection refused" logger="UnhandledError" Mar 19 11:57:01.454023 kubelet[2434]: I0319 11:57:01.453976 2434 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Mar 19 11:57:01.454841 kubelet[2434]: W0319 11:57:01.454793 2434 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Mar 19 11:57:01.455586 kubelet[2434]: I0319 11:57:01.455542 2434 server.go:1269] "Started kubelet" Mar 19 11:57:01.458100 kubelet[2434]: I0319 11:57:01.457126 2434 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Mar 19 11:57:01.460961 kubelet[2434]: I0319 11:57:01.460186 2434 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 19 11:57:01.465664 kubelet[2434]: I0319 11:57:01.465138 2434 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 19 11:57:01.465664 kubelet[2434]: I0319 11:57:01.465488 2434 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 19 11:57:01.469684 kubelet[2434]: I0319 11:57:01.469648 2434 server.go:460] "Adding debug handlers to kubelet server" Mar 19 11:57:01.474602 kubelet[2434]: I0319 11:57:01.473707 2434 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Mar 19 11:57:01.474602 kubelet[2434]: I0319 11:57:01.474060 2434 volume_manager.go:289] "Starting Kubelet Volume Manager" Mar 19 11:57:01.474602 kubelet[2434]: E0319 11:57:01.474282 2434 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4230-1-0-d-e284590f49\" not found" Mar 19 11:57:01.476378 kubelet[2434]: E0319 11:57:01.472233 2434 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://37.27.246.226:6443/api/v1/namespaces/default/events\": dial tcp 37.27.246.226:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4230-1-0-d-e284590f49.182e324f2bacdfa1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4230-1-0-d-e284590f49,UID:ci-4230-1-0-d-e284590f49,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4230-1-0-d-e284590f49,},FirstTimestamp:2025-03-19 11:57:01.455515553 +0000 UTC m=+0.431831299,LastTimestamp:2025-03-19 11:57:01.455515553 +0000 UTC m=+0.431831299,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4230-1-0-d-e284590f49,}" Mar 19 11:57:01.478897 kubelet[2434]: I0319 11:57:01.478852 2434 factory.go:221] Registration of the systemd container factory successfully Mar 19 11:57:01.480101 kubelet[2434]: I0319 11:57:01.480063 2434 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Mar 19 11:57:01.480336 kubelet[2434]: E0319 11:57:01.480308 2434 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://37.27.246.226:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4230-1-0-d-e284590f49?timeout=10s\": dial tcp 37.27.246.226:6443: connect: connection refused" interval="200ms" Mar 19 11:57:01.482734 kubelet[2434]: I0319 11:57:01.481262 2434 reconciler.go:26] "Reconciler: start to sync state" Mar 19 11:57:01.482734 kubelet[2434]: I0319 11:57:01.481353 2434 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 19 11:57:01.482734 kubelet[2434]: I0319 11:57:01.482032 2434 factory.go:221] Registration of the containerd container factory successfully Mar 19 11:57:01.492705 kubelet[2434]: I0319 11:57:01.492658 2434 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Mar 19 11:57:01.493481 kubelet[2434]: I0319 11:57:01.493457 2434 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Mar 19 11:57:01.493541 kubelet[2434]: I0319 11:57:01.493490 2434 status_manager.go:217] "Starting to sync pod status with apiserver" Mar 19 11:57:01.493541 kubelet[2434]: I0319 11:57:01.493506 2434 kubelet.go:2321] "Starting kubelet main sync loop" Mar 19 11:57:01.493597 kubelet[2434]: E0319 11:57:01.493539 2434 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 19 11:57:01.500946 kubelet[2434]: W0319 11:57:01.500857 2434 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://37.27.246.226:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 37.27.246.226:6443: connect: connection refused Mar 19 11:57:01.501155 kubelet[2434]: E0319 11:57:01.501131 2434 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://37.27.246.226:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 37.27.246.226:6443: connect: connection refused" logger="UnhandledError" Mar 19 11:57:01.501786 kubelet[2434]: W0319 11:57:01.501731 2434 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://37.27.246.226:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 37.27.246.226:6443: connect: connection refused Mar 19 11:57:01.501928 kubelet[2434]: E0319 11:57:01.501860 2434 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://37.27.246.226:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 37.27.246.226:6443: connect: connection refused" logger="UnhandledError" Mar 19 11:57:01.502123 kubelet[2434]: E0319 11:57:01.502106 2434 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Mar 19 11:57:01.511475 kubelet[2434]: I0319 11:57:01.511406 2434 cpu_manager.go:214] "Starting CPU manager" policy="none" Mar 19 11:57:01.511475 kubelet[2434]: I0319 11:57:01.511455 2434 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Mar 19 11:57:01.511475 kubelet[2434]: I0319 11:57:01.511474 2434 state_mem.go:36] "Initialized new in-memory state store" Mar 19 11:57:01.512920 kubelet[2434]: I0319 11:57:01.512898 2434 policy_none.go:49] "None policy: Start" Mar 19 11:57:01.513393 kubelet[2434]: I0319 11:57:01.513367 2434 memory_manager.go:170] "Starting memorymanager" policy="None" Mar 19 11:57:01.513393 kubelet[2434]: I0319 11:57:01.513390 2434 state_mem.go:35] "Initializing new in-memory state store" Mar 19 11:57:01.519063 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Mar 19 11:57:01.530249 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Mar 19 11:57:01.533501 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Mar 19 11:57:01.541185 kubelet[2434]: I0319 11:57:01.540818 2434 manager.go:510] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Mar 19 11:57:01.541302 kubelet[2434]: I0319 11:57:01.541237 2434 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 19 11:57:01.541302 kubelet[2434]: I0319 11:57:01.541248 2434 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 19 11:57:01.541513 kubelet[2434]: I0319 11:57:01.541488 2434 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 19 11:57:01.544189 kubelet[2434]: E0319 11:57:01.544151 2434 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4230-1-0-d-e284590f49\" not found" Mar 19 11:57:01.604170 systemd[1]: Created slice kubepods-burstable-pod0c2706b2816e41ac0a665b5d91275921.slice - libcontainer container kubepods-burstable-pod0c2706b2816e41ac0a665b5d91275921.slice. Mar 19 11:57:01.616683 systemd[1]: Created slice kubepods-burstable-podad332766918da75e9877490f1a9e7817.slice - libcontainer container kubepods-burstable-podad332766918da75e9877490f1a9e7817.slice. Mar 19 11:57:01.624658 systemd[1]: Created slice kubepods-burstable-pod8863b05d3ad38ff6b9286f44b660a4c1.slice - libcontainer container kubepods-burstable-pod8863b05d3ad38ff6b9286f44b660a4c1.slice. Mar 19 11:57:01.643844 kubelet[2434]: I0319 11:57:01.643762 2434 kubelet_node_status.go:72] "Attempting to register node" node="ci-4230-1-0-d-e284590f49" Mar 19 11:57:01.644156 kubelet[2434]: E0319 11:57:01.644121 2434 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://37.27.246.226:6443/api/v1/nodes\": dial tcp 37.27.246.226:6443: connect: connection refused" node="ci-4230-1-0-d-e284590f49" Mar 19 11:57:01.680843 kubelet[2434]: E0319 11:57:01.680777 2434 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://37.27.246.226:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4230-1-0-d-e284590f49?timeout=10s\": dial tcp 37.27.246.226:6443: connect: connection refused" interval="400ms" Mar 19 11:57:01.682234 kubelet[2434]: I0319 11:57:01.682204 2434 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/ad332766918da75e9877490f1a9e7817-flexvolume-dir\") pod \"kube-controller-manager-ci-4230-1-0-d-e284590f49\" (UID: \"ad332766918da75e9877490f1a9e7817\") " pod="kube-system/kube-controller-manager-ci-4230-1-0-d-e284590f49" Mar 19 11:57:01.682234 kubelet[2434]: I0319 11:57:01.682239 2434 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/ad332766918da75e9877490f1a9e7817-k8s-certs\") pod \"kube-controller-manager-ci-4230-1-0-d-e284590f49\" (UID: \"ad332766918da75e9877490f1a9e7817\") " pod="kube-system/kube-controller-manager-ci-4230-1-0-d-e284590f49" Mar 19 11:57:01.682461 kubelet[2434]: I0319 11:57:01.682268 2434 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/ad332766918da75e9877490f1a9e7817-kubeconfig\") pod \"kube-controller-manager-ci-4230-1-0-d-e284590f49\" (UID: \"ad332766918da75e9877490f1a9e7817\") " pod="kube-system/kube-controller-manager-ci-4230-1-0-d-e284590f49" Mar 19 11:57:01.682461 kubelet[2434]: I0319 11:57:01.682287 2434 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/ad332766918da75e9877490f1a9e7817-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4230-1-0-d-e284590f49\" (UID: \"ad332766918da75e9877490f1a9e7817\") " pod="kube-system/kube-controller-manager-ci-4230-1-0-d-e284590f49" Mar 19 11:57:01.682461 kubelet[2434]: I0319 11:57:01.682306 2434 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/0c2706b2816e41ac0a665b5d91275921-ca-certs\") pod \"kube-apiserver-ci-4230-1-0-d-e284590f49\" (UID: \"0c2706b2816e41ac0a665b5d91275921\") " pod="kube-system/kube-apiserver-ci-4230-1-0-d-e284590f49" Mar 19 11:57:01.682461 kubelet[2434]: I0319 11:57:01.682323 2434 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/0c2706b2816e41ac0a665b5d91275921-k8s-certs\") pod \"kube-apiserver-ci-4230-1-0-d-e284590f49\" (UID: \"0c2706b2816e41ac0a665b5d91275921\") " pod="kube-system/kube-apiserver-ci-4230-1-0-d-e284590f49" Mar 19 11:57:01.682461 kubelet[2434]: I0319 11:57:01.682340 2434 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/0c2706b2816e41ac0a665b5d91275921-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4230-1-0-d-e284590f49\" (UID: \"0c2706b2816e41ac0a665b5d91275921\") " pod="kube-system/kube-apiserver-ci-4230-1-0-d-e284590f49" Mar 19 11:57:01.682679 kubelet[2434]: I0319 11:57:01.682362 2434 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/ad332766918da75e9877490f1a9e7817-ca-certs\") pod \"kube-controller-manager-ci-4230-1-0-d-e284590f49\" (UID: \"ad332766918da75e9877490f1a9e7817\") " pod="kube-system/kube-controller-manager-ci-4230-1-0-d-e284590f49" Mar 19 11:57:01.682679 kubelet[2434]: I0319 11:57:01.682389 2434 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/8863b05d3ad38ff6b9286f44b660a4c1-kubeconfig\") pod \"kube-scheduler-ci-4230-1-0-d-e284590f49\" (UID: \"8863b05d3ad38ff6b9286f44b660a4c1\") " pod="kube-system/kube-scheduler-ci-4230-1-0-d-e284590f49" Mar 19 11:57:01.847721 kubelet[2434]: I0319 11:57:01.847668 2434 kubelet_node_status.go:72] "Attempting to register node" node="ci-4230-1-0-d-e284590f49" Mar 19 11:57:01.848307 kubelet[2434]: E0319 11:57:01.848255 2434 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://37.27.246.226:6443/api/v1/nodes\": dial tcp 37.27.246.226:6443: connect: connection refused" node="ci-4230-1-0-d-e284590f49" Mar 19 11:57:01.913809 containerd[1531]: time="2025-03-19T11:57:01.913613052Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4230-1-0-d-e284590f49,Uid:0c2706b2816e41ac0a665b5d91275921,Namespace:kube-system,Attempt:0,}" Mar 19 11:57:01.925912 containerd[1531]: time="2025-03-19T11:57:01.925830128Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4230-1-0-d-e284590f49,Uid:ad332766918da75e9877490f1a9e7817,Namespace:kube-system,Attempt:0,}" Mar 19 11:57:01.929742 containerd[1531]: time="2025-03-19T11:57:01.929594726Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4230-1-0-d-e284590f49,Uid:8863b05d3ad38ff6b9286f44b660a4c1,Namespace:kube-system,Attempt:0,}" Mar 19 11:57:02.082039 kubelet[2434]: E0319 11:57:02.081940 2434 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://37.27.246.226:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4230-1-0-d-e284590f49?timeout=10s\": dial tcp 37.27.246.226:6443: connect: connection refused" interval="800ms" Mar 19 11:57:02.252412 kubelet[2434]: I0319 11:57:02.252091 2434 kubelet_node_status.go:72] "Attempting to register node" node="ci-4230-1-0-d-e284590f49" Mar 19 11:57:02.252855 kubelet[2434]: E0319 11:57:02.252789 2434 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://37.27.246.226:6443/api/v1/nodes\": dial tcp 37.27.246.226:6443: connect: connection refused" node="ci-4230-1-0-d-e284590f49" Mar 19 11:57:02.422385 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3251602807.mount: Deactivated successfully. Mar 19 11:57:02.427556 containerd[1531]: time="2025-03-19T11:57:02.427490047Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 19 11:57:02.431310 containerd[1531]: time="2025-03-19T11:57:02.431219720Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312078" Mar 19 11:57:02.433949 containerd[1531]: time="2025-03-19T11:57:02.433832816Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 19 11:57:02.436158 containerd[1531]: time="2025-03-19T11:57:02.436076490Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 19 11:57:02.437616 containerd[1531]: time="2025-03-19T11:57:02.437560364Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Mar 19 11:57:02.438659 containerd[1531]: time="2025-03-19T11:57:02.438632538Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 19 11:57:02.439526 kubelet[2434]: W0319 11:57:02.439446 2434 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://37.27.246.226:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 37.27.246.226:6443: connect: connection refused Mar 19 11:57:02.439823 kubelet[2434]: E0319 11:57:02.439742 2434 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://37.27.246.226:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 37.27.246.226:6443: connect: connection refused" logger="UnhandledError" Mar 19 11:57:02.440556 containerd[1531]: time="2025-03-19T11:57:02.440497484Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Mar 19 11:57:02.441302 containerd[1531]: time="2025-03-19T11:57:02.441259709Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 19 11:57:02.443295 containerd[1531]: time="2025-03-19T11:57:02.442829514Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 527.548645ms" Mar 19 11:57:02.445702 containerd[1531]: time="2025-03-19T11:57:02.445424706Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 519.410723ms" Mar 19 11:57:02.454441 containerd[1531]: time="2025-03-19T11:57:02.454279048Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 524.583152ms" Mar 19 11:57:02.530510 kubelet[2434]: W0319 11:57:02.530297 2434 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://37.27.246.226:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 37.27.246.226:6443: connect: connection refused Mar 19 11:57:02.530510 kubelet[2434]: E0319 11:57:02.530368 2434 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://37.27.246.226:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 37.27.246.226:6443: connect: connection refused" logger="UnhandledError" Mar 19 11:57:02.609211 containerd[1531]: time="2025-03-19T11:57:02.609097114Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 19 11:57:02.609593 containerd[1531]: time="2025-03-19T11:57:02.609427511Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 19 11:57:02.609860 containerd[1531]: time="2025-03-19T11:57:02.609511719Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 19 11:57:02.610302 containerd[1531]: time="2025-03-19T11:57:02.610177945Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 19 11:57:02.610618 containerd[1531]: time="2025-03-19T11:57:02.606544091Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 19 11:57:02.610618 containerd[1531]: time="2025-03-19T11:57:02.610389449Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 19 11:57:02.610618 containerd[1531]: time="2025-03-19T11:57:02.610407623Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 19 11:57:02.610618 containerd[1531]: time="2025-03-19T11:57:02.610478767Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 19 11:57:02.615404 containerd[1531]: time="2025-03-19T11:57:02.615311942Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 19 11:57:02.615404 containerd[1531]: time="2025-03-19T11:57:02.615367707Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 19 11:57:02.615404 containerd[1531]: time="2025-03-19T11:57:02.615385981Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 19 11:57:02.615996 containerd[1531]: time="2025-03-19T11:57:02.615464107Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 19 11:57:02.643076 systemd[1]: Started cri-containerd-10a9e8444d39b2481e25c41a1fb3fd0e41a47b4b6e056dc4d826c7f907ccb429.scope - libcontainer container 10a9e8444d39b2481e25c41a1fb3fd0e41a47b4b6e056dc4d826c7f907ccb429. Mar 19 11:57:02.644283 systemd[1]: Started cri-containerd-9463a572447583072e0149cece1f95fead2bb9245fb79e581359238a0a0ac2be.scope - libcontainer container 9463a572447583072e0149cece1f95fead2bb9245fb79e581359238a0a0ac2be. Mar 19 11:57:02.650183 systemd[1]: Started cri-containerd-80cd9f67a2d0eac0b86838b5ee8d1be8c588b648e121a1b5142d246175a5396d.scope - libcontainer container 80cd9f67a2d0eac0b86838b5ee8d1be8c588b648e121a1b5142d246175a5396d. Mar 19 11:57:02.710219 containerd[1531]: time="2025-03-19T11:57:02.709463955Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4230-1-0-d-e284590f49,Uid:8863b05d3ad38ff6b9286f44b660a4c1,Namespace:kube-system,Attempt:0,} returns sandbox id \"9463a572447583072e0149cece1f95fead2bb9245fb79e581359238a0a0ac2be\"" Mar 19 11:57:02.719623 containerd[1531]: time="2025-03-19T11:57:02.719561873Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4230-1-0-d-e284590f49,Uid:0c2706b2816e41ac0a665b5d91275921,Namespace:kube-system,Attempt:0,} returns sandbox id \"80cd9f67a2d0eac0b86838b5ee8d1be8c588b648e121a1b5142d246175a5396d\"" Mar 19 11:57:02.724915 containerd[1531]: time="2025-03-19T11:57:02.724847825Z" level=info msg="CreateContainer within sandbox \"9463a572447583072e0149cece1f95fead2bb9245fb79e581359238a0a0ac2be\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Mar 19 11:57:02.726993 containerd[1531]: time="2025-03-19T11:57:02.726601473Z" level=info msg="CreateContainer within sandbox \"80cd9f67a2d0eac0b86838b5ee8d1be8c588b648e121a1b5142d246175a5396d\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Mar 19 11:57:02.727180 containerd[1531]: time="2025-03-19T11:57:02.727105556Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4230-1-0-d-e284590f49,Uid:ad332766918da75e9877490f1a9e7817,Namespace:kube-system,Attempt:0,} returns sandbox id \"10a9e8444d39b2481e25c41a1fb3fd0e41a47b4b6e056dc4d826c7f907ccb429\"" Mar 19 11:57:02.730322 containerd[1531]: time="2025-03-19T11:57:02.730186155Z" level=info msg="CreateContainer within sandbox \"10a9e8444d39b2481e25c41a1fb3fd0e41a47b4b6e056dc4d826c7f907ccb429\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Mar 19 11:57:02.742519 containerd[1531]: time="2025-03-19T11:57:02.742441505Z" level=info msg="CreateContainer within sandbox \"9463a572447583072e0149cece1f95fead2bb9245fb79e581359238a0a0ac2be\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"c49290af89039639c3abe160eed3e98e253f51662a295c8d10c71a212cbc9f6d\"" Mar 19 11:57:02.743293 containerd[1531]: time="2025-03-19T11:57:02.743172381Z" level=info msg="StartContainer for \"c49290af89039639c3abe160eed3e98e253f51662a295c8d10c71a212cbc9f6d\"" Mar 19 11:57:02.748248 containerd[1531]: time="2025-03-19T11:57:02.748104893Z" level=info msg="CreateContainer within sandbox \"80cd9f67a2d0eac0b86838b5ee8d1be8c588b648e121a1b5142d246175a5396d\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"122877ffcbd219d054c6a57067aad884b83ba48abffec4d113c516ed0fa28c61\"" Mar 19 11:57:02.748951 containerd[1531]: time="2025-03-19T11:57:02.748740882Z" level=info msg="StartContainer for \"122877ffcbd219d054c6a57067aad884b83ba48abffec4d113c516ed0fa28c61\"" Mar 19 11:57:02.753888 containerd[1531]: time="2025-03-19T11:57:02.753808796Z" level=info msg="CreateContainer within sandbox \"10a9e8444d39b2481e25c41a1fb3fd0e41a47b4b6e056dc4d826c7f907ccb429\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"9f10d9e4adbc8baea36d71180b83e9c5adc60a9f1123335a1e2543df3606fa2b\"" Mar 19 11:57:02.754929 containerd[1531]: time="2025-03-19T11:57:02.754815398Z" level=info msg="StartContainer for \"9f10d9e4adbc8baea36d71180b83e9c5adc60a9f1123335a1e2543df3606fa2b\"" Mar 19 11:57:02.767615 systemd[1]: Started cri-containerd-c49290af89039639c3abe160eed3e98e253f51662a295c8d10c71a212cbc9f6d.scope - libcontainer container c49290af89039639c3abe160eed3e98e253f51662a295c8d10c71a212cbc9f6d. Mar 19 11:57:02.774648 kubelet[2434]: W0319 11:57:02.774549 2434 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://37.27.246.226:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4230-1-0-d-e284590f49&limit=500&resourceVersion=0": dial tcp 37.27.246.226:6443: connect: connection refused Mar 19 11:57:02.774829 kubelet[2434]: E0319 11:57:02.774651 2434 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://37.27.246.226:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4230-1-0-d-e284590f49&limit=500&resourceVersion=0\": dial tcp 37.27.246.226:6443: connect: connection refused" logger="UnhandledError" Mar 19 11:57:02.804349 systemd[1]: Started cri-containerd-9f10d9e4adbc8baea36d71180b83e9c5adc60a9f1123335a1e2543df3606fa2b.scope - libcontainer container 9f10d9e4adbc8baea36d71180b83e9c5adc60a9f1123335a1e2543df3606fa2b. Mar 19 11:57:02.813050 systemd[1]: Started cri-containerd-122877ffcbd219d054c6a57067aad884b83ba48abffec4d113c516ed0fa28c61.scope - libcontainer container 122877ffcbd219d054c6a57067aad884b83ba48abffec4d113c516ed0fa28c61. Mar 19 11:57:02.837851 containerd[1531]: time="2025-03-19T11:57:02.837799492Z" level=info msg="StartContainer for \"c49290af89039639c3abe160eed3e98e253f51662a295c8d10c71a212cbc9f6d\" returns successfully" Mar 19 11:57:02.880074 containerd[1531]: time="2025-03-19T11:57:02.879519115Z" level=info msg="StartContainer for \"9f10d9e4adbc8baea36d71180b83e9c5adc60a9f1123335a1e2543df3606fa2b\" returns successfully" Mar 19 11:57:02.883474 kubelet[2434]: E0319 11:57:02.883423 2434 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://37.27.246.226:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4230-1-0-d-e284590f49?timeout=10s\": dial tcp 37.27.246.226:6443: connect: connection refused" interval="1.6s" Mar 19 11:57:02.889313 containerd[1531]: time="2025-03-19T11:57:02.889245558Z" level=info msg="StartContainer for \"122877ffcbd219d054c6a57067aad884b83ba48abffec4d113c516ed0fa28c61\" returns successfully" Mar 19 11:57:03.055074 kubelet[2434]: I0319 11:57:03.054964 2434 kubelet_node_status.go:72] "Attempting to register node" node="ci-4230-1-0-d-e284590f49" Mar 19 11:57:03.055901 kubelet[2434]: E0319 11:57:03.055849 2434 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://37.27.246.226:6443/api/v1/nodes\": dial tcp 37.27.246.226:6443: connect: connection refused" node="ci-4230-1-0-d-e284590f49" Mar 19 11:57:03.079797 kubelet[2434]: W0319 11:57:03.079704 2434 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://37.27.246.226:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 37.27.246.226:6443: connect: connection refused Mar 19 11:57:03.079797 kubelet[2434]: E0319 11:57:03.079798 2434 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://37.27.246.226:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 37.27.246.226:6443: connect: connection refused" logger="UnhandledError" Mar 19 11:57:04.648123 kubelet[2434]: E0319 11:57:04.648078 2434 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-4230-1-0-d-e284590f49\" not found" node="ci-4230-1-0-d-e284590f49" Mar 19 11:57:04.659725 kubelet[2434]: I0319 11:57:04.659268 2434 kubelet_node_status.go:72] "Attempting to register node" node="ci-4230-1-0-d-e284590f49" Mar 19 11:57:04.787630 kubelet[2434]: I0319 11:57:04.787594 2434 kubelet_node_status.go:75] "Successfully registered node" node="ci-4230-1-0-d-e284590f49" Mar 19 11:57:04.787965 kubelet[2434]: E0319 11:57:04.787850 2434 kubelet_node_status.go:535] "Error updating node status, will retry" err="error getting node \"ci-4230-1-0-d-e284590f49\": node \"ci-4230-1-0-d-e284590f49\" not found" Mar 19 11:57:04.804485 kubelet[2434]: E0319 11:57:04.804437 2434 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4230-1-0-d-e284590f49\" not found" Mar 19 11:57:04.905493 kubelet[2434]: E0319 11:57:04.905289 2434 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4230-1-0-d-e284590f49\" not found" Mar 19 11:57:05.006464 kubelet[2434]: E0319 11:57:05.006399 2434 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4230-1-0-d-e284590f49\" not found" Mar 19 11:57:05.107418 kubelet[2434]: E0319 11:57:05.107366 2434 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4230-1-0-d-e284590f49\" not found" Mar 19 11:57:05.208214 kubelet[2434]: E0319 11:57:05.208020 2434 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4230-1-0-d-e284590f49\" not found" Mar 19 11:57:05.449617 kubelet[2434]: I0319 11:57:05.449559 2434 apiserver.go:52] "Watching apiserver" Mar 19 11:57:05.482381 kubelet[2434]: I0319 11:57:05.482204 2434 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 19 11:57:06.977977 systemd[1]: Reload requested from client PID 2709 ('systemctl') (unit session-7.scope)... Mar 19 11:57:06.977999 systemd[1]: Reloading... Mar 19 11:57:07.107934 zram_generator::config[2757]: No configuration found. Mar 19 11:57:07.251710 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 19 11:57:07.420699 systemd[1]: Reloading finished in 442 ms. Mar 19 11:57:07.448566 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Mar 19 11:57:07.458241 systemd[1]: kubelet.service: Deactivated successfully. Mar 19 11:57:07.458491 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 19 11:57:07.458560 systemd[1]: kubelet.service: Consumed 856ms CPU time, 113.4M memory peak. Mar 19 11:57:07.465273 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 19 11:57:07.591071 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 19 11:57:07.595403 (kubelet)[2805]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Mar 19 11:57:07.660000 kubelet[2805]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 19 11:57:07.661377 kubelet[2805]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Mar 19 11:57:07.661377 kubelet[2805]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 19 11:57:07.661377 kubelet[2805]: I0319 11:57:07.660434 2805 server.go:206] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 19 11:57:07.669770 kubelet[2805]: I0319 11:57:07.669730 2805 server.go:486] "Kubelet version" kubeletVersion="v1.31.0" Mar 19 11:57:07.669897 kubelet[2805]: I0319 11:57:07.669890 2805 server.go:488] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 19 11:57:07.670158 kubelet[2805]: I0319 11:57:07.670148 2805 server.go:929] "Client rotation is on, will bootstrap in background" Mar 19 11:57:07.671358 kubelet[2805]: I0319 11:57:07.671290 2805 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Mar 19 11:57:07.675845 kubelet[2805]: I0319 11:57:07.675815 2805 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Mar 19 11:57:07.679932 kubelet[2805]: E0319 11:57:07.678984 2805 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Mar 19 11:57:07.679932 kubelet[2805]: I0319 11:57:07.679010 2805 server.go:1403] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Mar 19 11:57:07.681491 kubelet[2805]: I0319 11:57:07.681457 2805 server.go:744] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Mar 19 11:57:07.682238 kubelet[2805]: I0319 11:57:07.682206 2805 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Mar 19 11:57:07.682358 kubelet[2805]: I0319 11:57:07.682320 2805 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 19 11:57:07.682510 kubelet[2805]: I0319 11:57:07.682347 2805 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4230-1-0-d-e284590f49","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 19 11:57:07.682510 kubelet[2805]: I0319 11:57:07.682502 2805 topology_manager.go:138] "Creating topology manager with none policy" Mar 19 11:57:07.682510 kubelet[2805]: I0319 11:57:07.682511 2805 container_manager_linux.go:300] "Creating device plugin manager" Mar 19 11:57:07.682661 kubelet[2805]: I0319 11:57:07.682538 2805 state_mem.go:36] "Initialized new in-memory state store" Mar 19 11:57:07.682661 kubelet[2805]: I0319 11:57:07.682618 2805 kubelet.go:408] "Attempting to sync node with API server" Mar 19 11:57:07.682661 kubelet[2805]: I0319 11:57:07.682627 2805 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 19 11:57:07.682661 kubelet[2805]: I0319 11:57:07.682651 2805 kubelet.go:314] "Adding apiserver pod source" Mar 19 11:57:07.682661 kubelet[2805]: I0319 11:57:07.682663 2805 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 19 11:57:07.689152 kubelet[2805]: I0319 11:57:07.688987 2805 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v1.7.23" apiVersion="v1" Mar 19 11:57:07.689500 kubelet[2805]: I0319 11:57:07.689481 2805 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Mar 19 11:57:07.698687 kubelet[2805]: I0319 11:57:07.698500 2805 server.go:1269] "Started kubelet" Mar 19 11:57:07.700888 kubelet[2805]: I0319 11:57:07.700831 2805 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 19 11:57:07.701405 kubelet[2805]: I0319 11:57:07.701170 2805 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 19 11:57:07.705341 kubelet[2805]: I0319 11:57:07.704602 2805 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Mar 19 11:57:07.705894 kubelet[2805]: I0319 11:57:07.705502 2805 server.go:460] "Adding debug handlers to kubelet server" Mar 19 11:57:07.706027 kubelet[2805]: I0319 11:57:07.706014 2805 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 19 11:57:07.712520 kubelet[2805]: I0319 11:57:07.712478 2805 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Mar 19 11:57:07.714605 kubelet[2805]: I0319 11:57:07.714584 2805 volume_manager.go:289] "Starting Kubelet Volume Manager" Mar 19 11:57:07.714992 kubelet[2805]: E0319 11:57:07.714963 2805 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4230-1-0-d-e284590f49\" not found" Mar 19 11:57:07.717770 kubelet[2805]: I0319 11:57:07.717724 2805 factory.go:221] Registration of the systemd container factory successfully Mar 19 11:57:07.717916 kubelet[2805]: I0319 11:57:07.717892 2805 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Mar 19 11:57:07.718524 kubelet[2805]: I0319 11:57:07.718299 2805 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 19 11:57:07.718524 kubelet[2805]: I0319 11:57:07.718405 2805 reconciler.go:26] "Reconciler: start to sync state" Mar 19 11:57:07.720398 kubelet[2805]: I0319 11:57:07.720261 2805 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Mar 19 11:57:07.725031 kubelet[2805]: I0319 11:57:07.724947 2805 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Mar 19 11:57:07.725031 kubelet[2805]: I0319 11:57:07.724978 2805 status_manager.go:217] "Starting to sync pod status with apiserver" Mar 19 11:57:07.725031 kubelet[2805]: I0319 11:57:07.724998 2805 kubelet.go:2321] "Starting kubelet main sync loop" Mar 19 11:57:07.725404 kubelet[2805]: E0319 11:57:07.725049 2805 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 19 11:57:07.727355 kubelet[2805]: I0319 11:57:07.727338 2805 factory.go:221] Registration of the containerd container factory successfully Mar 19 11:57:07.780497 kubelet[2805]: I0319 11:57:07.780451 2805 cpu_manager.go:214] "Starting CPU manager" policy="none" Mar 19 11:57:07.780497 kubelet[2805]: I0319 11:57:07.780468 2805 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Mar 19 11:57:07.780497 kubelet[2805]: I0319 11:57:07.780485 2805 state_mem.go:36] "Initialized new in-memory state store" Mar 19 11:57:07.780708 kubelet[2805]: I0319 11:57:07.780657 2805 state_mem.go:88] "Updated default CPUSet" cpuSet="" Mar 19 11:57:07.780708 kubelet[2805]: I0319 11:57:07.780667 2805 state_mem.go:96] "Updated CPUSet assignments" assignments={} Mar 19 11:57:07.780708 kubelet[2805]: I0319 11:57:07.780684 2805 policy_none.go:49] "None policy: Start" Mar 19 11:57:07.781418 kubelet[2805]: I0319 11:57:07.781390 2805 memory_manager.go:170] "Starting memorymanager" policy="None" Mar 19 11:57:07.781418 kubelet[2805]: I0319 11:57:07.781408 2805 state_mem.go:35] "Initializing new in-memory state store" Mar 19 11:57:07.781585 kubelet[2805]: I0319 11:57:07.781556 2805 state_mem.go:75] "Updated machine memory state" Mar 19 11:57:07.787253 kubelet[2805]: I0319 11:57:07.787223 2805 manager.go:510] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Mar 19 11:57:07.787830 kubelet[2805]: I0319 11:57:07.787592 2805 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 19 11:57:07.787830 kubelet[2805]: I0319 11:57:07.787611 2805 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 19 11:57:07.787932 kubelet[2805]: I0319 11:57:07.787909 2805 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 19 11:57:07.833916 kubelet[2805]: E0319 11:57:07.833862 2805 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-ci-4230-1-0-d-e284590f49\" already exists" pod="kube-system/kube-apiserver-ci-4230-1-0-d-e284590f49" Mar 19 11:57:07.892432 kubelet[2805]: I0319 11:57:07.891823 2805 kubelet_node_status.go:72] "Attempting to register node" node="ci-4230-1-0-d-e284590f49" Mar 19 11:57:07.900240 kubelet[2805]: I0319 11:57:07.899925 2805 kubelet_node_status.go:111] "Node was previously registered" node="ci-4230-1-0-d-e284590f49" Mar 19 11:57:07.900240 kubelet[2805]: I0319 11:57:07.900033 2805 kubelet_node_status.go:75] "Successfully registered node" node="ci-4230-1-0-d-e284590f49" Mar 19 11:57:07.992259 sudo[2837]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Mar 19 11:57:07.992789 sudo[2837]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=0) Mar 19 11:57:08.020134 kubelet[2805]: I0319 11:57:08.020035 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/0c2706b2816e41ac0a665b5d91275921-ca-certs\") pod \"kube-apiserver-ci-4230-1-0-d-e284590f49\" (UID: \"0c2706b2816e41ac0a665b5d91275921\") " pod="kube-system/kube-apiserver-ci-4230-1-0-d-e284590f49" Mar 19 11:57:08.021208 kubelet[2805]: I0319 11:57:08.020760 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/0c2706b2816e41ac0a665b5d91275921-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4230-1-0-d-e284590f49\" (UID: \"0c2706b2816e41ac0a665b5d91275921\") " pod="kube-system/kube-apiserver-ci-4230-1-0-d-e284590f49" Mar 19 11:57:08.021567 kubelet[2805]: I0319 11:57:08.020958 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/ad332766918da75e9877490f1a9e7817-ca-certs\") pod \"kube-controller-manager-ci-4230-1-0-d-e284590f49\" (UID: \"ad332766918da75e9877490f1a9e7817\") " pod="kube-system/kube-controller-manager-ci-4230-1-0-d-e284590f49" Mar 19 11:57:08.021567 kubelet[2805]: I0319 11:57:08.021759 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/ad332766918da75e9877490f1a9e7817-flexvolume-dir\") pod \"kube-controller-manager-ci-4230-1-0-d-e284590f49\" (UID: \"ad332766918da75e9877490f1a9e7817\") " pod="kube-system/kube-controller-manager-ci-4230-1-0-d-e284590f49" Mar 19 11:57:08.021567 kubelet[2805]: I0319 11:57:08.021823 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/ad332766918da75e9877490f1a9e7817-kubeconfig\") pod \"kube-controller-manager-ci-4230-1-0-d-e284590f49\" (UID: \"ad332766918da75e9877490f1a9e7817\") " pod="kube-system/kube-controller-manager-ci-4230-1-0-d-e284590f49" Mar 19 11:57:08.021567 kubelet[2805]: I0319 11:57:08.021850 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/ad332766918da75e9877490f1a9e7817-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4230-1-0-d-e284590f49\" (UID: \"ad332766918da75e9877490f1a9e7817\") " pod="kube-system/kube-controller-manager-ci-4230-1-0-d-e284590f49" Mar 19 11:57:08.023276 kubelet[2805]: I0319 11:57:08.023017 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/8863b05d3ad38ff6b9286f44b660a4c1-kubeconfig\") pod \"kube-scheduler-ci-4230-1-0-d-e284590f49\" (UID: \"8863b05d3ad38ff6b9286f44b660a4c1\") " pod="kube-system/kube-scheduler-ci-4230-1-0-d-e284590f49" Mar 19 11:57:08.023276 kubelet[2805]: I0319 11:57:08.023136 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/0c2706b2816e41ac0a665b5d91275921-k8s-certs\") pod \"kube-apiserver-ci-4230-1-0-d-e284590f49\" (UID: \"0c2706b2816e41ac0a665b5d91275921\") " pod="kube-system/kube-apiserver-ci-4230-1-0-d-e284590f49" Mar 19 11:57:08.023276 kubelet[2805]: I0319 11:57:08.023212 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/ad332766918da75e9877490f1a9e7817-k8s-certs\") pod \"kube-controller-manager-ci-4230-1-0-d-e284590f49\" (UID: \"ad332766918da75e9877490f1a9e7817\") " pod="kube-system/kube-controller-manager-ci-4230-1-0-d-e284590f49" Mar 19 11:57:08.549160 sudo[2837]: pam_unix(sudo:session): session closed for user root Mar 19 11:57:08.682970 kubelet[2805]: I0319 11:57:08.682922 2805 apiserver.go:52] "Watching apiserver" Mar 19 11:57:08.718944 kubelet[2805]: I0319 11:57:08.718896 2805 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 19 11:57:08.766729 kubelet[2805]: E0319 11:57:08.766687 2805 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-ci-4230-1-0-d-e284590f49\" already exists" pod="kube-system/kube-apiserver-ci-4230-1-0-d-e284590f49" Mar 19 11:57:08.811694 kubelet[2805]: I0319 11:57:08.811482 2805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4230-1-0-d-e284590f49" podStartSLOduration=2.811462598 podStartE2EDuration="2.811462598s" podCreationTimestamp="2025-03-19 11:57:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-19 11:57:08.809320652 +0000 UTC m=+1.198515138" watchObservedRunningTime="2025-03-19 11:57:08.811462598 +0000 UTC m=+1.200657075" Mar 19 11:57:08.811694 kubelet[2805]: I0319 11:57:08.811606 2805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4230-1-0-d-e284590f49" podStartSLOduration=1.811602259 podStartE2EDuration="1.811602259s" podCreationTimestamp="2025-03-19 11:57:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-19 11:57:08.797995971 +0000 UTC m=+1.187190447" watchObservedRunningTime="2025-03-19 11:57:08.811602259 +0000 UTC m=+1.200796736" Mar 19 11:57:08.837237 kubelet[2805]: I0319 11:57:08.837172 2805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4230-1-0-d-e284590f49" podStartSLOduration=1.8371536480000001 podStartE2EDuration="1.837153648s" podCreationTimestamp="2025-03-19 11:57:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-19 11:57:08.824849395 +0000 UTC m=+1.214043872" watchObservedRunningTime="2025-03-19 11:57:08.837153648 +0000 UTC m=+1.226348125" Mar 19 11:57:10.209561 sudo[1893]: pam_unix(sudo:session): session closed for user root Mar 19 11:57:10.371164 sshd[1892]: Connection closed by 139.178.68.195 port 48482 Mar 19 11:57:10.372266 sshd-session[1890]: pam_unix(sshd:session): session closed for user core Mar 19 11:57:10.376616 systemd[1]: sshd@6-37.27.246.226:22-139.178.68.195:48482.service: Deactivated successfully. Mar 19 11:57:10.379930 systemd[1]: session-7.scope: Deactivated successfully. Mar 19 11:57:10.380119 systemd[1]: session-7.scope: Consumed 5.983s CPU time, 213M memory peak. Mar 19 11:57:10.381485 systemd-logind[1502]: Session 7 logged out. Waiting for processes to exit. Mar 19 11:57:10.382839 systemd-logind[1502]: Removed session 7. Mar 19 11:57:13.319397 kubelet[2805]: I0319 11:57:13.319324 2805 kuberuntime_manager.go:1633] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Mar 19 11:57:13.320109 containerd[1531]: time="2025-03-19T11:57:13.319972322Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Mar 19 11:57:13.320496 kubelet[2805]: I0319 11:57:13.320386 2805 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Mar 19 11:57:14.181970 systemd[1]: Created slice kubepods-besteffort-poda649c4aa_342a_4186_9c4e_ce243df6a712.slice - libcontainer container kubepods-besteffort-poda649c4aa_342a_4186_9c4e_ce243df6a712.slice. Mar 19 11:57:14.209492 systemd[1]: Created slice kubepods-burstable-pod40347df1_b19f_4102_8759_3c41dc3a468d.slice - libcontainer container kubepods-burstable-pod40347df1_b19f_4102_8759_3c41dc3a468d.slice. Mar 19 11:57:14.215385 kubelet[2805]: W0319 11:57:14.215324 2805 reflector.go:561] object-"kube-system"/"cilium-config": failed to list *v1.ConfigMap: configmaps "cilium-config" is forbidden: User "system:node:ci-4230-1-0-d-e284590f49" cannot list resource "configmaps" in API group "" in the namespace "kube-system": no relationship found between node 'ci-4230-1-0-d-e284590f49' and this object Mar 19 11:57:14.215385 kubelet[2805]: E0319 11:57:14.215373 2805 reflector.go:158] "Unhandled Error" err="object-\"kube-system\"/\"cilium-config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"cilium-config\" is forbidden: User \"system:node:ci-4230-1-0-d-e284590f49\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"kube-system\": no relationship found between node 'ci-4230-1-0-d-e284590f49' and this object" logger="UnhandledError" Mar 19 11:57:14.215570 kubelet[2805]: W0319 11:57:14.215411 2805 reflector.go:561] object-"kube-system"/"cilium-clustermesh": failed to list *v1.Secret: secrets "cilium-clustermesh" is forbidden: User "system:node:ci-4230-1-0-d-e284590f49" cannot list resource "secrets" in API group "" in the namespace "kube-system": no relationship found between node 'ci-4230-1-0-d-e284590f49' and this object Mar 19 11:57:14.215570 kubelet[2805]: E0319 11:57:14.215420 2805 reflector.go:158] "Unhandled Error" err="object-\"kube-system\"/\"cilium-clustermesh\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"cilium-clustermesh\" is forbidden: User \"system:node:ci-4230-1-0-d-e284590f49\" cannot list resource \"secrets\" in API group \"\" in the namespace \"kube-system\": no relationship found between node 'ci-4230-1-0-d-e284590f49' and this object" logger="UnhandledError" Mar 19 11:57:14.215570 kubelet[2805]: W0319 11:57:14.215442 2805 reflector.go:561] object-"kube-system"/"hubble-server-certs": failed to list *v1.Secret: secrets "hubble-server-certs" is forbidden: User "system:node:ci-4230-1-0-d-e284590f49" cannot list resource "secrets" in API group "" in the namespace "kube-system": no relationship found between node 'ci-4230-1-0-d-e284590f49' and this object Mar 19 11:57:14.215570 kubelet[2805]: E0319 11:57:14.215454 2805 reflector.go:158] "Unhandled Error" err="object-\"kube-system\"/\"hubble-server-certs\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"hubble-server-certs\" is forbidden: User \"system:node:ci-4230-1-0-d-e284590f49\" cannot list resource \"secrets\" in API group \"\" in the namespace \"kube-system\": no relationship found between node 'ci-4230-1-0-d-e284590f49' and this object" logger="UnhandledError" Mar 19 11:57:14.260986 kubelet[2805]: I0319 11:57:14.260529 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xs64r\" (UniqueName: \"kubernetes.io/projected/a649c4aa-342a-4186-9c4e-ce243df6a712-kube-api-access-xs64r\") pod \"kube-proxy-tv9z2\" (UID: \"a649c4aa-342a-4186-9c4e-ce243df6a712\") " pod="kube-system/kube-proxy-tv9z2" Mar 19 11:57:14.260986 kubelet[2805]: I0319 11:57:14.260562 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-cni-path\") pod \"cilium-fx77j\" (UID: \"40347df1-b19f-4102-8759-3c41dc3a468d\") " pod="kube-system/cilium-fx77j" Mar 19 11:57:14.260986 kubelet[2805]: I0319 11:57:14.260578 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/a649c4aa-342a-4186-9c4e-ce243df6a712-xtables-lock\") pod \"kube-proxy-tv9z2\" (UID: \"a649c4aa-342a-4186-9c4e-ce243df6a712\") " pod="kube-system/kube-proxy-tv9z2" Mar 19 11:57:14.260986 kubelet[2805]: I0319 11:57:14.260596 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-bpf-maps\") pod \"cilium-fx77j\" (UID: \"40347df1-b19f-4102-8759-3c41dc3a468d\") " pod="kube-system/cilium-fx77j" Mar 19 11:57:14.260986 kubelet[2805]: I0319 11:57:14.260614 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/40347df1-b19f-4102-8759-3c41dc3a468d-hubble-tls\") pod \"cilium-fx77j\" (UID: \"40347df1-b19f-4102-8759-3c41dc3a468d\") " pod="kube-system/cilium-fx77j" Mar 19 11:57:14.260986 kubelet[2805]: I0319 11:57:14.260631 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-cilium-run\") pod \"cilium-fx77j\" (UID: \"40347df1-b19f-4102-8759-3c41dc3a468d\") " pod="kube-system/cilium-fx77j" Mar 19 11:57:14.261235 kubelet[2805]: I0319 11:57:14.260642 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-lib-modules\") pod \"cilium-fx77j\" (UID: \"40347df1-b19f-4102-8759-3c41dc3a468d\") " pod="kube-system/cilium-fx77j" Mar 19 11:57:14.261235 kubelet[2805]: I0319 11:57:14.260657 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-host-proc-sys-net\") pod \"cilium-fx77j\" (UID: \"40347df1-b19f-4102-8759-3c41dc3a468d\") " pod="kube-system/cilium-fx77j" Mar 19 11:57:14.261235 kubelet[2805]: I0319 11:57:14.260672 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/40347df1-b19f-4102-8759-3c41dc3a468d-cilium-config-path\") pod \"cilium-fx77j\" (UID: \"40347df1-b19f-4102-8759-3c41dc3a468d\") " pod="kube-system/cilium-fx77j" Mar 19 11:57:14.261235 kubelet[2805]: I0319 11:57:14.260689 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsttr\" (UniqueName: \"kubernetes.io/projected/40347df1-b19f-4102-8759-3c41dc3a468d-kube-api-access-dsttr\") pod \"cilium-fx77j\" (UID: \"40347df1-b19f-4102-8759-3c41dc3a468d\") " pod="kube-system/cilium-fx77j" Mar 19 11:57:14.261235 kubelet[2805]: I0319 11:57:14.260702 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a649c4aa-342a-4186-9c4e-ce243df6a712-lib-modules\") pod \"kube-proxy-tv9z2\" (UID: \"a649c4aa-342a-4186-9c4e-ce243df6a712\") " pod="kube-system/kube-proxy-tv9z2" Mar 19 11:57:14.261235 kubelet[2805]: I0319 11:57:14.260714 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-hostproc\") pod \"cilium-fx77j\" (UID: \"40347df1-b19f-4102-8759-3c41dc3a468d\") " pod="kube-system/cilium-fx77j" Mar 19 11:57:14.261380 kubelet[2805]: I0319 11:57:14.260727 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-xtables-lock\") pod \"cilium-fx77j\" (UID: \"40347df1-b19f-4102-8759-3c41dc3a468d\") " pod="kube-system/cilium-fx77j" Mar 19 11:57:14.261380 kubelet[2805]: I0319 11:57:14.260769 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/a649c4aa-342a-4186-9c4e-ce243df6a712-kube-proxy\") pod \"kube-proxy-tv9z2\" (UID: \"a649c4aa-342a-4186-9c4e-ce243df6a712\") " pod="kube-system/kube-proxy-tv9z2" Mar 19 11:57:14.261380 kubelet[2805]: I0319 11:57:14.260782 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-host-proc-sys-kernel\") pod \"cilium-fx77j\" (UID: \"40347df1-b19f-4102-8759-3c41dc3a468d\") " pod="kube-system/cilium-fx77j" Mar 19 11:57:14.261380 kubelet[2805]: I0319 11:57:14.260797 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-cilium-cgroup\") pod \"cilium-fx77j\" (UID: \"40347df1-b19f-4102-8759-3c41dc3a468d\") " pod="kube-system/cilium-fx77j" Mar 19 11:57:14.261380 kubelet[2805]: I0319 11:57:14.260809 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-etc-cni-netd\") pod \"cilium-fx77j\" (UID: \"40347df1-b19f-4102-8759-3c41dc3a468d\") " pod="kube-system/cilium-fx77j" Mar 19 11:57:14.261380 kubelet[2805]: I0319 11:57:14.260822 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/40347df1-b19f-4102-8759-3c41dc3a468d-clustermesh-secrets\") pod \"cilium-fx77j\" (UID: \"40347df1-b19f-4102-8759-3c41dc3a468d\") " pod="kube-system/cilium-fx77j" Mar 19 11:57:14.427904 systemd[1]: Created slice kubepods-besteffort-pod4ad24ea6_768c_404d_9de4_78cdd82836aa.slice - libcontainer container kubepods-besteffort-pod4ad24ea6_768c_404d_9de4_78cdd82836aa.slice. Mar 19 11:57:14.463752 kubelet[2805]: I0319 11:57:14.463609 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/4ad24ea6-768c-404d-9de4-78cdd82836aa-cilium-config-path\") pod \"cilium-operator-5d85765b45-fcz7w\" (UID: \"4ad24ea6-768c-404d-9de4-78cdd82836aa\") " pod="kube-system/cilium-operator-5d85765b45-fcz7w" Mar 19 11:57:14.464535 kubelet[2805]: I0319 11:57:14.464482 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8kbf\" (UniqueName: \"kubernetes.io/projected/4ad24ea6-768c-404d-9de4-78cdd82836aa-kube-api-access-q8kbf\") pod \"cilium-operator-5d85765b45-fcz7w\" (UID: \"4ad24ea6-768c-404d-9de4-78cdd82836aa\") " pod="kube-system/cilium-operator-5d85765b45-fcz7w" Mar 19 11:57:14.490198 containerd[1531]: time="2025-03-19T11:57:14.490075103Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-tv9z2,Uid:a649c4aa-342a-4186-9c4e-ce243df6a712,Namespace:kube-system,Attempt:0,}" Mar 19 11:57:14.519187 containerd[1531]: time="2025-03-19T11:57:14.518898223Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 19 11:57:14.522361 containerd[1531]: time="2025-03-19T11:57:14.519100512Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 19 11:57:14.522361 containerd[1531]: time="2025-03-19T11:57:14.519337636Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 19 11:57:14.522361 containerd[1531]: time="2025-03-19T11:57:14.519996027Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 19 11:57:14.551179 systemd[1]: Started cri-containerd-b542b5b6eb4eda42d765d2a1010ca6909329ef3fd2f52e2ed213132f474b706a.scope - libcontainer container b542b5b6eb4eda42d765d2a1010ca6909329ef3fd2f52e2ed213132f474b706a. Mar 19 11:57:14.582973 containerd[1531]: time="2025-03-19T11:57:14.582821056Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-tv9z2,Uid:a649c4aa-342a-4186-9c4e-ce243df6a712,Namespace:kube-system,Attempt:0,} returns sandbox id \"b542b5b6eb4eda42d765d2a1010ca6909329ef3fd2f52e2ed213132f474b706a\"" Mar 19 11:57:14.586407 containerd[1531]: time="2025-03-19T11:57:14.586168910Z" level=info msg="CreateContainer within sandbox \"b542b5b6eb4eda42d765d2a1010ca6909329ef3fd2f52e2ed213132f474b706a\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Mar 19 11:57:14.612110 containerd[1531]: time="2025-03-19T11:57:14.612038333Z" level=info msg="CreateContainer within sandbox \"b542b5b6eb4eda42d765d2a1010ca6909329ef3fd2f52e2ed213132f474b706a\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"877a7f151073bddc8c69cea0f84e2f75cb6f4c71c35b3a6e5436d02c2d07159c\"" Mar 19 11:57:14.613196 containerd[1531]: time="2025-03-19T11:57:14.613140986Z" level=info msg="StartContainer for \"877a7f151073bddc8c69cea0f84e2f75cb6f4c71c35b3a6e5436d02c2d07159c\"" Mar 19 11:57:14.642131 systemd[1]: Started cri-containerd-877a7f151073bddc8c69cea0f84e2f75cb6f4c71c35b3a6e5436d02c2d07159c.scope - libcontainer container 877a7f151073bddc8c69cea0f84e2f75cb6f4c71c35b3a6e5436d02c2d07159c. Mar 19 11:57:14.678926 containerd[1531]: time="2025-03-19T11:57:14.678801812Z" level=info msg="StartContainer for \"877a7f151073bddc8c69cea0f84e2f75cb6f4c71c35b3a6e5436d02c2d07159c\" returns successfully" Mar 19 11:57:14.778412 kubelet[2805]: I0319 11:57:14.778235 2805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-tv9z2" podStartSLOduration=0.778217023 podStartE2EDuration="778.217023ms" podCreationTimestamp="2025-03-19 11:57:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-19 11:57:14.777991972 +0000 UTC m=+7.167186448" watchObservedRunningTime="2025-03-19 11:57:14.778217023 +0000 UTC m=+7.167411499" Mar 19 11:57:15.334885 containerd[1531]: time="2025-03-19T11:57:15.334786992Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-5d85765b45-fcz7w,Uid:4ad24ea6-768c-404d-9de4-78cdd82836aa,Namespace:kube-system,Attempt:0,}" Mar 19 11:57:15.365731 kubelet[2805]: E0319 11:57:15.363814 2805 secret.go:188] Couldn't get secret kube-system/cilium-clustermesh: failed to sync secret cache: timed out waiting for the condition Mar 19 11:57:15.365731 kubelet[2805]: E0319 11:57:15.363999 2805 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/40347df1-b19f-4102-8759-3c41dc3a468d-clustermesh-secrets podName:40347df1-b19f-4102-8759-3c41dc3a468d nodeName:}" failed. No retries permitted until 2025-03-19 11:57:15.863968846 +0000 UTC m=+8.253163342 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "clustermesh-secrets" (UniqueName: "kubernetes.io/secret/40347df1-b19f-4102-8759-3c41dc3a468d-clustermesh-secrets") pod "cilium-fx77j" (UID: "40347df1-b19f-4102-8759-3c41dc3a468d") : failed to sync secret cache: timed out waiting for the condition Mar 19 11:57:15.373230 containerd[1531]: time="2025-03-19T11:57:15.373069235Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 19 11:57:15.373230 containerd[1531]: time="2025-03-19T11:57:15.373154845Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 19 11:57:15.373230 containerd[1531]: time="2025-03-19T11:57:15.373175694Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 19 11:57:15.373500 containerd[1531]: time="2025-03-19T11:57:15.373289487Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 19 11:57:15.414685 systemd[1]: run-containerd-runc-k8s.io-bb72b6cb42b223cfe8537b613d2b5912f92fab908bf4d332a5d91b06d19e9421-runc.r7tiJf.mount: Deactivated successfully. Mar 19 11:57:15.427076 systemd[1]: Started cri-containerd-bb72b6cb42b223cfe8537b613d2b5912f92fab908bf4d332a5d91b06d19e9421.scope - libcontainer container bb72b6cb42b223cfe8537b613d2b5912f92fab908bf4d332a5d91b06d19e9421. Mar 19 11:57:15.476909 containerd[1531]: time="2025-03-19T11:57:15.476749937Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-5d85765b45-fcz7w,Uid:4ad24ea6-768c-404d-9de4-78cdd82836aa,Namespace:kube-system,Attempt:0,} returns sandbox id \"bb72b6cb42b223cfe8537b613d2b5912f92fab908bf4d332a5d91b06d19e9421\"" Mar 19 11:57:15.480874 containerd[1531]: time="2025-03-19T11:57:15.480723633Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Mar 19 11:57:16.013603 containerd[1531]: time="2025-03-19T11:57:16.013517158Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-fx77j,Uid:40347df1-b19f-4102-8759-3c41dc3a468d,Namespace:kube-system,Attempt:0,}" Mar 19 11:57:16.050248 containerd[1531]: time="2025-03-19T11:57:16.050115513Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 19 11:57:16.050248 containerd[1531]: time="2025-03-19T11:57:16.050200833Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 19 11:57:16.050571 containerd[1531]: time="2025-03-19T11:57:16.050229677Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 19 11:57:16.050571 containerd[1531]: time="2025-03-19T11:57:16.050380750Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 19 11:57:16.077246 systemd[1]: Started cri-containerd-e1ac54e4ade1aee1a79f559a223841973bf4d1dbd6882bd882ecd9a1600d4dfa.scope - libcontainer container e1ac54e4ade1aee1a79f559a223841973bf4d1dbd6882bd882ecd9a1600d4dfa. Mar 19 11:57:16.111050 containerd[1531]: time="2025-03-19T11:57:16.110990506Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-fx77j,Uid:40347df1-b19f-4102-8759-3c41dc3a468d,Namespace:kube-system,Attempt:0,} returns sandbox id \"e1ac54e4ade1aee1a79f559a223841973bf4d1dbd6882bd882ecd9a1600d4dfa\"" Mar 19 11:57:18.805043 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2631982192.mount: Deactivated successfully. Mar 19 11:57:22.670214 containerd[1531]: time="2025-03-19T11:57:22.670143566Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 19 11:57:22.671463 containerd[1531]: time="2025-03-19T11:57:22.671407242Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=18904197" Mar 19 11:57:22.672772 containerd[1531]: time="2025-03-19T11:57:22.672634860Z" level=info msg="ImageCreate event name:\"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 19 11:57:22.674248 containerd[1531]: time="2025-03-19T11:57:22.673702378Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"18897442\" in 7.192928009s" Mar 19 11:57:22.674248 containerd[1531]: time="2025-03-19T11:57:22.673736081Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"" Mar 19 11:57:22.676339 containerd[1531]: time="2025-03-19T11:57:22.676292937Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Mar 19 11:57:22.677984 containerd[1531]: time="2025-03-19T11:57:22.677941142Z" level=info msg="CreateContainer within sandbox \"bb72b6cb42b223cfe8537b613d2b5912f92fab908bf4d332a5d91b06d19e9421\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Mar 19 11:57:22.692761 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1534457509.mount: Deactivated successfully. Mar 19 11:57:22.696032 containerd[1531]: time="2025-03-19T11:57:22.695990858Z" level=info msg="CreateContainer within sandbox \"bb72b6cb42b223cfe8537b613d2b5912f92fab908bf4d332a5d91b06d19e9421\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"62602a2adab8d404be04a797fe34a5e26d855a2873ca6e298bd07f4e0488b846\"" Mar 19 11:57:22.696479 containerd[1531]: time="2025-03-19T11:57:22.696465878Z" level=info msg="StartContainer for \"62602a2adab8d404be04a797fe34a5e26d855a2873ca6e298bd07f4e0488b846\"" Mar 19 11:57:22.731071 systemd[1]: Started cri-containerd-62602a2adab8d404be04a797fe34a5e26d855a2873ca6e298bd07f4e0488b846.scope - libcontainer container 62602a2adab8d404be04a797fe34a5e26d855a2873ca6e298bd07f4e0488b846. Mar 19 11:57:22.750771 containerd[1531]: time="2025-03-19T11:57:22.750697580Z" level=info msg="StartContainer for \"62602a2adab8d404be04a797fe34a5e26d855a2873ca6e298bd07f4e0488b846\" returns successfully" Mar 19 11:57:29.972408 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount957947529.mount: Deactivated successfully. Mar 19 11:57:31.686315 containerd[1531]: time="2025-03-19T11:57:31.686252916Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 19 11:57:31.688419 containerd[1531]: time="2025-03-19T11:57:31.688348460Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=166730503" Mar 19 11:57:31.688744 containerd[1531]: time="2025-03-19T11:57:31.688694298Z" level=info msg="ImageCreate event name:\"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 19 11:57:31.690918 containerd[1531]: time="2025-03-19T11:57:31.690509387Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"166719855\" in 9.013522811s" Mar 19 11:57:31.690918 containerd[1531]: time="2025-03-19T11:57:31.690556856Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"" Mar 19 11:57:31.698350 containerd[1531]: time="2025-03-19T11:57:31.697729967Z" level=info msg="CreateContainer within sandbox \"e1ac54e4ade1aee1a79f559a223841973bf4d1dbd6882bd882ecd9a1600d4dfa\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Mar 19 11:57:31.748675 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount499489393.mount: Deactivated successfully. Mar 19 11:57:31.750380 containerd[1531]: time="2025-03-19T11:57:31.750316736Z" level=info msg="CreateContainer within sandbox \"e1ac54e4ade1aee1a79f559a223841973bf4d1dbd6882bd882ecd9a1600d4dfa\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"b4840c728b6ff37e8c43f65799cba5c27b86a145f3d77597741805d13079d723\"" Mar 19 11:57:31.751197 containerd[1531]: time="2025-03-19T11:57:31.750940523Z" level=info msg="StartContainer for \"b4840c728b6ff37e8c43f65799cba5c27b86a145f3d77597741805d13079d723\"" Mar 19 11:57:31.957163 systemd[1]: Started cri-containerd-b4840c728b6ff37e8c43f65799cba5c27b86a145f3d77597741805d13079d723.scope - libcontainer container b4840c728b6ff37e8c43f65799cba5c27b86a145f3d77597741805d13079d723. Mar 19 11:57:31.983795 containerd[1531]: time="2025-03-19T11:57:31.983631686Z" level=info msg="StartContainer for \"b4840c728b6ff37e8c43f65799cba5c27b86a145f3d77597741805d13079d723\" returns successfully" Mar 19 11:57:31.995694 systemd[1]: cri-containerd-b4840c728b6ff37e8c43f65799cba5c27b86a145f3d77597741805d13079d723.scope: Deactivated successfully. Mar 19 11:57:32.089827 containerd[1531]: time="2025-03-19T11:57:32.067103060Z" level=info msg="shim disconnected" id=b4840c728b6ff37e8c43f65799cba5c27b86a145f3d77597741805d13079d723 namespace=k8s.io Mar 19 11:57:32.089827 containerd[1531]: time="2025-03-19T11:57:32.089817084Z" level=warning msg="cleaning up after shim disconnected" id=b4840c728b6ff37e8c43f65799cba5c27b86a145f3d77597741805d13079d723 namespace=k8s.io Mar 19 11:57:32.090182 containerd[1531]: time="2025-03-19T11:57:32.089841298Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 19 11:57:32.742306 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-b4840c728b6ff37e8c43f65799cba5c27b86a145f3d77597741805d13079d723-rootfs.mount: Deactivated successfully. Mar 19 11:57:32.864906 containerd[1531]: time="2025-03-19T11:57:32.864819690Z" level=info msg="CreateContainer within sandbox \"e1ac54e4ade1aee1a79f559a223841973bf4d1dbd6882bd882ecd9a1600d4dfa\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Mar 19 11:57:32.893308 containerd[1531]: time="2025-03-19T11:57:32.893241552Z" level=info msg="CreateContainer within sandbox \"e1ac54e4ade1aee1a79f559a223841973bf4d1dbd6882bd882ecd9a1600d4dfa\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"5b31e2b3a79792be2a445f1ccc7411778302d0b2ad34abfd194de17bc0795665\"" Mar 19 11:57:32.895279 containerd[1531]: time="2025-03-19T11:57:32.894062460Z" level=info msg="StartContainer for \"5b31e2b3a79792be2a445f1ccc7411778302d0b2ad34abfd194de17bc0795665\"" Mar 19 11:57:32.919956 kubelet[2805]: I0319 11:57:32.916141 2805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-5d85765b45-fcz7w" podStartSLOduration=11.718329472 podStartE2EDuration="18.916116015s" podCreationTimestamp="2025-03-19 11:57:14 +0000 UTC" firstStartedPulling="2025-03-19 11:57:15.47832275 +0000 UTC m=+7.867517236" lastFinishedPulling="2025-03-19 11:57:22.676109303 +0000 UTC m=+15.065303779" observedRunningTime="2025-03-19 11:57:22.796280463 +0000 UTC m=+15.185474939" watchObservedRunningTime="2025-03-19 11:57:32.916116015 +0000 UTC m=+25.305310511" Mar 19 11:57:32.957046 systemd[1]: Started cri-containerd-5b31e2b3a79792be2a445f1ccc7411778302d0b2ad34abfd194de17bc0795665.scope - libcontainer container 5b31e2b3a79792be2a445f1ccc7411778302d0b2ad34abfd194de17bc0795665. Mar 19 11:57:32.988228 containerd[1531]: time="2025-03-19T11:57:32.988098628Z" level=info msg="StartContainer for \"5b31e2b3a79792be2a445f1ccc7411778302d0b2ad34abfd194de17bc0795665\" returns successfully" Mar 19 11:57:33.004728 systemd[1]: systemd-sysctl.service: Deactivated successfully. Mar 19 11:57:33.005702 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Mar 19 11:57:33.006305 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables... Mar 19 11:57:33.013232 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Mar 19 11:57:33.014048 systemd[1]: cri-containerd-5b31e2b3a79792be2a445f1ccc7411778302d0b2ad34abfd194de17bc0795665.scope: Deactivated successfully. Mar 19 11:57:33.048002 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Mar 19 11:57:33.050736 containerd[1531]: time="2025-03-19T11:57:33.050650502Z" level=info msg="shim disconnected" id=5b31e2b3a79792be2a445f1ccc7411778302d0b2ad34abfd194de17bc0795665 namespace=k8s.io Mar 19 11:57:33.050736 containerd[1531]: time="2025-03-19T11:57:33.050722507Z" level=warning msg="cleaning up after shim disconnected" id=5b31e2b3a79792be2a445f1ccc7411778302d0b2ad34abfd194de17bc0795665 namespace=k8s.io Mar 19 11:57:33.050736 containerd[1531]: time="2025-03-19T11:57:33.050731644Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 19 11:57:33.742368 systemd[1]: run-containerd-runc-k8s.io-5b31e2b3a79792be2a445f1ccc7411778302d0b2ad34abfd194de17bc0795665-runc.u8nb3b.mount: Deactivated successfully. Mar 19 11:57:33.742487 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-5b31e2b3a79792be2a445f1ccc7411778302d0b2ad34abfd194de17bc0795665-rootfs.mount: Deactivated successfully. Mar 19 11:57:33.866337 containerd[1531]: time="2025-03-19T11:57:33.866127238Z" level=info msg="CreateContainer within sandbox \"e1ac54e4ade1aee1a79f559a223841973bf4d1dbd6882bd882ecd9a1600d4dfa\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Mar 19 11:57:33.919700 containerd[1531]: time="2025-03-19T11:57:33.919636080Z" level=info msg="CreateContainer within sandbox \"e1ac54e4ade1aee1a79f559a223841973bf4d1dbd6882bd882ecd9a1600d4dfa\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"9745ab02b8e658d81a940624e38afabc8bf95f38df9236674e895fcdcc973f37\"" Mar 19 11:57:33.921550 containerd[1531]: time="2025-03-19T11:57:33.921119417Z" level=info msg="StartContainer for \"9745ab02b8e658d81a940624e38afabc8bf95f38df9236674e895fcdcc973f37\"" Mar 19 11:57:33.981050 systemd[1]: Started cri-containerd-9745ab02b8e658d81a940624e38afabc8bf95f38df9236674e895fcdcc973f37.scope - libcontainer container 9745ab02b8e658d81a940624e38afabc8bf95f38df9236674e895fcdcc973f37. Mar 19 11:57:34.016795 containerd[1531]: time="2025-03-19T11:57:34.016659408Z" level=info msg="StartContainer for \"9745ab02b8e658d81a940624e38afabc8bf95f38df9236674e895fcdcc973f37\" returns successfully" Mar 19 11:57:34.023808 systemd[1]: cri-containerd-9745ab02b8e658d81a940624e38afabc8bf95f38df9236674e895fcdcc973f37.scope: Deactivated successfully. Mar 19 11:57:34.024287 systemd[1]: cri-containerd-9745ab02b8e658d81a940624e38afabc8bf95f38df9236674e895fcdcc973f37.scope: Consumed 22ms CPU time, 3M memory peak, 1M read from disk. Mar 19 11:57:34.055684 containerd[1531]: time="2025-03-19T11:57:34.055612865Z" level=info msg="shim disconnected" id=9745ab02b8e658d81a940624e38afabc8bf95f38df9236674e895fcdcc973f37 namespace=k8s.io Mar 19 11:57:34.055684 containerd[1531]: time="2025-03-19T11:57:34.055667458Z" level=warning msg="cleaning up after shim disconnected" id=9745ab02b8e658d81a940624e38afabc8bf95f38df9236674e895fcdcc973f37 namespace=k8s.io Mar 19 11:57:34.055684 containerd[1531]: time="2025-03-19T11:57:34.055674500Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 19 11:57:34.743643 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-9745ab02b8e658d81a940624e38afabc8bf95f38df9236674e895fcdcc973f37-rootfs.mount: Deactivated successfully. Mar 19 11:57:34.867858 containerd[1531]: time="2025-03-19T11:57:34.867816060Z" level=info msg="CreateContainer within sandbox \"e1ac54e4ade1aee1a79f559a223841973bf4d1dbd6882bd882ecd9a1600d4dfa\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Mar 19 11:57:34.905984 containerd[1531]: time="2025-03-19T11:57:34.905892315Z" level=info msg="CreateContainer within sandbox \"e1ac54e4ade1aee1a79f559a223841973bf4d1dbd6882bd882ecd9a1600d4dfa\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"b7c808ce26d368c2639ba65a198b7646021751e63362e4aa0492551ab2280841\"" Mar 19 11:57:34.906846 containerd[1531]: time="2025-03-19T11:57:34.906801658Z" level=info msg="StartContainer for \"b7c808ce26d368c2639ba65a198b7646021751e63362e4aa0492551ab2280841\"" Mar 19 11:57:34.946316 systemd[1]: Started cri-containerd-b7c808ce26d368c2639ba65a198b7646021751e63362e4aa0492551ab2280841.scope - libcontainer container b7c808ce26d368c2639ba65a198b7646021751e63362e4aa0492551ab2280841. Mar 19 11:57:34.976401 systemd[1]: cri-containerd-b7c808ce26d368c2639ba65a198b7646021751e63362e4aa0492551ab2280841.scope: Deactivated successfully. Mar 19 11:57:34.978584 containerd[1531]: time="2025-03-19T11:57:34.978542625Z" level=info msg="StartContainer for \"b7c808ce26d368c2639ba65a198b7646021751e63362e4aa0492551ab2280841\" returns successfully" Mar 19 11:57:35.006840 containerd[1531]: time="2025-03-19T11:57:35.006605459Z" level=info msg="shim disconnected" id=b7c808ce26d368c2639ba65a198b7646021751e63362e4aa0492551ab2280841 namespace=k8s.io Mar 19 11:57:35.006840 containerd[1531]: time="2025-03-19T11:57:35.006684146Z" level=warning msg="cleaning up after shim disconnected" id=b7c808ce26d368c2639ba65a198b7646021751e63362e4aa0492551ab2280841 namespace=k8s.io Mar 19 11:57:35.006840 containerd[1531]: time="2025-03-19T11:57:35.006694606Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 19 11:57:35.018392 containerd[1531]: time="2025-03-19T11:57:35.018349903Z" level=warning msg="cleanup warnings time=\"2025-03-19T11:57:35Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Mar 19 11:57:35.743609 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-b7c808ce26d368c2639ba65a198b7646021751e63362e4aa0492551ab2280841-rootfs.mount: Deactivated successfully. Mar 19 11:57:35.871855 containerd[1531]: time="2025-03-19T11:57:35.871778648Z" level=info msg="CreateContainer within sandbox \"e1ac54e4ade1aee1a79f559a223841973bf4d1dbd6882bd882ecd9a1600d4dfa\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Mar 19 11:57:35.896077 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3448509812.mount: Deactivated successfully. Mar 19 11:57:35.900158 containerd[1531]: time="2025-03-19T11:57:35.900102642Z" level=info msg="CreateContainer within sandbox \"e1ac54e4ade1aee1a79f559a223841973bf4d1dbd6882bd882ecd9a1600d4dfa\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"e6b89623cb51d19964eb97e8eac9a82dd8778cfc1597b6f3525a394c57251513\"" Mar 19 11:57:35.900918 containerd[1531]: time="2025-03-19T11:57:35.900855832Z" level=info msg="StartContainer for \"e6b89623cb51d19964eb97e8eac9a82dd8778cfc1597b6f3525a394c57251513\"" Mar 19 11:57:35.942422 systemd[1]: Started cri-containerd-e6b89623cb51d19964eb97e8eac9a82dd8778cfc1597b6f3525a394c57251513.scope - libcontainer container e6b89623cb51d19964eb97e8eac9a82dd8778cfc1597b6f3525a394c57251513. Mar 19 11:57:35.991128 containerd[1531]: time="2025-03-19T11:57:35.990962257Z" level=info msg="StartContainer for \"e6b89623cb51d19964eb97e8eac9a82dd8778cfc1597b6f3525a394c57251513\" returns successfully" Mar 19 11:57:36.129189 kubelet[2805]: I0319 11:57:36.129156 2805 kubelet_node_status.go:488] "Fast updating node status as it just became ready" Mar 19 11:57:36.175212 systemd[1]: Created slice kubepods-burstable-pod5fb07861_4622_4fa3_b2c6_8fc002ee237c.slice - libcontainer container kubepods-burstable-pod5fb07861_4622_4fa3_b2c6_8fc002ee237c.slice. Mar 19 11:57:36.192816 systemd[1]: Created slice kubepods-burstable-podab6099c0_0f0b_4131_9d0a_2e2b7559e303.slice - libcontainer container kubepods-burstable-podab6099c0_0f0b_4131_9d0a_2e2b7559e303.slice. Mar 19 11:57:36.199898 kubelet[2805]: W0319 11:57:36.199720 2805 reflector.go:561] object-"kube-system"/"coredns": failed to list *v1.ConfigMap: configmaps "coredns" is forbidden: User "system:node:ci-4230-1-0-d-e284590f49" cannot list resource "configmaps" in API group "" in the namespace "kube-system": no relationship found between node 'ci-4230-1-0-d-e284590f49' and this object Mar 19 11:57:36.200285 kubelet[2805]: E0319 11:57:36.200068 2805 reflector.go:158] "Unhandled Error" err="object-\"kube-system\"/\"coredns\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"coredns\" is forbidden: User \"system:node:ci-4230-1-0-d-e284590f49\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"kube-system\": no relationship found between node 'ci-4230-1-0-d-e284590f49' and this object" logger="UnhandledError" Mar 19 11:57:36.228112 kubelet[2805]: I0319 11:57:36.228070 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ab6099c0-0f0b-4131-9d0a-2e2b7559e303-config-volume\") pod \"coredns-6f6b679f8f-jwjkr\" (UID: \"ab6099c0-0f0b-4131-9d0a-2e2b7559e303\") " pod="kube-system/coredns-6f6b679f8f-jwjkr" Mar 19 11:57:36.228372 kubelet[2805]: I0319 11:57:36.228277 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9b99v\" (UniqueName: \"kubernetes.io/projected/5fb07861-4622-4fa3-b2c6-8fc002ee237c-kube-api-access-9b99v\") pod \"coredns-6f6b679f8f-tv99h\" (UID: \"5fb07861-4622-4fa3-b2c6-8fc002ee237c\") " pod="kube-system/coredns-6f6b679f8f-tv99h" Mar 19 11:57:36.228372 kubelet[2805]: I0319 11:57:36.228299 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5fb07861-4622-4fa3-b2c6-8fc002ee237c-config-volume\") pod \"coredns-6f6b679f8f-tv99h\" (UID: \"5fb07861-4622-4fa3-b2c6-8fc002ee237c\") " pod="kube-system/coredns-6f6b679f8f-tv99h" Mar 19 11:57:36.228372 kubelet[2805]: I0319 11:57:36.228313 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-df56f\" (UniqueName: \"kubernetes.io/projected/ab6099c0-0f0b-4131-9d0a-2e2b7559e303-kube-api-access-df56f\") pod \"coredns-6f6b679f8f-jwjkr\" (UID: \"ab6099c0-0f0b-4131-9d0a-2e2b7559e303\") " pod="kube-system/coredns-6f6b679f8f-jwjkr" Mar 19 11:57:36.895337 kubelet[2805]: I0319 11:57:36.894964 2805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-fx77j" podStartSLOduration=7.312774608 podStartE2EDuration="22.894944475s" podCreationTimestamp="2025-03-19 11:57:14 +0000 UTC" firstStartedPulling="2025-03-19 11:57:16.113535068 +0000 UTC m=+8.502729574" lastFinishedPulling="2025-03-19 11:57:31.695704966 +0000 UTC m=+24.084899441" observedRunningTime="2025-03-19 11:57:36.893372801 +0000 UTC m=+29.282567277" watchObservedRunningTime="2025-03-19 11:57:36.894944475 +0000 UTC m=+29.284138961" Mar 19 11:57:37.379842 containerd[1531]: time="2025-03-19T11:57:37.379745716Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-tv99h,Uid:5fb07861-4622-4fa3-b2c6-8fc002ee237c,Namespace:kube-system,Attempt:0,}" Mar 19 11:57:37.400018 containerd[1531]: time="2025-03-19T11:57:37.399575393Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-jwjkr,Uid:ab6099c0-0f0b-4131-9d0a-2e2b7559e303,Namespace:kube-system,Attempt:0,}" Mar 19 11:57:38.107076 systemd-networkd[1415]: cilium_host: Link UP Mar 19 11:57:38.107290 systemd-networkd[1415]: cilium_net: Link UP Mar 19 11:57:38.107548 systemd-networkd[1415]: cilium_net: Gained carrier Mar 19 11:57:38.107769 systemd-networkd[1415]: cilium_host: Gained carrier Mar 19 11:57:38.107936 systemd-networkd[1415]: cilium_net: Gained IPv6LL Mar 19 11:57:38.109168 systemd-networkd[1415]: cilium_host: Gained IPv6LL Mar 19 11:57:38.232905 systemd-networkd[1415]: cilium_vxlan: Link UP Mar 19 11:57:38.232917 systemd-networkd[1415]: cilium_vxlan: Gained carrier Mar 19 11:57:38.632958 kernel: NET: Registered PF_ALG protocol family Mar 19 11:57:39.386750 systemd-networkd[1415]: lxc_health: Link UP Mar 19 11:57:39.395649 systemd-networkd[1415]: lxc_health: Gained carrier Mar 19 11:57:39.424028 systemd-networkd[1415]: cilium_vxlan: Gained IPv6LL Mar 19 11:57:40.004091 systemd-networkd[1415]: lxc541a2cca4d7e: Link UP Mar 19 11:57:40.016904 kernel: eth0: renamed from tmp3e43f Mar 19 11:57:40.028760 kernel: eth0: renamed from tmp2dacb Mar 19 11:57:40.035829 systemd-networkd[1415]: lxce7bfbacf70db: Link UP Mar 19 11:57:40.037563 systemd-networkd[1415]: lxc541a2cca4d7e: Gained carrier Mar 19 11:57:40.040632 systemd-networkd[1415]: lxce7bfbacf70db: Gained carrier Mar 19 11:57:40.448033 systemd-networkd[1415]: lxc_health: Gained IPv6LL Mar 19 11:57:41.792724 systemd-networkd[1415]: lxc541a2cca4d7e: Gained IPv6LL Mar 19 11:57:41.920131 systemd-networkd[1415]: lxce7bfbacf70db: Gained IPv6LL Mar 19 11:57:44.913017 containerd[1531]: time="2025-03-19T11:57:44.912427242Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 19 11:57:44.913017 containerd[1531]: time="2025-03-19T11:57:44.912514446Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 19 11:57:44.913017 containerd[1531]: time="2025-03-19T11:57:44.912533110Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 19 11:57:44.913017 containerd[1531]: time="2025-03-19T11:57:44.912632697Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 19 11:57:44.921978 containerd[1531]: time="2025-03-19T11:57:44.917572600Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 19 11:57:44.921978 containerd[1531]: time="2025-03-19T11:57:44.917642831Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 19 11:57:44.921978 containerd[1531]: time="2025-03-19T11:57:44.917661747Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 19 11:57:44.921978 containerd[1531]: time="2025-03-19T11:57:44.917752767Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 19 11:57:44.967322 systemd[1]: Started cri-containerd-3e43f10113dbc3608d6f270e0494122b3b01bd37950293dd2045fe8eedcf2d60.scope - libcontainer container 3e43f10113dbc3608d6f270e0494122b3b01bd37950293dd2045fe8eedcf2d60. Mar 19 11:57:44.987463 systemd[1]: Started cri-containerd-2dacb7ece9068ce8f00979bd781452e94588a7302d2ed74242f599280a333192.scope - libcontainer container 2dacb7ece9068ce8f00979bd781452e94588a7302d2ed74242f599280a333192. Mar 19 11:57:45.093358 containerd[1531]: time="2025-03-19T11:57:45.091481575Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-tv99h,Uid:5fb07861-4622-4fa3-b2c6-8fc002ee237c,Namespace:kube-system,Attempt:0,} returns sandbox id \"3e43f10113dbc3608d6f270e0494122b3b01bd37950293dd2045fe8eedcf2d60\"" Mar 19 11:57:45.094692 containerd[1531]: time="2025-03-19T11:57:45.094375998Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-jwjkr,Uid:ab6099c0-0f0b-4131-9d0a-2e2b7559e303,Namespace:kube-system,Attempt:0,} returns sandbox id \"2dacb7ece9068ce8f00979bd781452e94588a7302d2ed74242f599280a333192\"" Mar 19 11:57:45.099485 containerd[1531]: time="2025-03-19T11:57:45.098957550Z" level=info msg="CreateContainer within sandbox \"3e43f10113dbc3608d6f270e0494122b3b01bd37950293dd2045fe8eedcf2d60\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Mar 19 11:57:45.099485 containerd[1531]: time="2025-03-19T11:57:45.099373549Z" level=info msg="CreateContainer within sandbox \"2dacb7ece9068ce8f00979bd781452e94588a7302d2ed74242f599280a333192\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Mar 19 11:57:45.123617 containerd[1531]: time="2025-03-19T11:57:45.123569273Z" level=info msg="CreateContainer within sandbox \"3e43f10113dbc3608d6f270e0494122b3b01bd37950293dd2045fe8eedcf2d60\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"b28430bf8800bf69a591ea50b1ab22a17575bb78ffcb1966a002a8d0d0d35463\"" Mar 19 11:57:45.125323 containerd[1531]: time="2025-03-19T11:57:45.124835365Z" level=info msg="StartContainer for \"b28430bf8800bf69a591ea50b1ab22a17575bb78ffcb1966a002a8d0d0d35463\"" Mar 19 11:57:45.126000 containerd[1531]: time="2025-03-19T11:57:45.125263647Z" level=info msg="CreateContainer within sandbox \"2dacb7ece9068ce8f00979bd781452e94588a7302d2ed74242f599280a333192\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"aeb93578bc827c74ab5516e5ca28069f8e9d9fa4d489d9fe6fd6b87cb9bfd218\"" Mar 19 11:57:45.126581 containerd[1531]: time="2025-03-19T11:57:45.126545909Z" level=info msg="StartContainer for \"aeb93578bc827c74ab5516e5ca28069f8e9d9fa4d489d9fe6fd6b87cb9bfd218\"" Mar 19 11:57:45.161020 systemd[1]: Started cri-containerd-aeb93578bc827c74ab5516e5ca28069f8e9d9fa4d489d9fe6fd6b87cb9bfd218.scope - libcontainer container aeb93578bc827c74ab5516e5ca28069f8e9d9fa4d489d9fe6fd6b87cb9bfd218. Mar 19 11:57:45.161988 systemd[1]: Started cri-containerd-b28430bf8800bf69a591ea50b1ab22a17575bb78ffcb1966a002a8d0d0d35463.scope - libcontainer container b28430bf8800bf69a591ea50b1ab22a17575bb78ffcb1966a002a8d0d0d35463. Mar 19 11:57:45.198589 containerd[1531]: time="2025-03-19T11:57:45.198024845Z" level=info msg="StartContainer for \"aeb93578bc827c74ab5516e5ca28069f8e9d9fa4d489d9fe6fd6b87cb9bfd218\" returns successfully" Mar 19 11:57:45.201051 containerd[1531]: time="2025-03-19T11:57:45.200990610Z" level=info msg="StartContainer for \"b28430bf8800bf69a591ea50b1ab22a17575bb78ffcb1966a002a8d0d0d35463\" returns successfully" Mar 19 11:57:45.907615 kubelet[2805]: I0319 11:57:45.906791 2805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-6f6b679f8f-jwjkr" podStartSLOduration=31.906754949 podStartE2EDuration="31.906754949s" podCreationTimestamp="2025-03-19 11:57:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-19 11:57:45.906474624 +0000 UTC m=+38.295669110" watchObservedRunningTime="2025-03-19 11:57:45.906754949 +0000 UTC m=+38.295949435" Mar 19 11:57:45.932109 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2618863042.mount: Deactivated successfully. Mar 19 11:57:45.948732 kubelet[2805]: I0319 11:57:45.948039 2805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-6f6b679f8f-tv99h" podStartSLOduration=31.948021811 podStartE2EDuration="31.948021811s" podCreationTimestamp="2025-03-19 11:57:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-19 11:57:45.927645972 +0000 UTC m=+38.316840469" watchObservedRunningTime="2025-03-19 11:57:45.948021811 +0000 UTC m=+38.337216286" Mar 19 12:01:47.916722 systemd[1]: Started sshd@7-37.27.246.226:22-139.178.68.195:38154.service - OpenSSH per-connection server daemon (139.178.68.195:38154). Mar 19 12:01:48.928473 sshd[4210]: Accepted publickey for core from 139.178.68.195 port 38154 ssh2: RSA SHA256:mZbisEv4EcLuWBRjDimak4H4E9nhJ1zN2E80H6lgS54 Mar 19 12:01:48.930705 sshd-session[4210]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 19 12:01:48.938051 systemd-logind[1502]: New session 8 of user core. Mar 19 12:01:48.950176 systemd[1]: Started session-8.scope - Session 8 of User core. Mar 19 12:01:50.235712 sshd[4212]: Connection closed by 139.178.68.195 port 38154 Mar 19 12:01:50.236812 sshd-session[4210]: pam_unix(sshd:session): session closed for user core Mar 19 12:01:50.241724 systemd[1]: sshd@7-37.27.246.226:22-139.178.68.195:38154.service: Deactivated successfully. Mar 19 12:01:50.244779 systemd[1]: session-8.scope: Deactivated successfully. Mar 19 12:01:50.246445 systemd-logind[1502]: Session 8 logged out. Waiting for processes to exit. Mar 19 12:01:50.248226 systemd-logind[1502]: Removed session 8. Mar 19 12:01:55.412284 systemd[1]: Started sshd@8-37.27.246.226:22-139.178.68.195:38160.service - OpenSSH per-connection server daemon (139.178.68.195:38160). Mar 19 12:01:56.388350 sshd[4225]: Accepted publickey for core from 139.178.68.195 port 38160 ssh2: RSA SHA256:mZbisEv4EcLuWBRjDimak4H4E9nhJ1zN2E80H6lgS54 Mar 19 12:01:56.390063 sshd-session[4225]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 19 12:01:56.397230 systemd-logind[1502]: New session 9 of user core. Mar 19 12:01:56.403153 systemd[1]: Started session-9.scope - Session 9 of User core. Mar 19 12:01:57.166127 sshd[4227]: Connection closed by 139.178.68.195 port 38160 Mar 19 12:01:57.167329 sshd-session[4225]: pam_unix(sshd:session): session closed for user core Mar 19 12:01:57.173499 systemd[1]: sshd@8-37.27.246.226:22-139.178.68.195:38160.service: Deactivated successfully. Mar 19 12:01:57.177957 systemd[1]: session-9.scope: Deactivated successfully. Mar 19 12:01:57.180407 systemd-logind[1502]: Session 9 logged out. Waiting for processes to exit. Mar 19 12:01:57.182807 systemd-logind[1502]: Removed session 9. Mar 19 12:02:02.338340 systemd[1]: Started sshd@9-37.27.246.226:22-139.178.68.195:43872.service - OpenSSH per-connection server daemon (139.178.68.195:43872). Mar 19 12:02:03.309614 sshd[4240]: Accepted publickey for core from 139.178.68.195 port 43872 ssh2: RSA SHA256:mZbisEv4EcLuWBRjDimak4H4E9nhJ1zN2E80H6lgS54 Mar 19 12:02:03.311649 sshd-session[4240]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 19 12:02:03.318166 systemd-logind[1502]: New session 10 of user core. Mar 19 12:02:03.324201 systemd[1]: Started session-10.scope - Session 10 of User core. Mar 19 12:02:04.055452 sshd[4242]: Connection closed by 139.178.68.195 port 43872 Mar 19 12:02:04.056240 sshd-session[4240]: pam_unix(sshd:session): session closed for user core Mar 19 12:02:04.059792 systemd[1]: sshd@9-37.27.246.226:22-139.178.68.195:43872.service: Deactivated successfully. Mar 19 12:02:04.062258 systemd[1]: session-10.scope: Deactivated successfully. Mar 19 12:02:04.064437 systemd-logind[1502]: Session 10 logged out. Waiting for processes to exit. Mar 19 12:02:04.066208 systemd-logind[1502]: Removed session 10. Mar 19 12:02:04.266313 systemd[1]: Started sshd@10-37.27.246.226:22-139.178.68.195:43884.service - OpenSSH per-connection server daemon (139.178.68.195:43884). Mar 19 12:02:05.338137 sshd[4255]: Accepted publickey for core from 139.178.68.195 port 43884 ssh2: RSA SHA256:mZbisEv4EcLuWBRjDimak4H4E9nhJ1zN2E80H6lgS54 Mar 19 12:02:05.340459 sshd-session[4255]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 19 12:02:05.347992 systemd-logind[1502]: New session 11 of user core. Mar 19 12:02:05.357140 systemd[1]: Started session-11.scope - Session 11 of User core. Mar 19 12:02:06.255674 sshd[4257]: Connection closed by 139.178.68.195 port 43884 Mar 19 12:02:06.257461 sshd-session[4255]: pam_unix(sshd:session): session closed for user core Mar 19 12:02:06.264723 systemd[1]: sshd@10-37.27.246.226:22-139.178.68.195:43884.service: Deactivated successfully. Mar 19 12:02:06.265123 systemd-logind[1502]: Session 11 logged out. Waiting for processes to exit. Mar 19 12:02:06.268246 systemd[1]: session-11.scope: Deactivated successfully. Mar 19 12:02:06.270328 systemd-logind[1502]: Removed session 11. Mar 19 12:02:06.426503 systemd[1]: Started sshd@11-37.27.246.226:22-139.178.68.195:51570.service - OpenSSH per-connection server daemon (139.178.68.195:51570). Mar 19 12:02:07.446521 sshd[4268]: Accepted publickey for core from 139.178.68.195 port 51570 ssh2: RSA SHA256:mZbisEv4EcLuWBRjDimak4H4E9nhJ1zN2E80H6lgS54 Mar 19 12:02:07.448352 sshd-session[4268]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 19 12:02:07.453935 systemd-logind[1502]: New session 12 of user core. Mar 19 12:02:07.460181 systemd[1]: Started session-12.scope - Session 12 of User core. Mar 19 12:02:08.252421 sshd[4270]: Connection closed by 139.178.68.195 port 51570 Mar 19 12:02:08.253096 sshd-session[4268]: pam_unix(sshd:session): session closed for user core Mar 19 12:02:08.256899 systemd-logind[1502]: Session 12 logged out. Waiting for processes to exit. Mar 19 12:02:08.257601 systemd[1]: sshd@11-37.27.246.226:22-139.178.68.195:51570.service: Deactivated successfully. Mar 19 12:02:08.259838 systemd[1]: session-12.scope: Deactivated successfully. Mar 19 12:02:08.261588 systemd-logind[1502]: Removed session 12. Mar 19 12:02:13.427388 systemd[1]: Started sshd@12-37.27.246.226:22-139.178.68.195:51584.service - OpenSSH per-connection server daemon (139.178.68.195:51584). Mar 19 12:02:14.402405 sshd[4283]: Accepted publickey for core from 139.178.68.195 port 51584 ssh2: RSA SHA256:mZbisEv4EcLuWBRjDimak4H4E9nhJ1zN2E80H6lgS54 Mar 19 12:02:14.404143 sshd-session[4283]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 19 12:02:14.410706 systemd-logind[1502]: New session 13 of user core. Mar 19 12:02:14.420292 systemd[1]: Started session-13.scope - Session 13 of User core. Mar 19 12:02:15.148122 sshd[4285]: Connection closed by 139.178.68.195 port 51584 Mar 19 12:02:15.148990 sshd-session[4283]: pam_unix(sshd:session): session closed for user core Mar 19 12:02:15.154324 systemd-logind[1502]: Session 13 logged out. Waiting for processes to exit. Mar 19 12:02:15.154601 systemd[1]: sshd@12-37.27.246.226:22-139.178.68.195:51584.service: Deactivated successfully. Mar 19 12:02:15.157808 systemd[1]: session-13.scope: Deactivated successfully. Mar 19 12:02:15.159741 systemd-logind[1502]: Removed session 13. Mar 19 12:02:20.349021 systemd[1]: Started sshd@13-37.27.246.226:22-139.178.68.195:42652.service - OpenSSH per-connection server daemon (139.178.68.195:42652). Mar 19 12:02:21.426513 sshd[4299]: Accepted publickey for core from 139.178.68.195 port 42652 ssh2: RSA SHA256:mZbisEv4EcLuWBRjDimak4H4E9nhJ1zN2E80H6lgS54 Mar 19 12:02:21.428197 sshd-session[4299]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 19 12:02:21.434179 systemd-logind[1502]: New session 14 of user core. Mar 19 12:02:21.437132 systemd[1]: Started session-14.scope - Session 14 of User core. Mar 19 12:02:22.226606 sshd[4301]: Connection closed by 139.178.68.195 port 42652 Mar 19 12:02:22.227316 sshd-session[4299]: pam_unix(sshd:session): session closed for user core Mar 19 12:02:22.233616 systemd[1]: sshd@13-37.27.246.226:22-139.178.68.195:42652.service: Deactivated successfully. Mar 19 12:02:22.234737 systemd-logind[1502]: Session 14 logged out. Waiting for processes to exit. Mar 19 12:02:22.236510 systemd[1]: session-14.scope: Deactivated successfully. Mar 19 12:02:22.237848 systemd-logind[1502]: Removed session 14. Mar 19 12:02:22.387231 systemd[1]: Started sshd@14-37.27.246.226:22-139.178.68.195:42666.service - OpenSSH per-connection server daemon (139.178.68.195:42666). Mar 19 12:02:23.356860 sshd[4313]: Accepted publickey for core from 139.178.68.195 port 42666 ssh2: RSA SHA256:mZbisEv4EcLuWBRjDimak4H4E9nhJ1zN2E80H6lgS54 Mar 19 12:02:23.358643 sshd-session[4313]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 19 12:02:23.364768 systemd-logind[1502]: New session 15 of user core. Mar 19 12:02:23.372154 systemd[1]: Started session-15.scope - Session 15 of User core. Mar 19 12:02:24.329807 sshd[4315]: Connection closed by 139.178.68.195 port 42666 Mar 19 12:02:24.330629 sshd-session[4313]: pam_unix(sshd:session): session closed for user core Mar 19 12:02:24.336953 systemd[1]: sshd@14-37.27.246.226:22-139.178.68.195:42666.service: Deactivated successfully. Mar 19 12:02:24.339043 systemd[1]: session-15.scope: Deactivated successfully. Mar 19 12:02:24.340614 systemd-logind[1502]: Session 15 logged out. Waiting for processes to exit. Mar 19 12:02:24.342752 systemd-logind[1502]: Removed session 15. Mar 19 12:02:24.504314 systemd[1]: Started sshd@15-37.27.246.226:22-139.178.68.195:42670.service - OpenSSH per-connection server daemon (139.178.68.195:42670). Mar 19 12:02:25.491497 sshd[4325]: Accepted publickey for core from 139.178.68.195 port 42670 ssh2: RSA SHA256:mZbisEv4EcLuWBRjDimak4H4E9nhJ1zN2E80H6lgS54 Mar 19 12:02:25.493300 sshd-session[4325]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 19 12:02:25.499007 systemd-logind[1502]: New session 16 of user core. Mar 19 12:02:25.504133 systemd[1]: Started session-16.scope - Session 16 of User core. Mar 19 12:02:28.057492 sshd[4327]: Connection closed by 139.178.68.195 port 42670 Mar 19 12:02:28.058846 sshd-session[4325]: pam_unix(sshd:session): session closed for user core Mar 19 12:02:28.066023 systemd-logind[1502]: Session 16 logged out. Waiting for processes to exit. Mar 19 12:02:28.066392 systemd[1]: sshd@15-37.27.246.226:22-139.178.68.195:42670.service: Deactivated successfully. Mar 19 12:02:28.068999 systemd[1]: session-16.scope: Deactivated successfully. Mar 19 12:02:28.070322 systemd-logind[1502]: Removed session 16. Mar 19 12:02:28.231369 systemd[1]: Started sshd@16-37.27.246.226:22-139.178.68.195:45926.service - OpenSSH per-connection server daemon (139.178.68.195:45926). Mar 19 12:02:29.210428 sshd[4345]: Accepted publickey for core from 139.178.68.195 port 45926 ssh2: RSA SHA256:mZbisEv4EcLuWBRjDimak4H4E9nhJ1zN2E80H6lgS54 Mar 19 12:02:29.212067 sshd-session[4345]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 19 12:02:29.218028 systemd-logind[1502]: New session 17 of user core. Mar 19 12:02:29.223189 systemd[1]: Started session-17.scope - Session 17 of User core. Mar 19 12:02:30.483193 sshd[4347]: Connection closed by 139.178.68.195 port 45926 Mar 19 12:02:30.483845 sshd-session[4345]: pam_unix(sshd:session): session closed for user core Mar 19 12:02:30.488220 systemd[1]: sshd@16-37.27.246.226:22-139.178.68.195:45926.service: Deactivated successfully. Mar 19 12:02:30.492284 systemd[1]: session-17.scope: Deactivated successfully. Mar 19 12:02:30.495241 systemd-logind[1502]: Session 17 logged out. Waiting for processes to exit. Mar 19 12:02:30.497626 systemd-logind[1502]: Removed session 17. Mar 19 12:02:30.660837 systemd[1]: Started sshd@17-37.27.246.226:22-139.178.68.195:45934.service - OpenSSH per-connection server daemon (139.178.68.195:45934). Mar 19 12:02:31.654029 sshd[4357]: Accepted publickey for core from 139.178.68.195 port 45934 ssh2: RSA SHA256:mZbisEv4EcLuWBRjDimak4H4E9nhJ1zN2E80H6lgS54 Mar 19 12:02:31.656404 sshd-session[4357]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 19 12:02:31.663812 systemd-logind[1502]: New session 18 of user core. Mar 19 12:02:31.670181 systemd[1]: Started session-18.scope - Session 18 of User core. Mar 19 12:02:32.469600 sshd[4359]: Connection closed by 139.178.68.195 port 45934 Mar 19 12:02:32.470508 sshd-session[4357]: pam_unix(sshd:session): session closed for user core Mar 19 12:02:32.475096 systemd-logind[1502]: Session 18 logged out. Waiting for processes to exit. Mar 19 12:02:32.475314 systemd[1]: sshd@17-37.27.246.226:22-139.178.68.195:45934.service: Deactivated successfully. Mar 19 12:02:32.478061 systemd[1]: session-18.scope: Deactivated successfully. Mar 19 12:02:32.479551 systemd-logind[1502]: Removed session 18. Mar 19 12:02:37.650285 systemd[1]: Started sshd@18-37.27.246.226:22-139.178.68.195:55434.service - OpenSSH per-connection server daemon (139.178.68.195:55434). Mar 19 12:02:38.653954 sshd[4375]: Accepted publickey for core from 139.178.68.195 port 55434 ssh2: RSA SHA256:mZbisEv4EcLuWBRjDimak4H4E9nhJ1zN2E80H6lgS54 Mar 19 12:02:38.655440 sshd-session[4375]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 19 12:02:38.661037 systemd-logind[1502]: New session 19 of user core. Mar 19 12:02:38.667154 systemd[1]: Started session-19.scope - Session 19 of User core. Mar 19 12:02:39.431199 sshd[4377]: Connection closed by 139.178.68.195 port 55434 Mar 19 12:02:39.432035 sshd-session[4375]: pam_unix(sshd:session): session closed for user core Mar 19 12:02:39.438456 systemd[1]: sshd@18-37.27.246.226:22-139.178.68.195:55434.service: Deactivated successfully. Mar 19 12:02:39.442161 systemd[1]: session-19.scope: Deactivated successfully. Mar 19 12:02:39.443476 systemd-logind[1502]: Session 19 logged out. Waiting for processes to exit. Mar 19 12:02:39.444819 systemd-logind[1502]: Removed session 19. Mar 19 12:02:44.604299 systemd[1]: Started sshd@19-37.27.246.226:22-139.178.68.195:55446.service - OpenSSH per-connection server daemon (139.178.68.195:55446). Mar 19 12:02:45.588066 sshd[4390]: Accepted publickey for core from 139.178.68.195 port 55446 ssh2: RSA SHA256:mZbisEv4EcLuWBRjDimak4H4E9nhJ1zN2E80H6lgS54 Mar 19 12:02:45.590725 sshd-session[4390]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 19 12:02:45.597786 systemd-logind[1502]: New session 20 of user core. Mar 19 12:02:45.608188 systemd[1]: Started session-20.scope - Session 20 of User core. Mar 19 12:02:46.344965 sshd[4394]: Connection closed by 139.178.68.195 port 55446 Mar 19 12:02:46.345597 sshd-session[4390]: pam_unix(sshd:session): session closed for user core Mar 19 12:02:46.349297 systemd[1]: sshd@19-37.27.246.226:22-139.178.68.195:55446.service: Deactivated successfully. Mar 19 12:02:46.352113 systemd[1]: session-20.scope: Deactivated successfully. Mar 19 12:02:46.353030 systemd-logind[1502]: Session 20 logged out. Waiting for processes to exit. Mar 19 12:02:46.353994 systemd-logind[1502]: Removed session 20. Mar 19 12:02:46.529405 systemd[1]: Started sshd@20-37.27.246.226:22-139.178.68.195:33508.service - OpenSSH per-connection server daemon (139.178.68.195:33508). Mar 19 12:02:47.542626 sshd[4406]: Accepted publickey for core from 139.178.68.195 port 33508 ssh2: RSA SHA256:mZbisEv4EcLuWBRjDimak4H4E9nhJ1zN2E80H6lgS54 Mar 19 12:02:47.544131 sshd-session[4406]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 19 12:02:47.549580 systemd-logind[1502]: New session 21 of user core. Mar 19 12:02:47.552083 systemd[1]: Started session-21.scope - Session 21 of User core. Mar 19 12:02:49.666786 containerd[1531]: time="2025-03-19T12:02:49.665694918Z" level=info msg="StopContainer for \"62602a2adab8d404be04a797fe34a5e26d855a2873ca6e298bd07f4e0488b846\" with timeout 30 (s)" Mar 19 12:02:49.674664 containerd[1531]: time="2025-03-19T12:02:49.673655921Z" level=info msg="Stop container \"62602a2adab8d404be04a797fe34a5e26d855a2873ca6e298bd07f4e0488b846\" with signal terminated" Mar 19 12:02:49.692183 systemd[1]: run-containerd-runc-k8s.io-e6b89623cb51d19964eb97e8eac9a82dd8778cfc1597b6f3525a394c57251513-runc.685Nir.mount: Deactivated successfully. Mar 19 12:02:49.695596 systemd[1]: cri-containerd-62602a2adab8d404be04a797fe34a5e26d855a2873ca6e298bd07f4e0488b846.scope: Deactivated successfully. Mar 19 12:02:49.695909 systemd[1]: cri-containerd-62602a2adab8d404be04a797fe34a5e26d855a2873ca6e298bd07f4e0488b846.scope: Consumed 719ms CPU time, 27.9M memory peak, 2.2M read from disk, 4K written to disk. Mar 19 12:02:49.727594 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-62602a2adab8d404be04a797fe34a5e26d855a2873ca6e298bd07f4e0488b846-rootfs.mount: Deactivated successfully. Mar 19 12:02:49.735988 containerd[1531]: time="2025-03-19T12:02:49.735914615Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Mar 19 12:02:49.737512 containerd[1531]: time="2025-03-19T12:02:49.737466123Z" level=info msg="shim disconnected" id=62602a2adab8d404be04a797fe34a5e26d855a2873ca6e298bd07f4e0488b846 namespace=k8s.io Mar 19 12:02:49.737512 containerd[1531]: time="2025-03-19T12:02:49.737512641Z" level=warning msg="cleaning up after shim disconnected" id=62602a2adab8d404be04a797fe34a5e26d855a2873ca6e298bd07f4e0488b846 namespace=k8s.io Mar 19 12:02:49.737600 containerd[1531]: time="2025-03-19T12:02:49.737521037Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 19 12:02:49.747640 containerd[1531]: time="2025-03-19T12:02:49.747343789Z" level=info msg="StopContainer for \"e6b89623cb51d19964eb97e8eac9a82dd8778cfc1597b6f3525a394c57251513\" with timeout 2 (s)" Mar 19 12:02:49.749159 containerd[1531]: time="2025-03-19T12:02:49.748953026Z" level=info msg="Stop container \"e6b89623cb51d19964eb97e8eac9a82dd8778cfc1597b6f3525a394c57251513\" with signal terminated" Mar 19 12:02:49.761830 systemd-networkd[1415]: lxc_health: Link DOWN Mar 19 12:02:49.761842 systemd-networkd[1415]: lxc_health: Lost carrier Mar 19 12:02:49.775350 containerd[1531]: time="2025-03-19T12:02:49.773582404Z" level=info msg="StopContainer for \"62602a2adab8d404be04a797fe34a5e26d855a2873ca6e298bd07f4e0488b846\" returns successfully" Mar 19 12:02:49.778854 containerd[1531]: time="2025-03-19T12:02:49.776795718Z" level=info msg="StopPodSandbox for \"bb72b6cb42b223cfe8537b613d2b5912f92fab908bf4d332a5d91b06d19e9421\"" Mar 19 12:02:49.787290 containerd[1531]: time="2025-03-19T12:02:49.782236136Z" level=info msg="Container to stop \"62602a2adab8d404be04a797fe34a5e26d855a2873ca6e298bd07f4e0488b846\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 19 12:02:49.790734 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-bb72b6cb42b223cfe8537b613d2b5912f92fab908bf4d332a5d91b06d19e9421-shm.mount: Deactivated successfully. Mar 19 12:02:49.794702 systemd[1]: cri-containerd-e6b89623cb51d19964eb97e8eac9a82dd8778cfc1597b6f3525a394c57251513.scope: Deactivated successfully. Mar 19 12:02:49.795237 systemd[1]: cri-containerd-e6b89623cb51d19964eb97e8eac9a82dd8778cfc1597b6f3525a394c57251513.scope: Consumed 9.529s CPU time, 161.6M memory peak, 40.1M read from disk, 13.3M written to disk. Mar 19 12:02:49.804420 systemd[1]: cri-containerd-bb72b6cb42b223cfe8537b613d2b5912f92fab908bf4d332a5d91b06d19e9421.scope: Deactivated successfully. Mar 19 12:02:49.828524 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e6b89623cb51d19964eb97e8eac9a82dd8778cfc1597b6f3525a394c57251513-rootfs.mount: Deactivated successfully. Mar 19 12:02:49.844738 containerd[1531]: time="2025-03-19T12:02:49.844615706Z" level=info msg="shim disconnected" id=e6b89623cb51d19964eb97e8eac9a82dd8778cfc1597b6f3525a394c57251513 namespace=k8s.io Mar 19 12:02:49.845258 containerd[1531]: time="2025-03-19T12:02:49.844718158Z" level=warning msg="cleaning up after shim disconnected" id=e6b89623cb51d19964eb97e8eac9a82dd8778cfc1597b6f3525a394c57251513 namespace=k8s.io Mar 19 12:02:49.845258 containerd[1531]: time="2025-03-19T12:02:49.845053216Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 19 12:02:49.849058 containerd[1531]: time="2025-03-19T12:02:49.848748383Z" level=info msg="shim disconnected" id=bb72b6cb42b223cfe8537b613d2b5912f92fab908bf4d332a5d91b06d19e9421 namespace=k8s.io Mar 19 12:02:49.849058 containerd[1531]: time="2025-03-19T12:02:49.848974185Z" level=warning msg="cleaning up after shim disconnected" id=bb72b6cb42b223cfe8537b613d2b5912f92fab908bf4d332a5d91b06d19e9421 namespace=k8s.io Mar 19 12:02:49.851644 containerd[1531]: time="2025-03-19T12:02:49.848984756Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 19 12:02:49.865621 containerd[1531]: time="2025-03-19T12:02:49.865575716Z" level=info msg="StopContainer for \"e6b89623cb51d19964eb97e8eac9a82dd8778cfc1597b6f3525a394c57251513\" returns successfully" Mar 19 12:02:49.866314 containerd[1531]: time="2025-03-19T12:02:49.866296476Z" level=info msg="StopPodSandbox for \"e1ac54e4ade1aee1a79f559a223841973bf4d1dbd6882bd882ecd9a1600d4dfa\"" Mar 19 12:02:49.866483 containerd[1531]: time="2025-03-19T12:02:49.866418685Z" level=info msg="Container to stop \"5b31e2b3a79792be2a445f1ccc7411778302d0b2ad34abfd194de17bc0795665\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 19 12:02:49.866561 containerd[1531]: time="2025-03-19T12:02:49.866533130Z" level=info msg="Container to stop \"9745ab02b8e658d81a940624e38afabc8bf95f38df9236674e895fcdcc973f37\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 19 12:02:49.866633 containerd[1531]: time="2025-03-19T12:02:49.866610756Z" level=info msg="Container to stop \"b7c808ce26d368c2639ba65a198b7646021751e63362e4aa0492551ab2280841\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 19 12:02:49.866703 containerd[1531]: time="2025-03-19T12:02:49.866691036Z" level=info msg="Container to stop \"e6b89623cb51d19964eb97e8eac9a82dd8778cfc1597b6f3525a394c57251513\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 19 12:02:49.866790 containerd[1531]: time="2025-03-19T12:02:49.866749466Z" level=info msg="Container to stop \"b4840c728b6ff37e8c43f65799cba5c27b86a145f3d77597741805d13079d723\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 19 12:02:49.868567 containerd[1531]: time="2025-03-19T12:02:49.868547897Z" level=info msg="TearDown network for sandbox \"bb72b6cb42b223cfe8537b613d2b5912f92fab908bf4d332a5d91b06d19e9421\" successfully" Mar 19 12:02:49.868650 containerd[1531]: time="2025-03-19T12:02:49.868637766Z" level=info msg="StopPodSandbox for \"bb72b6cb42b223cfe8537b613d2b5912f92fab908bf4d332a5d91b06d19e9421\" returns successfully" Mar 19 12:02:49.874251 systemd[1]: cri-containerd-e1ac54e4ade1aee1a79f559a223841973bf4d1dbd6882bd882ecd9a1600d4dfa.scope: Deactivated successfully. Mar 19 12:02:49.905744 containerd[1531]: time="2025-03-19T12:02:49.905258060Z" level=info msg="shim disconnected" id=e1ac54e4ade1aee1a79f559a223841973bf4d1dbd6882bd882ecd9a1600d4dfa namespace=k8s.io Mar 19 12:02:49.905744 containerd[1531]: time="2025-03-19T12:02:49.905364380Z" level=warning msg="cleaning up after shim disconnected" id=e1ac54e4ade1aee1a79f559a223841973bf4d1dbd6882bd882ecd9a1600d4dfa namespace=k8s.io Mar 19 12:02:49.905744 containerd[1531]: time="2025-03-19T12:02:49.905375781Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 19 12:02:49.922674 kubelet[2805]: I0319 12:02:49.921505 2805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8kbf\" (UniqueName: \"kubernetes.io/projected/4ad24ea6-768c-404d-9de4-78cdd82836aa-kube-api-access-q8kbf\") pod \"4ad24ea6-768c-404d-9de4-78cdd82836aa\" (UID: \"4ad24ea6-768c-404d-9de4-78cdd82836aa\") " Mar 19 12:02:49.922674 kubelet[2805]: I0319 12:02:49.921567 2805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/4ad24ea6-768c-404d-9de4-78cdd82836aa-cilium-config-path\") pod \"4ad24ea6-768c-404d-9de4-78cdd82836aa\" (UID: \"4ad24ea6-768c-404d-9de4-78cdd82836aa\") " Mar 19 12:02:49.926001 containerd[1531]: time="2025-03-19T12:02:49.925706111Z" level=info msg="TearDown network for sandbox \"e1ac54e4ade1aee1a79f559a223841973bf4d1dbd6882bd882ecd9a1600d4dfa\" successfully" Mar 19 12:02:49.926001 containerd[1531]: time="2025-03-19T12:02:49.925749712Z" level=info msg="StopPodSandbox for \"e1ac54e4ade1aee1a79f559a223841973bf4d1dbd6882bd882ecd9a1600d4dfa\" returns successfully" Mar 19 12:02:49.935583 kubelet[2805]: I0319 12:02:49.933674 2805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ad24ea6-768c-404d-9de4-78cdd82836aa-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "4ad24ea6-768c-404d-9de4-78cdd82836aa" (UID: "4ad24ea6-768c-404d-9de4-78cdd82836aa"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 19 12:02:49.935583 kubelet[2805]: I0319 12:02:49.933695 2805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ad24ea6-768c-404d-9de4-78cdd82836aa-kube-api-access-q8kbf" (OuterVolumeSpecName: "kube-api-access-q8kbf") pod "4ad24ea6-768c-404d-9de4-78cdd82836aa" (UID: "4ad24ea6-768c-404d-9de4-78cdd82836aa"). InnerVolumeSpecName "kube-api-access-q8kbf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 19 12:02:50.022061 kubelet[2805]: I0319 12:02:50.022014 2805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/40347df1-b19f-4102-8759-3c41dc3a468d-clustermesh-secrets\") pod \"40347df1-b19f-4102-8759-3c41dc3a468d\" (UID: \"40347df1-b19f-4102-8759-3c41dc3a468d\") " Mar 19 12:02:50.022487 kubelet[2805]: I0319 12:02:50.022241 2805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dsttr\" (UniqueName: \"kubernetes.io/projected/40347df1-b19f-4102-8759-3c41dc3a468d-kube-api-access-dsttr\") pod \"40347df1-b19f-4102-8759-3c41dc3a468d\" (UID: \"40347df1-b19f-4102-8759-3c41dc3a468d\") " Mar 19 12:02:50.022487 kubelet[2805]: I0319 12:02:50.022261 2805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-cni-path\") pod \"40347df1-b19f-4102-8759-3c41dc3a468d\" (UID: \"40347df1-b19f-4102-8759-3c41dc3a468d\") " Mar 19 12:02:50.022487 kubelet[2805]: I0319 12:02:50.022275 2805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-xtables-lock\") pod \"40347df1-b19f-4102-8759-3c41dc3a468d\" (UID: \"40347df1-b19f-4102-8759-3c41dc3a468d\") " Mar 19 12:02:50.022487 kubelet[2805]: I0319 12:02:50.022287 2805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-etc-cni-netd\") pod \"40347df1-b19f-4102-8759-3c41dc3a468d\" (UID: \"40347df1-b19f-4102-8759-3c41dc3a468d\") " Mar 19 12:02:50.022487 kubelet[2805]: I0319 12:02:50.022303 2805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-cilium-run\") pod \"40347df1-b19f-4102-8759-3c41dc3a468d\" (UID: \"40347df1-b19f-4102-8759-3c41dc3a468d\") " Mar 19 12:02:50.022487 kubelet[2805]: I0319 12:02:50.022315 2805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-host-proc-sys-net\") pod \"40347df1-b19f-4102-8759-3c41dc3a468d\" (UID: \"40347df1-b19f-4102-8759-3c41dc3a468d\") " Mar 19 12:02:50.022684 kubelet[2805]: I0319 12:02:50.022326 2805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-host-proc-sys-kernel\") pod \"40347df1-b19f-4102-8759-3c41dc3a468d\" (UID: \"40347df1-b19f-4102-8759-3c41dc3a468d\") " Mar 19 12:02:50.022684 kubelet[2805]: I0319 12:02:50.022338 2805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-bpf-maps\") pod \"40347df1-b19f-4102-8759-3c41dc3a468d\" (UID: \"40347df1-b19f-4102-8759-3c41dc3a468d\") " Mar 19 12:02:50.022684 kubelet[2805]: I0319 12:02:50.022353 2805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-cilium-cgroup\") pod \"40347df1-b19f-4102-8759-3c41dc3a468d\" (UID: \"40347df1-b19f-4102-8759-3c41dc3a468d\") " Mar 19 12:02:50.022684 kubelet[2805]: I0319 12:02:50.022367 2805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-lib-modules\") pod \"40347df1-b19f-4102-8759-3c41dc3a468d\" (UID: \"40347df1-b19f-4102-8759-3c41dc3a468d\") " Mar 19 12:02:50.022684 kubelet[2805]: I0319 12:02:50.022380 2805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/40347df1-b19f-4102-8759-3c41dc3a468d-cilium-config-path\") pod \"40347df1-b19f-4102-8759-3c41dc3a468d\" (UID: \"40347df1-b19f-4102-8759-3c41dc3a468d\") " Mar 19 12:02:50.022684 kubelet[2805]: I0319 12:02:50.022394 2805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/40347df1-b19f-4102-8759-3c41dc3a468d-hubble-tls\") pod \"40347df1-b19f-4102-8759-3c41dc3a468d\" (UID: \"40347df1-b19f-4102-8759-3c41dc3a468d\") " Mar 19 12:02:50.022836 kubelet[2805]: I0319 12:02:50.022407 2805 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-hostproc\") pod \"40347df1-b19f-4102-8759-3c41dc3a468d\" (UID: \"40347df1-b19f-4102-8759-3c41dc3a468d\") " Mar 19 12:02:50.023219 kubelet[2805]: I0319 12:02:50.023165 2805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "40347df1-b19f-4102-8759-3c41dc3a468d" (UID: "40347df1-b19f-4102-8759-3c41dc3a468d"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 19 12:02:50.025252 kubelet[2805]: I0319 12:02:50.024981 2805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-cni-path" (OuterVolumeSpecName: "cni-path") pod "40347df1-b19f-4102-8759-3c41dc3a468d" (UID: "40347df1-b19f-4102-8759-3c41dc3a468d"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 19 12:02:50.025252 kubelet[2805]: I0319 12:02:50.025014 2805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "40347df1-b19f-4102-8759-3c41dc3a468d" (UID: "40347df1-b19f-4102-8759-3c41dc3a468d"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 19 12:02:50.025252 kubelet[2805]: I0319 12:02:50.025031 2805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "40347df1-b19f-4102-8759-3c41dc3a468d" (UID: "40347df1-b19f-4102-8759-3c41dc3a468d"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 19 12:02:50.025252 kubelet[2805]: I0319 12:02:50.025048 2805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "40347df1-b19f-4102-8759-3c41dc3a468d" (UID: "40347df1-b19f-4102-8759-3c41dc3a468d"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 19 12:02:50.025252 kubelet[2805]: I0319 12:02:50.025102 2805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40347df1-b19f-4102-8759-3c41dc3a468d-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "40347df1-b19f-4102-8759-3c41dc3a468d" (UID: "40347df1-b19f-4102-8759-3c41dc3a468d"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 19 12:02:50.025423 kubelet[2805]: I0319 12:02:50.025126 2805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "40347df1-b19f-4102-8759-3c41dc3a468d" (UID: "40347df1-b19f-4102-8759-3c41dc3a468d"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 19 12:02:50.025423 kubelet[2805]: I0319 12:02:50.025144 2805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "40347df1-b19f-4102-8759-3c41dc3a468d" (UID: "40347df1-b19f-4102-8759-3c41dc3a468d"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 19 12:02:50.025423 kubelet[2805]: I0319 12:02:50.025160 2805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "40347df1-b19f-4102-8759-3c41dc3a468d" (UID: "40347df1-b19f-4102-8759-3c41dc3a468d"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 19 12:02:50.026711 kubelet[2805]: I0319 12:02:50.026345 2805 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-q8kbf\" (UniqueName: \"kubernetes.io/projected/4ad24ea6-768c-404d-9de4-78cdd82836aa-kube-api-access-q8kbf\") on node \"ci-4230-1-0-d-e284590f49\" DevicePath \"\"" Mar 19 12:02:50.026711 kubelet[2805]: I0319 12:02:50.026391 2805 reconciler_common.go:288] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/4ad24ea6-768c-404d-9de4-78cdd82836aa-cilium-config-path\") on node \"ci-4230-1-0-d-e284590f49\" DevicePath \"\"" Mar 19 12:02:50.026711 kubelet[2805]: I0319 12:02:50.026425 2805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-hostproc" (OuterVolumeSpecName: "hostproc") pod "40347df1-b19f-4102-8759-3c41dc3a468d" (UID: "40347df1-b19f-4102-8759-3c41dc3a468d"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 19 12:02:50.027432 kubelet[2805]: I0319 12:02:50.027390 2805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40347df1-b19f-4102-8759-3c41dc3a468d-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "40347df1-b19f-4102-8759-3c41dc3a468d" (UID: "40347df1-b19f-4102-8759-3c41dc3a468d"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 19 12:02:50.027432 kubelet[2805]: I0319 12:02:50.027429 2805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "40347df1-b19f-4102-8759-3c41dc3a468d" (UID: "40347df1-b19f-4102-8759-3c41dc3a468d"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 19 12:02:50.028030 kubelet[2805]: I0319 12:02:50.028002 2805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40347df1-b19f-4102-8759-3c41dc3a468d-kube-api-access-dsttr" (OuterVolumeSpecName: "kube-api-access-dsttr") pod "40347df1-b19f-4102-8759-3c41dc3a468d" (UID: "40347df1-b19f-4102-8759-3c41dc3a468d"). InnerVolumeSpecName "kube-api-access-dsttr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 19 12:02:50.029280 kubelet[2805]: I0319 12:02:50.029247 2805 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40347df1-b19f-4102-8759-3c41dc3a468d-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "40347df1-b19f-4102-8759-3c41dc3a468d" (UID: "40347df1-b19f-4102-8759-3c41dc3a468d"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 19 12:02:50.126912 kubelet[2805]: I0319 12:02:50.126745 2805 reconciler_common.go:288] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/40347df1-b19f-4102-8759-3c41dc3a468d-clustermesh-secrets\") on node \"ci-4230-1-0-d-e284590f49\" DevicePath \"\"" Mar 19 12:02:50.126912 kubelet[2805]: I0319 12:02:50.126811 2805 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-dsttr\" (UniqueName: \"kubernetes.io/projected/40347df1-b19f-4102-8759-3c41dc3a468d-kube-api-access-dsttr\") on node \"ci-4230-1-0-d-e284590f49\" DevicePath \"\"" Mar 19 12:02:50.126912 kubelet[2805]: I0319 12:02:50.126825 2805 reconciler_common.go:288] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-cni-path\") on node \"ci-4230-1-0-d-e284590f49\" DevicePath \"\"" Mar 19 12:02:50.126912 kubelet[2805]: I0319 12:02:50.126839 2805 reconciler_common.go:288] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-host-proc-sys-net\") on node \"ci-4230-1-0-d-e284590f49\" DevicePath \"\"" Mar 19 12:02:50.126912 kubelet[2805]: I0319 12:02:50.126853 2805 reconciler_common.go:288] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-xtables-lock\") on node \"ci-4230-1-0-d-e284590f49\" DevicePath \"\"" Mar 19 12:02:50.126912 kubelet[2805]: I0319 12:02:50.126862 2805 reconciler_common.go:288] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-etc-cni-netd\") on node \"ci-4230-1-0-d-e284590f49\" DevicePath \"\"" Mar 19 12:02:50.126912 kubelet[2805]: I0319 12:02:50.126907 2805 reconciler_common.go:288] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-cilium-run\") on node \"ci-4230-1-0-d-e284590f49\" DevicePath \"\"" Mar 19 12:02:50.126912 kubelet[2805]: I0319 12:02:50.126917 2805 reconciler_common.go:288] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-host-proc-sys-kernel\") on node \"ci-4230-1-0-d-e284590f49\" DevicePath \"\"" Mar 19 12:02:50.127443 kubelet[2805]: I0319 12:02:50.126928 2805 reconciler_common.go:288] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-bpf-maps\") on node \"ci-4230-1-0-d-e284590f49\" DevicePath \"\"" Mar 19 12:02:50.127443 kubelet[2805]: I0319 12:02:50.126937 2805 reconciler_common.go:288] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-cilium-cgroup\") on node \"ci-4230-1-0-d-e284590f49\" DevicePath \"\"" Mar 19 12:02:50.127443 kubelet[2805]: I0319 12:02:50.126946 2805 reconciler_common.go:288] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/40347df1-b19f-4102-8759-3c41dc3a468d-hubble-tls\") on node \"ci-4230-1-0-d-e284590f49\" DevicePath \"\"" Mar 19 12:02:50.127443 kubelet[2805]: I0319 12:02:50.126955 2805 reconciler_common.go:288] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-lib-modules\") on node \"ci-4230-1-0-d-e284590f49\" DevicePath \"\"" Mar 19 12:02:50.127443 kubelet[2805]: I0319 12:02:50.126964 2805 reconciler_common.go:288] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/40347df1-b19f-4102-8759-3c41dc3a468d-cilium-config-path\") on node \"ci-4230-1-0-d-e284590f49\" DevicePath \"\"" Mar 19 12:02:50.127443 kubelet[2805]: I0319 12:02:50.126973 2805 reconciler_common.go:288] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/40347df1-b19f-4102-8759-3c41dc3a468d-hostproc\") on node \"ci-4230-1-0-d-e284590f49\" DevicePath \"\"" Mar 19 12:02:50.553698 systemd[1]: Removed slice kubepods-besteffort-pod4ad24ea6_768c_404d_9de4_78cdd82836aa.slice - libcontainer container kubepods-besteffort-pod4ad24ea6_768c_404d_9de4_78cdd82836aa.slice. Mar 19 12:02:50.554822 systemd[1]: kubepods-besteffort-pod4ad24ea6_768c_404d_9de4_78cdd82836aa.slice: Consumed 752ms CPU time, 28.2M memory peak, 2.2M read from disk, 4K written to disk. Mar 19 12:02:50.574935 kubelet[2805]: I0319 12:02:50.573493 2805 scope.go:117] "RemoveContainer" containerID="62602a2adab8d404be04a797fe34a5e26d855a2873ca6e298bd07f4e0488b846" Mar 19 12:02:50.603241 containerd[1531]: time="2025-03-19T12:02:50.602472893Z" level=info msg="RemoveContainer for \"62602a2adab8d404be04a797fe34a5e26d855a2873ca6e298bd07f4e0488b846\"" Mar 19 12:02:50.608497 containerd[1531]: time="2025-03-19T12:02:50.608434549Z" level=info msg="RemoveContainer for \"62602a2adab8d404be04a797fe34a5e26d855a2873ca6e298bd07f4e0488b846\" returns successfully" Mar 19 12:02:50.611037 systemd[1]: Removed slice kubepods-burstable-pod40347df1_b19f_4102_8759_3c41dc3a468d.slice - libcontainer container kubepods-burstable-pod40347df1_b19f_4102_8759_3c41dc3a468d.slice. Mar 19 12:02:50.611188 systemd[1]: kubepods-burstable-pod40347df1_b19f_4102_8759_3c41dc3a468d.slice: Consumed 9.626s CPU time, 161.9M memory peak, 41.1M read from disk, 13.3M written to disk. Mar 19 12:02:50.615734 kubelet[2805]: I0319 12:02:50.615695 2805 scope.go:117] "RemoveContainer" containerID="62602a2adab8d404be04a797fe34a5e26d855a2873ca6e298bd07f4e0488b846" Mar 19 12:02:50.630751 containerd[1531]: time="2025-03-19T12:02:50.630600648Z" level=error msg="ContainerStatus for \"62602a2adab8d404be04a797fe34a5e26d855a2873ca6e298bd07f4e0488b846\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"62602a2adab8d404be04a797fe34a5e26d855a2873ca6e298bd07f4e0488b846\": not found" Mar 19 12:02:50.634074 kubelet[2805]: E0319 12:02:50.633212 2805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"62602a2adab8d404be04a797fe34a5e26d855a2873ca6e298bd07f4e0488b846\": not found" containerID="62602a2adab8d404be04a797fe34a5e26d855a2873ca6e298bd07f4e0488b846" Mar 19 12:02:50.634074 kubelet[2805]: I0319 12:02:50.633364 2805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"62602a2adab8d404be04a797fe34a5e26d855a2873ca6e298bd07f4e0488b846"} err="failed to get container status \"62602a2adab8d404be04a797fe34a5e26d855a2873ca6e298bd07f4e0488b846\": rpc error: code = NotFound desc = an error occurred when try to find container \"62602a2adab8d404be04a797fe34a5e26d855a2873ca6e298bd07f4e0488b846\": not found" Mar 19 12:02:50.634074 kubelet[2805]: I0319 12:02:50.633542 2805 scope.go:117] "RemoveContainer" containerID="e6b89623cb51d19964eb97e8eac9a82dd8778cfc1597b6f3525a394c57251513" Mar 19 12:02:50.636021 containerd[1531]: time="2025-03-19T12:02:50.635555977Z" level=info msg="RemoveContainer for \"e6b89623cb51d19964eb97e8eac9a82dd8778cfc1597b6f3525a394c57251513\"" Mar 19 12:02:50.642565 containerd[1531]: time="2025-03-19T12:02:50.642503310Z" level=info msg="RemoveContainer for \"e6b89623cb51d19964eb97e8eac9a82dd8778cfc1597b6f3525a394c57251513\" returns successfully" Mar 19 12:02:50.643050 kubelet[2805]: I0319 12:02:50.643023 2805 scope.go:117] "RemoveContainer" containerID="b7c808ce26d368c2639ba65a198b7646021751e63362e4aa0492551ab2280841" Mar 19 12:02:50.645171 containerd[1531]: time="2025-03-19T12:02:50.644466631Z" level=info msg="RemoveContainer for \"b7c808ce26d368c2639ba65a198b7646021751e63362e4aa0492551ab2280841\"" Mar 19 12:02:50.649649 containerd[1531]: time="2025-03-19T12:02:50.649595865Z" level=info msg="RemoveContainer for \"b7c808ce26d368c2639ba65a198b7646021751e63362e4aa0492551ab2280841\" returns successfully" Mar 19 12:02:50.650175 kubelet[2805]: I0319 12:02:50.650096 2805 scope.go:117] "RemoveContainer" containerID="9745ab02b8e658d81a940624e38afabc8bf95f38df9236674e895fcdcc973f37" Mar 19 12:02:50.651428 containerd[1531]: time="2025-03-19T12:02:50.651401250Z" level=info msg="RemoveContainer for \"9745ab02b8e658d81a940624e38afabc8bf95f38df9236674e895fcdcc973f37\"" Mar 19 12:02:50.658568 containerd[1531]: time="2025-03-19T12:02:50.658364652Z" level=info msg="RemoveContainer for \"9745ab02b8e658d81a940624e38afabc8bf95f38df9236674e895fcdcc973f37\" returns successfully" Mar 19 12:02:50.659015 kubelet[2805]: I0319 12:02:50.658986 2805 scope.go:117] "RemoveContainer" containerID="5b31e2b3a79792be2a445f1ccc7411778302d0b2ad34abfd194de17bc0795665" Mar 19 12:02:50.660443 containerd[1531]: time="2025-03-19T12:02:50.660174405Z" level=info msg="RemoveContainer for \"5b31e2b3a79792be2a445f1ccc7411778302d0b2ad34abfd194de17bc0795665\"" Mar 19 12:02:50.663487 containerd[1531]: time="2025-03-19T12:02:50.663437241Z" level=info msg="RemoveContainer for \"5b31e2b3a79792be2a445f1ccc7411778302d0b2ad34abfd194de17bc0795665\" returns successfully" Mar 19 12:02:50.663770 kubelet[2805]: I0319 12:02:50.663687 2805 scope.go:117] "RemoveContainer" containerID="b4840c728b6ff37e8c43f65799cba5c27b86a145f3d77597741805d13079d723" Mar 19 12:02:50.664801 containerd[1531]: time="2025-03-19T12:02:50.664738912Z" level=info msg="RemoveContainer for \"b4840c728b6ff37e8c43f65799cba5c27b86a145f3d77597741805d13079d723\"" Mar 19 12:02:50.668176 containerd[1531]: time="2025-03-19T12:02:50.668119007Z" level=info msg="RemoveContainer for \"b4840c728b6ff37e8c43f65799cba5c27b86a145f3d77597741805d13079d723\" returns successfully" Mar 19 12:02:50.668558 kubelet[2805]: I0319 12:02:50.668340 2805 scope.go:117] "RemoveContainer" containerID="e6b89623cb51d19964eb97e8eac9a82dd8778cfc1597b6f3525a394c57251513" Mar 19 12:02:50.668602 containerd[1531]: time="2025-03-19T12:02:50.668572077Z" level=error msg="ContainerStatus for \"e6b89623cb51d19964eb97e8eac9a82dd8778cfc1597b6f3525a394c57251513\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"e6b89623cb51d19964eb97e8eac9a82dd8778cfc1597b6f3525a394c57251513\": not found" Mar 19 12:02:50.668733 kubelet[2805]: E0319 12:02:50.668684 2805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"e6b89623cb51d19964eb97e8eac9a82dd8778cfc1597b6f3525a394c57251513\": not found" containerID="e6b89623cb51d19964eb97e8eac9a82dd8778cfc1597b6f3525a394c57251513" Mar 19 12:02:50.668902 kubelet[2805]: I0319 12:02:50.668731 2805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"e6b89623cb51d19964eb97e8eac9a82dd8778cfc1597b6f3525a394c57251513"} err="failed to get container status \"e6b89623cb51d19964eb97e8eac9a82dd8778cfc1597b6f3525a394c57251513\": rpc error: code = NotFound desc = an error occurred when try to find container \"e6b89623cb51d19964eb97e8eac9a82dd8778cfc1597b6f3525a394c57251513\": not found" Mar 19 12:02:50.668902 kubelet[2805]: I0319 12:02:50.668757 2805 scope.go:117] "RemoveContainer" containerID="b7c808ce26d368c2639ba65a198b7646021751e63362e4aa0492551ab2280841" Mar 19 12:02:50.669033 kubelet[2805]: E0319 12:02:50.669017 2805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"b7c808ce26d368c2639ba65a198b7646021751e63362e4aa0492551ab2280841\": not found" containerID="b7c808ce26d368c2639ba65a198b7646021751e63362e4aa0492551ab2280841" Mar 19 12:02:50.669090 containerd[1531]: time="2025-03-19T12:02:50.668918235Z" level=error msg="ContainerStatus for \"b7c808ce26d368c2639ba65a198b7646021751e63362e4aa0492551ab2280841\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"b7c808ce26d368c2639ba65a198b7646021751e63362e4aa0492551ab2280841\": not found" Mar 19 12:02:50.669160 kubelet[2805]: I0319 12:02:50.669032 2805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"b7c808ce26d368c2639ba65a198b7646021751e63362e4aa0492551ab2280841"} err="failed to get container status \"b7c808ce26d368c2639ba65a198b7646021751e63362e4aa0492551ab2280841\": rpc error: code = NotFound desc = an error occurred when try to find container \"b7c808ce26d368c2639ba65a198b7646021751e63362e4aa0492551ab2280841\": not found" Mar 19 12:02:50.669160 kubelet[2805]: I0319 12:02:50.669045 2805 scope.go:117] "RemoveContainer" containerID="9745ab02b8e658d81a940624e38afabc8bf95f38df9236674e895fcdcc973f37" Mar 19 12:02:50.669267 containerd[1531]: time="2025-03-19T12:02:50.669144650Z" level=error msg="ContainerStatus for \"9745ab02b8e658d81a940624e38afabc8bf95f38df9236674e895fcdcc973f37\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"9745ab02b8e658d81a940624e38afabc8bf95f38df9236674e895fcdcc973f37\": not found" Mar 19 12:02:50.669301 kubelet[2805]: E0319 12:02:50.669231 2805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"9745ab02b8e658d81a940624e38afabc8bf95f38df9236674e895fcdcc973f37\": not found" containerID="9745ab02b8e658d81a940624e38afabc8bf95f38df9236674e895fcdcc973f37" Mar 19 12:02:50.669301 kubelet[2805]: I0319 12:02:50.669246 2805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"9745ab02b8e658d81a940624e38afabc8bf95f38df9236674e895fcdcc973f37"} err="failed to get container status \"9745ab02b8e658d81a940624e38afabc8bf95f38df9236674e895fcdcc973f37\": rpc error: code = NotFound desc = an error occurred when try to find container \"9745ab02b8e658d81a940624e38afabc8bf95f38df9236674e895fcdcc973f37\": not found" Mar 19 12:02:50.669301 kubelet[2805]: I0319 12:02:50.669256 2805 scope.go:117] "RemoveContainer" containerID="5b31e2b3a79792be2a445f1ccc7411778302d0b2ad34abfd194de17bc0795665" Mar 19 12:02:50.669503 containerd[1531]: time="2025-03-19T12:02:50.669440834Z" level=error msg="ContainerStatus for \"5b31e2b3a79792be2a445f1ccc7411778302d0b2ad34abfd194de17bc0795665\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"5b31e2b3a79792be2a445f1ccc7411778302d0b2ad34abfd194de17bc0795665\": not found" Mar 19 12:02:50.669544 kubelet[2805]: E0319 12:02:50.669527 2805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"5b31e2b3a79792be2a445f1ccc7411778302d0b2ad34abfd194de17bc0795665\": not found" containerID="5b31e2b3a79792be2a445f1ccc7411778302d0b2ad34abfd194de17bc0795665" Mar 19 12:02:50.669584 kubelet[2805]: I0319 12:02:50.669541 2805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"5b31e2b3a79792be2a445f1ccc7411778302d0b2ad34abfd194de17bc0795665"} err="failed to get container status \"5b31e2b3a79792be2a445f1ccc7411778302d0b2ad34abfd194de17bc0795665\": rpc error: code = NotFound desc = an error occurred when try to find container \"5b31e2b3a79792be2a445f1ccc7411778302d0b2ad34abfd194de17bc0795665\": not found" Mar 19 12:02:50.669584 kubelet[2805]: I0319 12:02:50.669551 2805 scope.go:117] "RemoveContainer" containerID="b4840c728b6ff37e8c43f65799cba5c27b86a145f3d77597741805d13079d723" Mar 19 12:02:50.669666 containerd[1531]: time="2025-03-19T12:02:50.669651600Z" level=error msg="ContainerStatus for \"b4840c728b6ff37e8c43f65799cba5c27b86a145f3d77597741805d13079d723\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"b4840c728b6ff37e8c43f65799cba5c27b86a145f3d77597741805d13079d723\": not found" Mar 19 12:02:50.669732 kubelet[2805]: E0319 12:02:50.669720 2805 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"b4840c728b6ff37e8c43f65799cba5c27b86a145f3d77597741805d13079d723\": not found" containerID="b4840c728b6ff37e8c43f65799cba5c27b86a145f3d77597741805d13079d723" Mar 19 12:02:50.669764 kubelet[2805]: I0319 12:02:50.669737 2805 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"b4840c728b6ff37e8c43f65799cba5c27b86a145f3d77597741805d13079d723"} err="failed to get container status \"b4840c728b6ff37e8c43f65799cba5c27b86a145f3d77597741805d13079d723\": rpc error: code = NotFound desc = an error occurred when try to find container \"b4840c728b6ff37e8c43f65799cba5c27b86a145f3d77597741805d13079d723\": not found" Mar 19 12:02:50.687851 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e1ac54e4ade1aee1a79f559a223841973bf4d1dbd6882bd882ecd9a1600d4dfa-rootfs.mount: Deactivated successfully. Mar 19 12:02:50.687964 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-e1ac54e4ade1aee1a79f559a223841973bf4d1dbd6882bd882ecd9a1600d4dfa-shm.mount: Deactivated successfully. Mar 19 12:02:50.688020 systemd[1]: var-lib-kubelet-pods-40347df1\x2db19f\x2d4102\x2d8759\x2d3c41dc3a468d-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Mar 19 12:02:50.688079 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-bb72b6cb42b223cfe8537b613d2b5912f92fab908bf4d332a5d91b06d19e9421-rootfs.mount: Deactivated successfully. Mar 19 12:02:50.688130 systemd[1]: var-lib-kubelet-pods-40347df1\x2db19f\x2d4102\x2d8759\x2d3c41dc3a468d-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Mar 19 12:02:50.688181 systemd[1]: var-lib-kubelet-pods-4ad24ea6\x2d768c\x2d404d\x2d9de4\x2d78cdd82836aa-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dq8kbf.mount: Deactivated successfully. Mar 19 12:02:50.688236 systemd[1]: var-lib-kubelet-pods-40347df1\x2db19f\x2d4102\x2d8759\x2d3c41dc3a468d-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2ddsttr.mount: Deactivated successfully. Mar 19 12:02:51.685520 sshd[4408]: Connection closed by 139.178.68.195 port 33508 Mar 19 12:02:51.686632 sshd-session[4406]: pam_unix(sshd:session): session closed for user core Mar 19 12:02:51.689742 systemd[1]: sshd@20-37.27.246.226:22-139.178.68.195:33508.service: Deactivated successfully. Mar 19 12:02:51.691434 systemd[1]: session-21.scope: Deactivated successfully. Mar 19 12:02:51.693324 systemd-logind[1502]: Session 21 logged out. Waiting for processes to exit. Mar 19 12:02:51.695025 systemd-logind[1502]: Removed session 21. Mar 19 12:02:51.732897 kubelet[2805]: I0319 12:02:51.731580 2805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40347df1-b19f-4102-8759-3c41dc3a468d" path="/var/lib/kubelet/pods/40347df1-b19f-4102-8759-3c41dc3a468d/volumes" Mar 19 12:02:51.732897 kubelet[2805]: I0319 12:02:51.732160 2805 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ad24ea6-768c-404d-9de4-78cdd82836aa" path="/var/lib/kubelet/pods/4ad24ea6-768c-404d-9de4-78cdd82836aa/volumes" Mar 19 12:02:51.856352 systemd[1]: Started sshd@21-37.27.246.226:22-139.178.68.195:33522.service - OpenSSH per-connection server daemon (139.178.68.195:33522). Mar 19 12:02:52.844937 sshd[4568]: Accepted publickey for core from 139.178.68.195 port 33522 ssh2: RSA SHA256:mZbisEv4EcLuWBRjDimak4H4E9nhJ1zN2E80H6lgS54 Mar 19 12:02:52.846635 sshd-session[4568]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 19 12:02:52.851770 systemd-logind[1502]: New session 22 of user core. Mar 19 12:02:52.861094 systemd[1]: Started session-22.scope - Session 22 of User core. Mar 19 12:02:52.887018 kubelet[2805]: E0319 12:02:52.886969 2805 kubelet.go:2901] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Mar 19 12:02:53.974310 kubelet[2805]: E0319 12:02:53.974176 2805 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="4ad24ea6-768c-404d-9de4-78cdd82836aa" containerName="cilium-operator" Mar 19 12:02:53.974310 kubelet[2805]: E0319 12:02:53.974207 2805 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="40347df1-b19f-4102-8759-3c41dc3a468d" containerName="mount-cgroup" Mar 19 12:02:53.974310 kubelet[2805]: E0319 12:02:53.974213 2805 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="40347df1-b19f-4102-8759-3c41dc3a468d" containerName="clean-cilium-state" Mar 19 12:02:53.974310 kubelet[2805]: E0319 12:02:53.974219 2805 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="40347df1-b19f-4102-8759-3c41dc3a468d" containerName="apply-sysctl-overwrites" Mar 19 12:02:53.974310 kubelet[2805]: E0319 12:02:53.974226 2805 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="40347df1-b19f-4102-8759-3c41dc3a468d" containerName="mount-bpf-fs" Mar 19 12:02:53.974310 kubelet[2805]: E0319 12:02:53.974230 2805 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="40347df1-b19f-4102-8759-3c41dc3a468d" containerName="cilium-agent" Mar 19 12:02:53.979424 kubelet[2805]: I0319 12:02:53.979392 2805 memory_manager.go:354] "RemoveStaleState removing state" podUID="40347df1-b19f-4102-8759-3c41dc3a468d" containerName="cilium-agent" Mar 19 12:02:53.979667 kubelet[2805]: I0319 12:02:53.979658 2805 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ad24ea6-768c-404d-9de4-78cdd82836aa" containerName="cilium-operator" Mar 19 12:02:53.998309 systemd[1]: Created slice kubepods-burstable-pod8a20f9f8_bb09_43c5_8f93_9bd4d552573d.slice - libcontainer container kubepods-burstable-pod8a20f9f8_bb09_43c5_8f93_9bd4d552573d.slice. Mar 19 12:02:54.052744 kubelet[2805]: I0319 12:02:54.052683 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/8a20f9f8-bb09-43c5-8f93-9bd4d552573d-cilium-cgroup\") pod \"cilium-bwzhc\" (UID: \"8a20f9f8-bb09-43c5-8f93-9bd4d552573d\") " pod="kube-system/cilium-bwzhc" Mar 19 12:02:54.052744 kubelet[2805]: I0319 12:02:54.052723 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/8a20f9f8-bb09-43c5-8f93-9bd4d552573d-hubble-tls\") pod \"cilium-bwzhc\" (UID: \"8a20f9f8-bb09-43c5-8f93-9bd4d552573d\") " pod="kube-system/cilium-bwzhc" Mar 19 12:02:54.052744 kubelet[2805]: I0319 12:02:54.052740 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8a20f9f8-bb09-43c5-8f93-9bd4d552573d-etc-cni-netd\") pod \"cilium-bwzhc\" (UID: \"8a20f9f8-bb09-43c5-8f93-9bd4d552573d\") " pod="kube-system/cilium-bwzhc" Mar 19 12:02:54.052744 kubelet[2805]: I0319 12:02:54.052755 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/8a20f9f8-bb09-43c5-8f93-9bd4d552573d-xtables-lock\") pod \"cilium-bwzhc\" (UID: \"8a20f9f8-bb09-43c5-8f93-9bd4d552573d\") " pod="kube-system/cilium-bwzhc" Mar 19 12:02:54.053068 kubelet[2805]: I0319 12:02:54.052768 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/8a20f9f8-bb09-43c5-8f93-9bd4d552573d-cilium-config-path\") pod \"cilium-bwzhc\" (UID: \"8a20f9f8-bb09-43c5-8f93-9bd4d552573d\") " pod="kube-system/cilium-bwzhc" Mar 19 12:02:54.053068 kubelet[2805]: I0319 12:02:54.052785 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/8a20f9f8-bb09-43c5-8f93-9bd4d552573d-host-proc-sys-kernel\") pod \"cilium-bwzhc\" (UID: \"8a20f9f8-bb09-43c5-8f93-9bd4d552573d\") " pod="kube-system/cilium-bwzhc" Mar 19 12:02:54.053068 kubelet[2805]: I0319 12:02:54.052818 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9m8w8\" (UniqueName: \"kubernetes.io/projected/8a20f9f8-bb09-43c5-8f93-9bd4d552573d-kube-api-access-9m8w8\") pod \"cilium-bwzhc\" (UID: \"8a20f9f8-bb09-43c5-8f93-9bd4d552573d\") " pod="kube-system/cilium-bwzhc" Mar 19 12:02:54.053068 kubelet[2805]: I0319 12:02:54.052834 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/8a20f9f8-bb09-43c5-8f93-9bd4d552573d-cilium-run\") pod \"cilium-bwzhc\" (UID: \"8a20f9f8-bb09-43c5-8f93-9bd4d552573d\") " pod="kube-system/cilium-bwzhc" Mar 19 12:02:54.053068 kubelet[2805]: I0319 12:02:54.052845 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/8a20f9f8-bb09-43c5-8f93-9bd4d552573d-cni-path\") pod \"cilium-bwzhc\" (UID: \"8a20f9f8-bb09-43c5-8f93-9bd4d552573d\") " pod="kube-system/cilium-bwzhc" Mar 19 12:02:54.053284 kubelet[2805]: I0319 12:02:54.052859 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/8a20f9f8-bb09-43c5-8f93-9bd4d552573d-host-proc-sys-net\") pod \"cilium-bwzhc\" (UID: \"8a20f9f8-bb09-43c5-8f93-9bd4d552573d\") " pod="kube-system/cilium-bwzhc" Mar 19 12:02:54.053284 kubelet[2805]: I0319 12:02:54.052892 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/8a20f9f8-bb09-43c5-8f93-9bd4d552573d-bpf-maps\") pod \"cilium-bwzhc\" (UID: \"8a20f9f8-bb09-43c5-8f93-9bd4d552573d\") " pod="kube-system/cilium-bwzhc" Mar 19 12:02:54.053284 kubelet[2805]: I0319 12:02:54.052906 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8a20f9f8-bb09-43c5-8f93-9bd4d552573d-lib-modules\") pod \"cilium-bwzhc\" (UID: \"8a20f9f8-bb09-43c5-8f93-9bd4d552573d\") " pod="kube-system/cilium-bwzhc" Mar 19 12:02:54.053284 kubelet[2805]: I0319 12:02:54.052921 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/8a20f9f8-bb09-43c5-8f93-9bd4d552573d-cilium-ipsec-secrets\") pod \"cilium-bwzhc\" (UID: \"8a20f9f8-bb09-43c5-8f93-9bd4d552573d\") " pod="kube-system/cilium-bwzhc" Mar 19 12:02:54.053284 kubelet[2805]: I0319 12:02:54.052937 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/8a20f9f8-bb09-43c5-8f93-9bd4d552573d-hostproc\") pod \"cilium-bwzhc\" (UID: \"8a20f9f8-bb09-43c5-8f93-9bd4d552573d\") " pod="kube-system/cilium-bwzhc" Mar 19 12:02:54.053284 kubelet[2805]: I0319 12:02:54.052950 2805 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/8a20f9f8-bb09-43c5-8f93-9bd4d552573d-clustermesh-secrets\") pod \"cilium-bwzhc\" (UID: \"8a20f9f8-bb09-43c5-8f93-9bd4d552573d\") " pod="kube-system/cilium-bwzhc" Mar 19 12:02:54.168904 sshd[4570]: Connection closed by 139.178.68.195 port 33522 Mar 19 12:02:54.170155 sshd-session[4568]: pam_unix(sshd:session): session closed for user core Mar 19 12:02:54.206827 systemd[1]: sshd@21-37.27.246.226:22-139.178.68.195:33522.service: Deactivated successfully. Mar 19 12:02:54.209643 systemd[1]: session-22.scope: Deactivated successfully. Mar 19 12:02:54.211221 systemd-logind[1502]: Session 22 logged out. Waiting for processes to exit. Mar 19 12:02:54.215534 systemd-logind[1502]: Removed session 22. Mar 19 12:02:54.309417 containerd[1531]: time="2025-03-19T12:02:54.309281557Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-bwzhc,Uid:8a20f9f8-bb09-43c5-8f93-9bd4d552573d,Namespace:kube-system,Attempt:0,}" Mar 19 12:02:54.345239 containerd[1531]: time="2025-03-19T12:02:54.344914731Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 19 12:02:54.345239 containerd[1531]: time="2025-03-19T12:02:54.345002867Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 19 12:02:54.345239 containerd[1531]: time="2025-03-19T12:02:54.345018756Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 19 12:02:54.345239 containerd[1531]: time="2025-03-19T12:02:54.345153318Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 19 12:02:54.374191 systemd[1]: Started sshd@22-37.27.246.226:22-139.178.68.195:33524.service - OpenSSH per-connection server daemon (139.178.68.195:33524). Mar 19 12:02:54.383371 systemd[1]: Started cri-containerd-a713ce80d1e0f75dde7b6dd8529bc1d762de7d0f9aaa8201037a98d384694cec.scope - libcontainer container a713ce80d1e0f75dde7b6dd8529bc1d762de7d0f9aaa8201037a98d384694cec. Mar 19 12:02:54.411341 containerd[1531]: time="2025-03-19T12:02:54.411268422Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-bwzhc,Uid:8a20f9f8-bb09-43c5-8f93-9bd4d552573d,Namespace:kube-system,Attempt:0,} returns sandbox id \"a713ce80d1e0f75dde7b6dd8529bc1d762de7d0f9aaa8201037a98d384694cec\"" Mar 19 12:02:54.424611 containerd[1531]: time="2025-03-19T12:02:54.424545370Z" level=info msg="CreateContainer within sandbox \"a713ce80d1e0f75dde7b6dd8529bc1d762de7d0f9aaa8201037a98d384694cec\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Mar 19 12:02:54.438504 containerd[1531]: time="2025-03-19T12:02:54.438426960Z" level=info msg="CreateContainer within sandbox \"a713ce80d1e0f75dde7b6dd8529bc1d762de7d0f9aaa8201037a98d384694cec\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"db6b463beddc18316983b07c31b3101fb54e1fcea7ddf5b8c51bb9b279cddbd2\"" Mar 19 12:02:54.439405 containerd[1531]: time="2025-03-19T12:02:54.439194269Z" level=info msg="StartContainer for \"db6b463beddc18316983b07c31b3101fb54e1fcea7ddf5b8c51bb9b279cddbd2\"" Mar 19 12:02:54.476015 systemd[1]: Started cri-containerd-db6b463beddc18316983b07c31b3101fb54e1fcea7ddf5b8c51bb9b279cddbd2.scope - libcontainer container db6b463beddc18316983b07c31b3101fb54e1fcea7ddf5b8c51bb9b279cddbd2. Mar 19 12:02:54.500550 containerd[1531]: time="2025-03-19T12:02:54.500006371Z" level=info msg="StartContainer for \"db6b463beddc18316983b07c31b3101fb54e1fcea7ddf5b8c51bb9b279cddbd2\" returns successfully" Mar 19 12:02:54.513853 systemd[1]: cri-containerd-db6b463beddc18316983b07c31b3101fb54e1fcea7ddf5b8c51bb9b279cddbd2.scope: Deactivated successfully. Mar 19 12:02:54.514341 systemd[1]: cri-containerd-db6b463beddc18316983b07c31b3101fb54e1fcea7ddf5b8c51bb9b279cddbd2.scope: Consumed 21ms CPU time, 9.1M memory peak, 2.5M read from disk. Mar 19 12:02:54.549831 containerd[1531]: time="2025-03-19T12:02:54.549718257Z" level=info msg="shim disconnected" id=db6b463beddc18316983b07c31b3101fb54e1fcea7ddf5b8c51bb9b279cddbd2 namespace=k8s.io Mar 19 12:02:54.549831 containerd[1531]: time="2025-03-19T12:02:54.549783459Z" level=warning msg="cleaning up after shim disconnected" id=db6b463beddc18316983b07c31b3101fb54e1fcea7ddf5b8c51bb9b279cddbd2 namespace=k8s.io Mar 19 12:02:54.549831 containerd[1531]: time="2025-03-19T12:02:54.549807644Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 19 12:02:54.620926 containerd[1531]: time="2025-03-19T12:02:54.619921143Z" level=info msg="CreateContainer within sandbox \"a713ce80d1e0f75dde7b6dd8529bc1d762de7d0f9aaa8201037a98d384694cec\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Mar 19 12:02:54.643083 containerd[1531]: time="2025-03-19T12:02:54.642020928Z" level=info msg="CreateContainer within sandbox \"a713ce80d1e0f75dde7b6dd8529bc1d762de7d0f9aaa8201037a98d384694cec\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"3664eb155fc2f472bd5a7a55e84b997156e085011ab60fdd6c0e7d4f38a78472\"" Mar 19 12:02:54.645134 containerd[1531]: time="2025-03-19T12:02:54.645091344Z" level=info msg="StartContainer for \"3664eb155fc2f472bd5a7a55e84b997156e085011ab60fdd6c0e7d4f38a78472\"" Mar 19 12:02:54.681395 systemd[1]: Started cri-containerd-3664eb155fc2f472bd5a7a55e84b997156e085011ab60fdd6c0e7d4f38a78472.scope - libcontainer container 3664eb155fc2f472bd5a7a55e84b997156e085011ab60fdd6c0e7d4f38a78472. Mar 19 12:02:54.715489 containerd[1531]: time="2025-03-19T12:02:54.715286054Z" level=info msg="StartContainer for \"3664eb155fc2f472bd5a7a55e84b997156e085011ab60fdd6c0e7d4f38a78472\" returns successfully" Mar 19 12:02:54.729982 systemd[1]: cri-containerd-3664eb155fc2f472bd5a7a55e84b997156e085011ab60fdd6c0e7d4f38a78472.scope: Deactivated successfully. Mar 19 12:02:54.730436 systemd[1]: cri-containerd-3664eb155fc2f472bd5a7a55e84b997156e085011ab60fdd6c0e7d4f38a78472.scope: Consumed 21ms CPU time, 7.1M memory peak, 1.8M read from disk. Mar 19 12:02:54.756928 containerd[1531]: time="2025-03-19T12:02:54.756810162Z" level=info msg="shim disconnected" id=3664eb155fc2f472bd5a7a55e84b997156e085011ab60fdd6c0e7d4f38a78472 namespace=k8s.io Mar 19 12:02:54.756928 containerd[1531]: time="2025-03-19T12:02:54.756912864Z" level=warning msg="cleaning up after shim disconnected" id=3664eb155fc2f472bd5a7a55e84b997156e085011ab60fdd6c0e7d4f38a78472 namespace=k8s.io Mar 19 12:02:54.756928 containerd[1531]: time="2025-03-19T12:02:54.756924706Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 19 12:02:54.881261 kubelet[2805]: I0319 12:02:54.880101 2805 setters.go:600] "Node became not ready" node="ci-4230-1-0-d-e284590f49" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-03-19T12:02:54Z","lastTransitionTime":"2025-03-19T12:02:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} Mar 19 12:02:55.452424 sshd[4611]: Accepted publickey for core from 139.178.68.195 port 33524 ssh2: RSA SHA256:mZbisEv4EcLuWBRjDimak4H4E9nhJ1zN2E80H6lgS54 Mar 19 12:02:55.454360 sshd-session[4611]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 19 12:02:55.461053 systemd-logind[1502]: New session 23 of user core. Mar 19 12:02:55.474183 systemd[1]: Started session-23.scope - Session 23 of User core. Mar 19 12:02:55.618920 containerd[1531]: time="2025-03-19T12:02:55.618846133Z" level=info msg="CreateContainer within sandbox \"a713ce80d1e0f75dde7b6dd8529bc1d762de7d0f9aaa8201037a98d384694cec\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Mar 19 12:02:55.643413 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount883523381.mount: Deactivated successfully. Mar 19 12:02:55.645147 containerd[1531]: time="2025-03-19T12:02:55.643632124Z" level=info msg="CreateContainer within sandbox \"a713ce80d1e0f75dde7b6dd8529bc1d762de7d0f9aaa8201037a98d384694cec\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"1eccbcbec4f3bb6825d51fde83bd5a1e1d8e187d3a6b6a7cab29868c7e5bd017\"" Mar 19 12:02:55.647035 containerd[1531]: time="2025-03-19T12:02:55.646094670Z" level=info msg="StartContainer for \"1eccbcbec4f3bb6825d51fde83bd5a1e1d8e187d3a6b6a7cab29868c7e5bd017\"" Mar 19 12:02:55.692133 systemd[1]: Started cri-containerd-1eccbcbec4f3bb6825d51fde83bd5a1e1d8e187d3a6b6a7cab29868c7e5bd017.scope - libcontainer container 1eccbcbec4f3bb6825d51fde83bd5a1e1d8e187d3a6b6a7cab29868c7e5bd017. Mar 19 12:02:55.731378 containerd[1531]: time="2025-03-19T12:02:55.731134719Z" level=info msg="StartContainer for \"1eccbcbec4f3bb6825d51fde83bd5a1e1d8e187d3a6b6a7cab29868c7e5bd017\" returns successfully" Mar 19 12:02:55.737268 systemd[1]: cri-containerd-1eccbcbec4f3bb6825d51fde83bd5a1e1d8e187d3a6b6a7cab29868c7e5bd017.scope: Deactivated successfully. Mar 19 12:02:55.766070 containerd[1531]: time="2025-03-19T12:02:55.766003310Z" level=info msg="shim disconnected" id=1eccbcbec4f3bb6825d51fde83bd5a1e1d8e187d3a6b6a7cab29868c7e5bd017 namespace=k8s.io Mar 19 12:02:55.766070 containerd[1531]: time="2025-03-19T12:02:55.766053283Z" level=warning msg="cleaning up after shim disconnected" id=1eccbcbec4f3bb6825d51fde83bd5a1e1d8e187d3a6b6a7cab29868c7e5bd017 namespace=k8s.io Mar 19 12:02:55.766070 containerd[1531]: time="2025-03-19T12:02:55.766061759Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 19 12:02:56.163719 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-1eccbcbec4f3bb6825d51fde83bd5a1e1d8e187d3a6b6a7cab29868c7e5bd017-rootfs.mount: Deactivated successfully. Mar 19 12:02:56.183312 sshd[4751]: Connection closed by 139.178.68.195 port 33524 Mar 19 12:02:56.183980 sshd-session[4611]: pam_unix(sshd:session): session closed for user core Mar 19 12:02:56.186692 systemd[1]: sshd@22-37.27.246.226:22-139.178.68.195:33524.service: Deactivated successfully. Mar 19 12:02:56.188759 systemd[1]: session-23.scope: Deactivated successfully. Mar 19 12:02:56.190021 systemd-logind[1502]: Session 23 logged out. Waiting for processes to exit. Mar 19 12:02:56.191546 systemd-logind[1502]: Removed session 23. Mar 19 12:02:56.349537 systemd[1]: Started sshd@23-37.27.246.226:22-139.178.68.195:48598.service - OpenSSH per-connection server daemon (139.178.68.195:48598). Mar 19 12:02:56.624251 containerd[1531]: time="2025-03-19T12:02:56.624193873Z" level=info msg="CreateContainer within sandbox \"a713ce80d1e0f75dde7b6dd8529bc1d762de7d0f9aaa8201037a98d384694cec\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Mar 19 12:02:56.648099 containerd[1531]: time="2025-03-19T12:02:56.648028101Z" level=info msg="CreateContainer within sandbox \"a713ce80d1e0f75dde7b6dd8529bc1d762de7d0f9aaa8201037a98d384694cec\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"24c2ae48388735f55b298aee4159871570273fb7702be145c52d7b93bc509e24\"" Mar 19 12:02:56.653212 containerd[1531]: time="2025-03-19T12:02:56.653141766Z" level=info msg="StartContainer for \"24c2ae48388735f55b298aee4159871570273fb7702be145c52d7b93bc509e24\"" Mar 19 12:02:56.697122 systemd[1]: Started cri-containerd-24c2ae48388735f55b298aee4159871570273fb7702be145c52d7b93bc509e24.scope - libcontainer container 24c2ae48388735f55b298aee4159871570273fb7702be145c52d7b93bc509e24. Mar 19 12:02:56.731554 systemd[1]: cri-containerd-24c2ae48388735f55b298aee4159871570273fb7702be145c52d7b93bc509e24.scope: Deactivated successfully. Mar 19 12:02:56.732472 containerd[1531]: time="2025-03-19T12:02:56.732338341Z" level=info msg="StartContainer for \"24c2ae48388735f55b298aee4159871570273fb7702be145c52d7b93bc509e24\" returns successfully" Mar 19 12:02:56.755873 containerd[1531]: time="2025-03-19T12:02:56.755794098Z" level=info msg="shim disconnected" id=24c2ae48388735f55b298aee4159871570273fb7702be145c52d7b93bc509e24 namespace=k8s.io Mar 19 12:02:56.756164 containerd[1531]: time="2025-03-19T12:02:56.756032685Z" level=warning msg="cleaning up after shim disconnected" id=24c2ae48388735f55b298aee4159871570273fb7702be145c52d7b93bc509e24 namespace=k8s.io Mar 19 12:02:56.756164 containerd[1531]: time="2025-03-19T12:02:56.756045870Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 19 12:02:57.163692 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-24c2ae48388735f55b298aee4159871570273fb7702be145c52d7b93bc509e24-rootfs.mount: Deactivated successfully. Mar 19 12:02:57.369464 sshd[4813]: Accepted publickey for core from 139.178.68.195 port 48598 ssh2: RSA SHA256:mZbisEv4EcLuWBRjDimak4H4E9nhJ1zN2E80H6lgS54 Mar 19 12:02:57.371294 sshd-session[4813]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 19 12:02:57.379552 systemd-logind[1502]: New session 24 of user core. Mar 19 12:02:57.391214 systemd[1]: Started session-24.scope - Session 24 of User core. Mar 19 12:02:57.632970 containerd[1531]: time="2025-03-19T12:02:57.632891703Z" level=info msg="CreateContainer within sandbox \"a713ce80d1e0f75dde7b6dd8529bc1d762de7d0f9aaa8201037a98d384694cec\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Mar 19 12:02:57.652497 containerd[1531]: time="2025-03-19T12:02:57.651960588Z" level=info msg="CreateContainer within sandbox \"a713ce80d1e0f75dde7b6dd8529bc1d762de7d0f9aaa8201037a98d384694cec\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"7adac9a35676e79b14a49cfba62912084a3c9d2df212dbcf82a5bccc62d8649c\"" Mar 19 12:02:57.656666 containerd[1531]: time="2025-03-19T12:02:57.656521567Z" level=info msg="StartContainer for \"7adac9a35676e79b14a49cfba62912084a3c9d2df212dbcf82a5bccc62d8649c\"" Mar 19 12:02:57.712154 systemd[1]: Started cri-containerd-7adac9a35676e79b14a49cfba62912084a3c9d2df212dbcf82a5bccc62d8649c.scope - libcontainer container 7adac9a35676e79b14a49cfba62912084a3c9d2df212dbcf82a5bccc62d8649c. Mar 19 12:02:57.750102 containerd[1531]: time="2025-03-19T12:02:57.749831817Z" level=info msg="StartContainer for \"7adac9a35676e79b14a49cfba62912084a3c9d2df212dbcf82a5bccc62d8649c\" returns successfully" Mar 19 12:02:58.226921 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aesni)) Mar 19 12:02:58.668408 kubelet[2805]: I0319 12:02:58.667748 2805 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-bwzhc" podStartSLOduration=5.667714262 podStartE2EDuration="5.667714262s" podCreationTimestamp="2025-03-19 12:02:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-19 12:02:58.667549103 +0000 UTC m=+351.056743579" watchObservedRunningTime="2025-03-19 12:02:58.667714262 +0000 UTC m=+351.056908739" Mar 19 12:03:00.412851 systemd[1]: run-containerd-runc-k8s.io-7adac9a35676e79b14a49cfba62912084a3c9d2df212dbcf82a5bccc62d8649c-runc.8e6JKI.mount: Deactivated successfully. Mar 19 12:03:01.198296 systemd-networkd[1415]: lxc_health: Link UP Mar 19 12:03:01.206776 systemd-networkd[1415]: lxc_health: Gained carrier Mar 19 12:03:02.686470 kubelet[2805]: E0319 12:03:02.686385 2805 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 127.0.0.1:50080->127.0.0.1:33187: write tcp 127.0.0.1:50080->127.0.0.1:33187: write: broken pipe Mar 19 12:03:03.072307 systemd-networkd[1415]: lxc_health: Gained IPv6LL Mar 19 12:03:06.936184 systemd[1]: run-containerd-runc-k8s.io-7adac9a35676e79b14a49cfba62912084a3c9d2df212dbcf82a5bccc62d8649c-runc.HOgGgf.mount: Deactivated successfully. Mar 19 12:03:07.210346 sshd[4871]: Connection closed by 139.178.68.195 port 48598 Mar 19 12:03:07.211770 sshd-session[4813]: pam_unix(sshd:session): session closed for user core Mar 19 12:03:07.216439 systemd[1]: sshd@23-37.27.246.226:22-139.178.68.195:48598.service: Deactivated successfully. Mar 19 12:03:07.218772 systemd[1]: session-24.scope: Deactivated successfully. Mar 19 12:03:07.219814 systemd-logind[1502]: Session 24 logged out. Waiting for processes to exit. Mar 19 12:03:07.221603 systemd-logind[1502]: Removed session 24. Mar 19 12:03:07.756726 containerd[1531]: time="2025-03-19T12:03:07.756648629Z" level=info msg="StopPodSandbox for \"bb72b6cb42b223cfe8537b613d2b5912f92fab908bf4d332a5d91b06d19e9421\"" Mar 19 12:03:07.757377 containerd[1531]: time="2025-03-19T12:03:07.756768664Z" level=info msg="TearDown network for sandbox \"bb72b6cb42b223cfe8537b613d2b5912f92fab908bf4d332a5d91b06d19e9421\" successfully" Mar 19 12:03:07.757377 containerd[1531]: time="2025-03-19T12:03:07.756783823Z" level=info msg="StopPodSandbox for \"bb72b6cb42b223cfe8537b613d2b5912f92fab908bf4d332a5d91b06d19e9421\" returns successfully" Mar 19 12:03:07.757377 containerd[1531]: time="2025-03-19T12:03:07.757237844Z" level=info msg="RemovePodSandbox for \"bb72b6cb42b223cfe8537b613d2b5912f92fab908bf4d332a5d91b06d19e9421\"" Mar 19 12:03:07.757377 containerd[1531]: time="2025-03-19T12:03:07.757271086Z" level=info msg="Forcibly stopping sandbox \"bb72b6cb42b223cfe8537b613d2b5912f92fab908bf4d332a5d91b06d19e9421\"" Mar 19 12:03:07.757377 containerd[1531]: time="2025-03-19T12:03:07.757315449Z" level=info msg="TearDown network for sandbox \"bb72b6cb42b223cfe8537b613d2b5912f92fab908bf4d332a5d91b06d19e9421\" successfully" Mar 19 12:03:07.762284 containerd[1531]: time="2025-03-19T12:03:07.762214723Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"bb72b6cb42b223cfe8537b613d2b5912f92fab908bf4d332a5d91b06d19e9421\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 19 12:03:07.762473 containerd[1531]: time="2025-03-19T12:03:07.762320641Z" level=info msg="RemovePodSandbox \"bb72b6cb42b223cfe8537b613d2b5912f92fab908bf4d332a5d91b06d19e9421\" returns successfully" Mar 19 12:03:07.763011 containerd[1531]: time="2025-03-19T12:03:07.762983364Z" level=info msg="StopPodSandbox for \"e1ac54e4ade1aee1a79f559a223841973bf4d1dbd6882bd882ecd9a1600d4dfa\"" Mar 19 12:03:07.763075 containerd[1531]: time="2025-03-19T12:03:07.763065919Z" level=info msg="TearDown network for sandbox \"e1ac54e4ade1aee1a79f559a223841973bf4d1dbd6882bd882ecd9a1600d4dfa\" successfully" Mar 19 12:03:07.763107 containerd[1531]: time="2025-03-19T12:03:07.763078883Z" level=info msg="StopPodSandbox for \"e1ac54e4ade1aee1a79f559a223841973bf4d1dbd6882bd882ecd9a1600d4dfa\" returns successfully" Mar 19 12:03:07.763343 containerd[1531]: time="2025-03-19T12:03:07.763313022Z" level=info msg="RemovePodSandbox for \"e1ac54e4ade1aee1a79f559a223841973bf4d1dbd6882bd882ecd9a1600d4dfa\"" Mar 19 12:03:07.763343 containerd[1531]: time="2025-03-19T12:03:07.763338540Z" level=info msg="Forcibly stopping sandbox \"e1ac54e4ade1aee1a79f559a223841973bf4d1dbd6882bd882ecd9a1600d4dfa\"" Mar 19 12:03:07.763552 containerd[1531]: time="2025-03-19T12:03:07.763388032Z" level=info msg="TearDown network for sandbox \"e1ac54e4ade1aee1a79f559a223841973bf4d1dbd6882bd882ecd9a1600d4dfa\" successfully" Mar 19 12:03:07.768024 containerd[1531]: time="2025-03-19T12:03:07.767968067Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"e1ac54e4ade1aee1a79f559a223841973bf4d1dbd6882bd882ecd9a1600d4dfa\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 19 12:03:07.768243 containerd[1531]: time="2025-03-19T12:03:07.768055821Z" level=info msg="RemovePodSandbox \"e1ac54e4ade1aee1a79f559a223841973bf4d1dbd6882bd882ecd9a1600d4dfa\" returns successfully" Mar 19 12:03:23.682415 systemd[1]: cri-containerd-9f10d9e4adbc8baea36d71180b83e9c5adc60a9f1123335a1e2543df3606fa2b.scope: Deactivated successfully. Mar 19 12:03:23.682900 systemd[1]: cri-containerd-9f10d9e4adbc8baea36d71180b83e9c5adc60a9f1123335a1e2543df3606fa2b.scope: Consumed 6.934s CPU time, 74.6M memory peak, 20.4M read from disk. Mar 19 12:03:23.706716 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-9f10d9e4adbc8baea36d71180b83e9c5adc60a9f1123335a1e2543df3606fa2b-rootfs.mount: Deactivated successfully. Mar 19 12:03:23.725773 containerd[1531]: time="2025-03-19T12:03:23.725704759Z" level=info msg="shim disconnected" id=9f10d9e4adbc8baea36d71180b83e9c5adc60a9f1123335a1e2543df3606fa2b namespace=k8s.io Mar 19 12:03:23.725773 containerd[1531]: time="2025-03-19T12:03:23.725768529Z" level=warning msg="cleaning up after shim disconnected" id=9f10d9e4adbc8baea36d71180b83e9c5adc60a9f1123335a1e2543df3606fa2b namespace=k8s.io Mar 19 12:03:23.725773 containerd[1531]: time="2025-03-19T12:03:23.725778498Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 19 12:03:23.811309 kubelet[2805]: E0319 12:03:23.811244 2805 controller.go:195] "Failed to update lease" err="rpc error: code = Unavailable desc = error reading from server: read tcp 10.0.0.3:43650->10.0.0.2:2379: read: connection timed out" Mar 19 12:03:24.724061 kubelet[2805]: I0319 12:03:24.723814 2805 scope.go:117] "RemoveContainer" containerID="9f10d9e4adbc8baea36d71180b83e9c5adc60a9f1123335a1e2543df3606fa2b" Mar 19 12:03:24.727157 containerd[1531]: time="2025-03-19T12:03:24.726911161Z" level=info msg="CreateContainer within sandbox \"10a9e8444d39b2481e25c41a1fb3fd0e41a47b4b6e056dc4d826c7f907ccb429\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:1,}" Mar 19 12:03:24.747661 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4105407342.mount: Deactivated successfully. Mar 19 12:03:24.756929 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount781500621.mount: Deactivated successfully. Mar 19 12:03:24.758540 containerd[1531]: time="2025-03-19T12:03:24.758442067Z" level=info msg="CreateContainer within sandbox \"10a9e8444d39b2481e25c41a1fb3fd0e41a47b4b6e056dc4d826c7f907ccb429\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:1,} returns container id \"704143469274684a9ad6081646f956b5be663ea1899c10c7c507df0de9e109d4\"" Mar 19 12:03:24.759432 containerd[1531]: time="2025-03-19T12:03:24.759383941Z" level=info msg="StartContainer for \"704143469274684a9ad6081646f956b5be663ea1899c10c7c507df0de9e109d4\"" Mar 19 12:03:24.785098 systemd[1]: Started cri-containerd-704143469274684a9ad6081646f956b5be663ea1899c10c7c507df0de9e109d4.scope - libcontainer container 704143469274684a9ad6081646f956b5be663ea1899c10c7c507df0de9e109d4. Mar 19 12:03:24.824912 containerd[1531]: time="2025-03-19T12:03:24.824825966Z" level=info msg="StartContainer for \"704143469274684a9ad6081646f956b5be663ea1899c10c7c507df0de9e109d4\" returns successfully" Mar 19 12:03:25.222625 kubelet[2805]: E0319 12:03:25.219762 2805 event.go:359] "Server rejected event (will not retry!)" err="rpc error: code = Unavailable desc = error reading from server: read tcp 10.0.0.3:43458->10.0.0.2:2379: read: connection timed out" event="&Event{ObjectMeta:{kube-apiserver-ci-4230-1-0-d-e284590f49.182e32a6be86efab kube-system 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:kube-system,Name:kube-apiserver-ci-4230-1-0-d-e284590f49,UID:0c2706b2816e41ac0a665b5d91275921,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Readiness probe failed: HTTP probe failed with statuscode: 500,Source:EventSource{Component:kubelet,Host:ci-4230-1-0-d-e284590f49,},FirstTimestamp:2025-03-19 12:03:17.581434795 +0000 UTC m=+369.970629272,LastTimestamp:2025-03-19 12:03:17.581434795 +0000 UTC m=+369.970629272,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4230-1-0-d-e284590f49,}" Mar 19 12:03:28.899431 systemd[1]: cri-containerd-c49290af89039639c3abe160eed3e98e253f51662a295c8d10c71a212cbc9f6d.scope: Deactivated successfully. Mar 19 12:03:28.901079 systemd[1]: cri-containerd-c49290af89039639c3abe160eed3e98e253f51662a295c8d10c71a212cbc9f6d.scope: Consumed 1.984s CPU time, 31.5M memory peak, 13.1M read from disk. Mar 19 12:03:28.932190 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-c49290af89039639c3abe160eed3e98e253f51662a295c8d10c71a212cbc9f6d-rootfs.mount: Deactivated successfully. Mar 19 12:03:28.939606 containerd[1531]: time="2025-03-19T12:03:28.939464944Z" level=info msg="shim disconnected" id=c49290af89039639c3abe160eed3e98e253f51662a295c8d10c71a212cbc9f6d namespace=k8s.io Mar 19 12:03:28.939606 containerd[1531]: time="2025-03-19T12:03:28.939571783Z" level=warning msg="cleaning up after shim disconnected" id=c49290af89039639c3abe160eed3e98e253f51662a295c8d10c71a212cbc9f6d namespace=k8s.io Mar 19 12:03:28.939606 containerd[1531]: time="2025-03-19T12:03:28.939583105Z" level=info msg="cleaning up dead shim" namespace=k8s.io