Dec 16 09:36:54.871601 kernel: Linux version 6.6.65-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Thu Dec 12 23:15:00 -00 2024 Dec 16 09:36:54.871622 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=hetzner verity.usrhash=2fdbba50b59d8c8a9877a81151806ddc16f473fe99b9ba0d8825997d654583ff Dec 16 09:36:54.871630 kernel: BIOS-provided physical RAM map: Dec 16 09:36:54.871636 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable Dec 16 09:36:54.871640 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved Dec 16 09:36:54.871645 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Dec 16 09:36:54.871651 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000007cfdbfff] usable Dec 16 09:36:54.871656 kernel: BIOS-e820: [mem 0x000000007cfdc000-0x000000007cffffff] reserved Dec 16 09:36:54.871664 kernel: BIOS-e820: [mem 0x00000000b0000000-0x00000000bfffffff] reserved Dec 16 09:36:54.871669 kernel: BIOS-e820: [mem 0x00000000fed1c000-0x00000000fed1ffff] reserved Dec 16 09:36:54.871674 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Dec 16 09:36:54.871679 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Dec 16 09:36:54.871684 kernel: NX (Execute Disable) protection: active Dec 16 09:36:54.871689 kernel: APIC: Static calls initialized Dec 16 09:36:54.871697 kernel: SMBIOS 2.8 present. Dec 16 09:36:54.871703 kernel: DMI: Hetzner vServer/Standard PC (Q35 + ICH9, 2009), BIOS 20171111 11/11/2017 Dec 16 09:36:54.871709 kernel: Hypervisor detected: KVM Dec 16 09:36:54.871714 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Dec 16 09:36:54.871719 kernel: kvm-clock: using sched offset of 2896591470 cycles Dec 16 09:36:54.871725 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Dec 16 09:36:54.871731 kernel: tsc: Detected 2445.404 MHz processor Dec 16 09:36:54.871736 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Dec 16 09:36:54.871742 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Dec 16 09:36:54.871750 kernel: last_pfn = 0x7cfdc max_arch_pfn = 0x400000000 Dec 16 09:36:54.871756 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs Dec 16 09:36:54.871761 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Dec 16 09:36:54.871767 kernel: Using GB pages for direct mapping Dec 16 09:36:54.871772 kernel: ACPI: Early table checksum verification disabled Dec 16 09:36:54.871778 kernel: ACPI: RSDP 0x00000000000F51F0 000014 (v00 BOCHS ) Dec 16 09:36:54.871783 kernel: ACPI: RSDT 0x000000007CFE265D 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 16 09:36:54.871789 kernel: ACPI: FACP 0x000000007CFE244D 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Dec 16 09:36:54.871794 kernel: ACPI: DSDT 0x000000007CFE0040 00240D (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 16 09:36:54.871802 kernel: ACPI: FACS 0x000000007CFE0000 000040 Dec 16 09:36:54.871807 kernel: ACPI: APIC 0x000000007CFE2541 000080 (v03 BOCHS BXPC 00000001 BXPC 00000001) Dec 16 09:36:54.871813 kernel: ACPI: HPET 0x000000007CFE25C1 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 16 09:36:54.871818 kernel: ACPI: MCFG 0x000000007CFE25F9 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 16 09:36:54.871824 kernel: ACPI: WAET 0x000000007CFE2635 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 16 09:36:54.871830 kernel: ACPI: Reserving FACP table memory at [mem 0x7cfe244d-0x7cfe2540] Dec 16 09:36:54.871835 kernel: ACPI: Reserving DSDT table memory at [mem 0x7cfe0040-0x7cfe244c] Dec 16 09:36:54.871841 kernel: ACPI: Reserving FACS table memory at [mem 0x7cfe0000-0x7cfe003f] Dec 16 09:36:54.871851 kernel: ACPI: Reserving APIC table memory at [mem 0x7cfe2541-0x7cfe25c0] Dec 16 09:36:54.871857 kernel: ACPI: Reserving HPET table memory at [mem 0x7cfe25c1-0x7cfe25f8] Dec 16 09:36:54.871863 kernel: ACPI: Reserving MCFG table memory at [mem 0x7cfe25f9-0x7cfe2634] Dec 16 09:36:54.871869 kernel: ACPI: Reserving WAET table memory at [mem 0x7cfe2635-0x7cfe265c] Dec 16 09:36:54.871874 kernel: No NUMA configuration found Dec 16 09:36:54.871880 kernel: Faking a node at [mem 0x0000000000000000-0x000000007cfdbfff] Dec 16 09:36:54.871888 kernel: NODE_DATA(0) allocated [mem 0x7cfd6000-0x7cfdbfff] Dec 16 09:36:54.871894 kernel: Zone ranges: Dec 16 09:36:54.871900 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Dec 16 09:36:54.871905 kernel: DMA32 [mem 0x0000000001000000-0x000000007cfdbfff] Dec 16 09:36:54.871911 kernel: Normal empty Dec 16 09:36:54.871917 kernel: Movable zone start for each node Dec 16 09:36:54.871922 kernel: Early memory node ranges Dec 16 09:36:54.871928 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Dec 16 09:36:54.871934 kernel: node 0: [mem 0x0000000000100000-0x000000007cfdbfff] Dec 16 09:36:54.871940 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000007cfdbfff] Dec 16 09:36:54.871948 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Dec 16 09:36:54.871953 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Dec 16 09:36:54.871959 kernel: On node 0, zone DMA32: 12324 pages in unavailable ranges Dec 16 09:36:54.871965 kernel: ACPI: PM-Timer IO Port: 0x608 Dec 16 09:36:54.871971 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Dec 16 09:36:54.871976 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Dec 16 09:36:54.871982 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Dec 16 09:36:54.871988 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Dec 16 09:36:54.871994 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Dec 16 09:36:54.872001 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Dec 16 09:36:54.872007 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Dec 16 09:36:54.872013 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Dec 16 09:36:54.872019 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Dec 16 09:36:54.872024 kernel: smpboot: Allowing 2 CPUs, 0 hotplug CPUs Dec 16 09:36:54.872030 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Dec 16 09:36:54.872036 kernel: [mem 0xc0000000-0xfed1bfff] available for PCI devices Dec 16 09:36:54.872041 kernel: Booting paravirtualized kernel on KVM Dec 16 09:36:54.872047 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Dec 16 09:36:54.872055 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1 Dec 16 09:36:54.872061 kernel: percpu: Embedded 58 pages/cpu s197032 r8192 d32344 u1048576 Dec 16 09:36:54.872067 kernel: pcpu-alloc: s197032 r8192 d32344 u1048576 alloc=1*2097152 Dec 16 09:36:54.872072 kernel: pcpu-alloc: [0] 0 1 Dec 16 09:36:54.872078 kernel: kvm-guest: PV spinlocks disabled, no host support Dec 16 09:36:54.872085 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=hetzner verity.usrhash=2fdbba50b59d8c8a9877a81151806ddc16f473fe99b9ba0d8825997d654583ff Dec 16 09:36:54.872091 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Dec 16 09:36:54.872097 kernel: random: crng init done Dec 16 09:36:54.872104 kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Dec 16 09:36:54.872110 kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear) Dec 16 09:36:54.872116 kernel: Fallback order for Node 0: 0 Dec 16 09:36:54.872122 kernel: Built 1 zonelists, mobility grouping on. Total pages: 503708 Dec 16 09:36:54.872127 kernel: Policy zone: DMA32 Dec 16 09:36:54.872133 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Dec 16 09:36:54.872139 kernel: Memory: 1922056K/2047464K available (12288K kernel code, 2299K rwdata, 22724K rodata, 42844K init, 2348K bss, 125148K reserved, 0K cma-reserved) Dec 16 09:36:54.872145 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Dec 16 09:36:54.872151 kernel: ftrace: allocating 37902 entries in 149 pages Dec 16 09:36:54.872158 kernel: ftrace: allocated 149 pages with 4 groups Dec 16 09:36:54.872164 kernel: Dynamic Preempt: voluntary Dec 16 09:36:54.872170 kernel: rcu: Preemptible hierarchical RCU implementation. Dec 16 09:36:54.872176 kernel: rcu: RCU event tracing is enabled. Dec 16 09:36:54.872183 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Dec 16 09:36:54.872189 kernel: Trampoline variant of Tasks RCU enabled. Dec 16 09:36:54.872194 kernel: Rude variant of Tasks RCU enabled. Dec 16 09:36:54.872200 kernel: Tracing variant of Tasks RCU enabled. Dec 16 09:36:54.872206 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Dec 16 09:36:54.872212 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Dec 16 09:36:54.872220 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 Dec 16 09:36:54.872226 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Dec 16 09:36:54.872232 kernel: Console: colour VGA+ 80x25 Dec 16 09:36:54.872237 kernel: printk: console [tty0] enabled Dec 16 09:36:54.872243 kernel: printk: console [ttyS0] enabled Dec 16 09:36:54.872252 kernel: ACPI: Core revision 20230628 Dec 16 09:36:54.872267 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Dec 16 09:36:54.872280 kernel: APIC: Switch to symmetric I/O mode setup Dec 16 09:36:54.872290 kernel: x2apic enabled Dec 16 09:36:54.872304 kernel: APIC: Switched APIC routing to: physical x2apic Dec 16 09:36:54.872310 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Dec 16 09:36:54.872316 kernel: tsc: Marking TSC unstable due to TSCs unsynchronized Dec 16 09:36:54.872321 kernel: Calibrating delay loop (skipped) preset value.. 4890.80 BogoMIPS (lpj=2445404) Dec 16 09:36:54.872327 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Dec 16 09:36:54.872334 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Dec 16 09:36:54.872340 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Dec 16 09:36:54.872346 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Dec 16 09:36:54.872360 kernel: Spectre V2 : Mitigation: Retpolines Dec 16 09:36:54.872366 kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch Dec 16 09:36:54.872372 kernel: Spectre V2 : Spectre v2 / SpectreRSB : Filling RSB on VMEXIT Dec 16 09:36:54.872380 kernel: Spectre V2 : Enabling Speculation Barrier for firmware calls Dec 16 09:36:54.872387 kernel: RETBleed: Mitigation: untrained return thunk Dec 16 09:36:54.872398 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Dec 16 09:36:54.872413 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Dec 16 09:36:54.872426 kernel: Speculative Return Stack Overflow: IBPB-extending microcode not applied! Dec 16 09:36:54.872438 kernel: Speculative Return Stack Overflow: WARNING: See https://kernel.org/doc/html/latest/admin-guide/hw-vuln/srso.html for mitigation options. Dec 16 09:36:54.872452 kernel: Speculative Return Stack Overflow: Vulnerable: Safe RET, no microcode Dec 16 09:36:54.872465 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Dec 16 09:36:54.872480 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Dec 16 09:36:54.872486 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Dec 16 09:36:54.872493 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Dec 16 09:36:54.872499 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. Dec 16 09:36:54.872505 kernel: Freeing SMP alternatives memory: 32K Dec 16 09:36:54.872513 kernel: pid_max: default: 32768 minimum: 301 Dec 16 09:36:54.872519 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Dec 16 09:36:54.872526 kernel: landlock: Up and running. Dec 16 09:36:54.872532 kernel: SELinux: Initializing. Dec 16 09:36:54.872538 kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Dec 16 09:36:54.872544 kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Dec 16 09:36:54.872550 kernel: smpboot: CPU0: AMD EPYC Processor (family: 0x17, model: 0x31, stepping: 0x0) Dec 16 09:36:54.872557 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Dec 16 09:36:54.872577 kernel: RCU Tasks Rude: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Dec 16 09:36:54.872587 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Dec 16 09:36:54.872597 kernel: Performance Events: Fam17h+ core perfctr, AMD PMU driver. Dec 16 09:36:54.872608 kernel: ... version: 0 Dec 16 09:36:54.872619 kernel: ... bit width: 48 Dec 16 09:36:54.872629 kernel: ... generic registers: 6 Dec 16 09:36:54.872640 kernel: ... value mask: 0000ffffffffffff Dec 16 09:36:54.872651 kernel: ... max period: 00007fffffffffff Dec 16 09:36:54.872661 kernel: ... fixed-purpose events: 0 Dec 16 09:36:54.872667 kernel: ... event mask: 000000000000003f Dec 16 09:36:54.872676 kernel: signal: max sigframe size: 1776 Dec 16 09:36:54.872682 kernel: rcu: Hierarchical SRCU implementation. Dec 16 09:36:54.872689 kernel: rcu: Max phase no-delay instances is 400. Dec 16 09:36:54.872695 kernel: smp: Bringing up secondary CPUs ... Dec 16 09:36:54.872701 kernel: smpboot: x86: Booting SMP configuration: Dec 16 09:36:54.872707 kernel: .... node #0, CPUs: #1 Dec 16 09:36:54.872713 kernel: smp: Brought up 1 node, 2 CPUs Dec 16 09:36:54.872719 kernel: smpboot: Max logical packages: 1 Dec 16 09:36:54.872725 kernel: smpboot: Total of 2 processors activated (9781.61 BogoMIPS) Dec 16 09:36:54.872731 kernel: devtmpfs: initialized Dec 16 09:36:54.872740 kernel: x86/mm: Memory block size: 128MB Dec 16 09:36:54.872746 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Dec 16 09:36:54.872752 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Dec 16 09:36:54.872758 kernel: pinctrl core: initialized pinctrl subsystem Dec 16 09:36:54.872764 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Dec 16 09:36:54.872770 kernel: audit: initializing netlink subsys (disabled) Dec 16 09:36:54.872776 kernel: audit: type=2000 audit(1734341813.538:1): state=initialized audit_enabled=0 res=1 Dec 16 09:36:54.872782 kernel: thermal_sys: Registered thermal governor 'step_wise' Dec 16 09:36:54.872789 kernel: thermal_sys: Registered thermal governor 'user_space' Dec 16 09:36:54.872796 kernel: cpuidle: using governor menu Dec 16 09:36:54.872803 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Dec 16 09:36:54.872809 kernel: dca service started, version 1.12.1 Dec 16 09:36:54.872815 kernel: PCI: MMCONFIG for domain 0000 [bus 00-ff] at [mem 0xb0000000-0xbfffffff] (base 0xb0000000) Dec 16 09:36:54.872821 kernel: PCI: Using configuration type 1 for base access Dec 16 09:36:54.872827 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Dec 16 09:36:54.872833 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Dec 16 09:36:54.872839 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Dec 16 09:36:54.872845 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Dec 16 09:36:54.872854 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Dec 16 09:36:54.872860 kernel: ACPI: Added _OSI(Module Device) Dec 16 09:36:54.872866 kernel: ACPI: Added _OSI(Processor Device) Dec 16 09:36:54.872872 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Dec 16 09:36:54.872878 kernel: ACPI: Added _OSI(Processor Aggregator Device) Dec 16 09:36:54.872884 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Dec 16 09:36:54.872890 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Dec 16 09:36:54.872896 kernel: ACPI: Interpreter enabled Dec 16 09:36:54.872902 kernel: ACPI: PM: (supports S0 S5) Dec 16 09:36:54.872910 kernel: ACPI: Using IOAPIC for interrupt routing Dec 16 09:36:54.872917 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Dec 16 09:36:54.872923 kernel: PCI: Using E820 reservations for host bridge windows Dec 16 09:36:54.872929 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Dec 16 09:36:54.872935 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Dec 16 09:36:54.873107 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Dec 16 09:36:54.873223 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Dec 16 09:36:54.873333 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Dec 16 09:36:54.873341 kernel: PCI host bridge to bus 0000:00 Dec 16 09:36:54.873451 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Dec 16 09:36:54.873547 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Dec 16 09:36:54.873696 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Dec 16 09:36:54.873796 kernel: pci_bus 0000:00: root bus resource [mem 0x7d000000-0xafffffff window] Dec 16 09:36:54.873889 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window] Dec 16 09:36:54.873987 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x8ffffffff window] Dec 16 09:36:54.874080 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Dec 16 09:36:54.874199 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 Dec 16 09:36:54.874313 kernel: pci 0000:00:01.0: [1af4:1050] type 00 class 0x030000 Dec 16 09:36:54.874417 kernel: pci 0000:00:01.0: reg 0x10: [mem 0xfb800000-0xfbffffff pref] Dec 16 09:36:54.874521 kernel: pci 0000:00:01.0: reg 0x18: [mem 0xfd200000-0xfd203fff 64bit pref] Dec 16 09:36:54.874662 kernel: pci 0000:00:01.0: reg 0x20: [mem 0xfea10000-0xfea10fff] Dec 16 09:36:54.874773 kernel: pci 0000:00:01.0: reg 0x30: [mem 0xfea00000-0xfea0ffff pref] Dec 16 09:36:54.874876 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Dec 16 09:36:54.874988 kernel: pci 0000:00:02.0: [1b36:000c] type 01 class 0x060400 Dec 16 09:36:54.875092 kernel: pci 0000:00:02.0: reg 0x10: [mem 0xfea11000-0xfea11fff] Dec 16 09:36:54.875203 kernel: pci 0000:00:02.1: [1b36:000c] type 01 class 0x060400 Dec 16 09:36:54.875306 kernel: pci 0000:00:02.1: reg 0x10: [mem 0xfea12000-0xfea12fff] Dec 16 09:36:54.875422 kernel: pci 0000:00:02.2: [1b36:000c] type 01 class 0x060400 Dec 16 09:36:54.875613 kernel: pci 0000:00:02.2: reg 0x10: [mem 0xfea13000-0xfea13fff] Dec 16 09:36:54.875799 kernel: pci 0000:00:02.3: [1b36:000c] type 01 class 0x060400 Dec 16 09:36:54.875913 kernel: pci 0000:00:02.3: reg 0x10: [mem 0xfea14000-0xfea14fff] Dec 16 09:36:54.876025 kernel: pci 0000:00:02.4: [1b36:000c] type 01 class 0x060400 Dec 16 09:36:54.876128 kernel: pci 0000:00:02.4: reg 0x10: [mem 0xfea15000-0xfea15fff] Dec 16 09:36:54.876248 kernel: pci 0000:00:02.5: [1b36:000c] type 01 class 0x060400 Dec 16 09:36:54.876351 kernel: pci 0000:00:02.5: reg 0x10: [mem 0xfea16000-0xfea16fff] Dec 16 09:36:54.876460 kernel: pci 0000:00:02.6: [1b36:000c] type 01 class 0x060400 Dec 16 09:36:54.876675 kernel: pci 0000:00:02.6: reg 0x10: [mem 0xfea17000-0xfea17fff] Dec 16 09:36:54.876802 kernel: pci 0000:00:02.7: [1b36:000c] type 01 class 0x060400 Dec 16 09:36:54.876910 kernel: pci 0000:00:02.7: reg 0x10: [mem 0xfea18000-0xfea18fff] Dec 16 09:36:54.877030 kernel: pci 0000:00:03.0: [1b36:000c] type 01 class 0x060400 Dec 16 09:36:54.877159 kernel: pci 0000:00:03.0: reg 0x10: [mem 0xfea19000-0xfea19fff] Dec 16 09:36:54.877274 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 Dec 16 09:36:54.877380 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Dec 16 09:36:54.877507 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 Dec 16 09:36:54.877686 kernel: pci 0000:00:1f.2: reg 0x20: [io 0xc040-0xc05f] Dec 16 09:36:54.877805 kernel: pci 0000:00:1f.2: reg 0x24: [mem 0xfea1a000-0xfea1afff] Dec 16 09:36:54.877918 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 Dec 16 09:36:54.878022 kernel: pci 0000:00:1f.3: reg 0x20: [io 0x0700-0x073f] Dec 16 09:36:54.878140 kernel: pci 0000:01:00.0: [1af4:1041] type 00 class 0x020000 Dec 16 09:36:54.878252 kernel: pci 0000:01:00.0: reg 0x14: [mem 0xfe880000-0xfe880fff] Dec 16 09:36:54.878363 kernel: pci 0000:01:00.0: reg 0x20: [mem 0xfd000000-0xfd003fff 64bit pref] Dec 16 09:36:54.878474 kernel: pci 0000:01:00.0: reg 0x30: [mem 0xfe800000-0xfe87ffff pref] Dec 16 09:36:54.878661 kernel: pci 0000:00:02.0: PCI bridge to [bus 01] Dec 16 09:36:54.878775 kernel: pci 0000:00:02.0: bridge window [mem 0xfe800000-0xfe9fffff] Dec 16 09:36:54.878884 kernel: pci 0000:00:02.0: bridge window [mem 0xfd000000-0xfd1fffff 64bit pref] Dec 16 09:36:54.879007 kernel: pci 0000:02:00.0: [1b36:000d] type 00 class 0x0c0330 Dec 16 09:36:54.879121 kernel: pci 0000:02:00.0: reg 0x10: [mem 0xfe600000-0xfe603fff 64bit] Dec 16 09:36:54.879230 kernel: pci 0000:00:02.1: PCI bridge to [bus 02] Dec 16 09:36:54.879338 kernel: pci 0000:00:02.1: bridge window [mem 0xfe600000-0xfe7fffff] Dec 16 09:36:54.879452 kernel: pci 0000:00:02.1: bridge window [mem 0xfce00000-0xfcffffff 64bit pref] Dec 16 09:36:54.879636 kernel: pci 0000:03:00.0: [1af4:1043] type 00 class 0x078000 Dec 16 09:36:54.879755 kernel: pci 0000:03:00.0: reg 0x14: [mem 0xfe400000-0xfe400fff] Dec 16 09:36:54.879867 kernel: pci 0000:03:00.0: reg 0x20: [mem 0xfcc00000-0xfcc03fff 64bit pref] Dec 16 09:36:54.879973 kernel: pci 0000:00:02.2: PCI bridge to [bus 03] Dec 16 09:36:54.880081 kernel: pci 0000:00:02.2: bridge window [mem 0xfe400000-0xfe5fffff] Dec 16 09:36:54.880185 kernel: pci 0000:00:02.2: bridge window [mem 0xfcc00000-0xfcdfffff 64bit pref] Dec 16 09:36:54.880349 kernel: pci 0000:04:00.0: [1af4:1045] type 00 class 0x00ff00 Dec 16 09:36:54.880469 kernel: pci 0000:04:00.0: reg 0x20: [mem 0xfca00000-0xfca03fff 64bit pref] Dec 16 09:36:54.880669 kernel: pci 0000:00:02.3: PCI bridge to [bus 04] Dec 16 09:36:54.880789 kernel: pci 0000:00:02.3: bridge window [mem 0xfe200000-0xfe3fffff] Dec 16 09:36:54.880898 kernel: pci 0000:00:02.3: bridge window [mem 0xfca00000-0xfcbfffff 64bit pref] Dec 16 09:36:54.881020 kernel: pci 0000:05:00.0: [1af4:1044] type 00 class 0x00ff00 Dec 16 09:36:54.881156 kernel: pci 0000:05:00.0: reg 0x20: [mem 0xfc800000-0xfc803fff 64bit pref] Dec 16 09:36:54.881275 kernel: pci 0000:00:02.4: PCI bridge to [bus 05] Dec 16 09:36:54.881383 kernel: pci 0000:00:02.4: bridge window [mem 0xfe000000-0xfe1fffff] Dec 16 09:36:54.881497 kernel: pci 0000:00:02.4: bridge window [mem 0xfc800000-0xfc9fffff 64bit pref] Dec 16 09:36:54.881703 kernel: pci 0000:06:00.0: [1af4:1048] type 00 class 0x010000 Dec 16 09:36:54.881820 kernel: pci 0000:06:00.0: reg 0x14: [mem 0xfde00000-0xfde00fff] Dec 16 09:36:54.881932 kernel: pci 0000:06:00.0: reg 0x20: [mem 0xfc600000-0xfc603fff 64bit pref] Dec 16 09:36:54.882039 kernel: pci 0000:00:02.5: PCI bridge to [bus 06] Dec 16 09:36:54.882151 kernel: pci 0000:00:02.5: bridge window [mem 0xfde00000-0xfdffffff] Dec 16 09:36:54.882275 kernel: pci 0000:00:02.5: bridge window [mem 0xfc600000-0xfc7fffff 64bit pref] Dec 16 09:36:54.882285 kernel: acpiphp: Slot [0] registered Dec 16 09:36:54.882402 kernel: pci 0000:07:00.0: [1af4:1041] type 00 class 0x020000 Dec 16 09:36:54.882518 kernel: pci 0000:07:00.0: reg 0x14: [mem 0xfdc80000-0xfdc80fff] Dec 16 09:36:54.882790 kernel: pci 0000:07:00.0: reg 0x20: [mem 0xfc400000-0xfc403fff 64bit pref] Dec 16 09:36:54.882908 kernel: pci 0000:07:00.0: reg 0x30: [mem 0xfdc00000-0xfdc7ffff pref] Dec 16 09:36:54.883017 kernel: pci 0000:00:02.6: PCI bridge to [bus 07] Dec 16 09:36:54.883130 kernel: pci 0000:00:02.6: bridge window [mem 0xfdc00000-0xfddfffff] Dec 16 09:36:54.883257 kernel: pci 0000:00:02.6: bridge window [mem 0xfc400000-0xfc5fffff 64bit pref] Dec 16 09:36:54.883274 kernel: acpiphp: Slot [0-2] registered Dec 16 09:36:54.883404 kernel: pci 0000:00:02.7: PCI bridge to [bus 08] Dec 16 09:36:54.883535 kernel: pci 0000:00:02.7: bridge window [mem 0xfda00000-0xfdbfffff] Dec 16 09:36:54.884206 kernel: pci 0000:00:02.7: bridge window [mem 0xfc200000-0xfc3fffff 64bit pref] Dec 16 09:36:54.884220 kernel: acpiphp: Slot [0-3] registered Dec 16 09:36:54.884329 kernel: pci 0000:00:03.0: PCI bridge to [bus 09] Dec 16 09:36:54.884442 kernel: pci 0000:00:03.0: bridge window [mem 0xfd800000-0xfd9fffff] Dec 16 09:36:54.884653 kernel: pci 0000:00:03.0: bridge window [mem 0xfc000000-0xfc1fffff 64bit pref] Dec 16 09:36:54.884671 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Dec 16 09:36:54.884679 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Dec 16 09:36:54.884686 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Dec 16 09:36:54.884692 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Dec 16 09:36:54.884698 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Dec 16 09:36:54.884704 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Dec 16 09:36:54.884710 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Dec 16 09:36:54.884721 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Dec 16 09:36:54.884727 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Dec 16 09:36:54.884733 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Dec 16 09:36:54.884739 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Dec 16 09:36:54.884745 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Dec 16 09:36:54.884752 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Dec 16 09:36:54.884758 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Dec 16 09:36:54.884764 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Dec 16 09:36:54.884770 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Dec 16 09:36:54.884779 kernel: iommu: Default domain type: Translated Dec 16 09:36:54.884786 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Dec 16 09:36:54.884792 kernel: PCI: Using ACPI for IRQ routing Dec 16 09:36:54.884798 kernel: PCI: pci_cache_line_size set to 64 bytes Dec 16 09:36:54.884804 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] Dec 16 09:36:54.884810 kernel: e820: reserve RAM buffer [mem 0x7cfdc000-0x7fffffff] Dec 16 09:36:54.884926 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Dec 16 09:36:54.885034 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Dec 16 09:36:54.885163 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Dec 16 09:36:54.885177 kernel: vgaarb: loaded Dec 16 09:36:54.885184 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Dec 16 09:36:54.885190 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Dec 16 09:36:54.885196 kernel: clocksource: Switched to clocksource kvm-clock Dec 16 09:36:54.885202 kernel: VFS: Disk quotas dquot_6.6.0 Dec 16 09:36:54.885209 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Dec 16 09:36:54.885215 kernel: pnp: PnP ACPI init Dec 16 09:36:54.885331 kernel: system 00:04: [mem 0xb0000000-0xbfffffff window] has been reserved Dec 16 09:36:54.885344 kernel: pnp: PnP ACPI: found 5 devices Dec 16 09:36:54.885352 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Dec 16 09:36:54.885358 kernel: NET: Registered PF_INET protocol family Dec 16 09:36:54.885364 kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear) Dec 16 09:36:54.885370 kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear) Dec 16 09:36:54.885377 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Dec 16 09:36:54.885383 kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear) Dec 16 09:36:54.885389 kernel: TCP bind hash table entries: 16384 (order: 7, 524288 bytes, linear) Dec 16 09:36:54.885396 kernel: TCP: Hash tables configured (established 16384 bind 16384) Dec 16 09:36:54.885404 kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear) Dec 16 09:36:54.885410 kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear) Dec 16 09:36:54.885416 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Dec 16 09:36:54.885422 kernel: NET: Registered PF_XDP protocol family Dec 16 09:36:54.885560 kernel: pci 0000:00:02.6: bridge window [io 0x1000-0x0fff] to [bus 07] add_size 1000 Dec 16 09:36:54.885700 kernel: pci 0000:00:02.7: bridge window [io 0x1000-0x0fff] to [bus 08] add_size 1000 Dec 16 09:36:54.885806 kernel: pci 0000:00:03.0: bridge window [io 0x1000-0x0fff] to [bus 09] add_size 1000 Dec 16 09:36:54.885920 kernel: pci 0000:00:02.6: BAR 13: assigned [io 0x1000-0x1fff] Dec 16 09:36:54.886029 kernel: pci 0000:00:02.7: BAR 13: assigned [io 0x2000-0x2fff] Dec 16 09:36:54.886136 kernel: pci 0000:00:03.0: BAR 13: assigned [io 0x3000-0x3fff] Dec 16 09:36:54.886242 kernel: pci 0000:00:02.0: PCI bridge to [bus 01] Dec 16 09:36:54.886351 kernel: pci 0000:00:02.0: bridge window [mem 0xfe800000-0xfe9fffff] Dec 16 09:36:54.886457 kernel: pci 0000:00:02.0: bridge window [mem 0xfd000000-0xfd1fffff 64bit pref] Dec 16 09:36:54.886616 kernel: pci 0000:00:02.1: PCI bridge to [bus 02] Dec 16 09:36:54.887642 kernel: pci 0000:00:02.1: bridge window [mem 0xfe600000-0xfe7fffff] Dec 16 09:36:54.887765 kernel: pci 0000:00:02.1: bridge window [mem 0xfce00000-0xfcffffff 64bit pref] Dec 16 09:36:54.887951 kernel: pci 0000:00:02.2: PCI bridge to [bus 03] Dec 16 09:36:54.888090 kernel: pci 0000:00:02.2: bridge window [mem 0xfe400000-0xfe5fffff] Dec 16 09:36:54.888249 kernel: pci 0000:00:02.2: bridge window [mem 0xfcc00000-0xfcdfffff 64bit pref] Dec 16 09:36:54.888354 kernel: pci 0000:00:02.3: PCI bridge to [bus 04] Dec 16 09:36:54.888456 kernel: pci 0000:00:02.3: bridge window [mem 0xfe200000-0xfe3fffff] Dec 16 09:36:54.889268 kernel: pci 0000:00:02.3: bridge window [mem 0xfca00000-0xfcbfffff 64bit pref] Dec 16 09:36:54.889389 kernel: pci 0000:00:02.4: PCI bridge to [bus 05] Dec 16 09:36:54.889530 kernel: pci 0000:00:02.4: bridge window [mem 0xfe000000-0xfe1fffff] Dec 16 09:36:54.890377 kernel: pci 0000:00:02.4: bridge window [mem 0xfc800000-0xfc9fffff 64bit pref] Dec 16 09:36:54.890499 kernel: pci 0000:00:02.5: PCI bridge to [bus 06] Dec 16 09:36:54.890698 kernel: pci 0000:00:02.5: bridge window [mem 0xfde00000-0xfdffffff] Dec 16 09:36:54.890810 kernel: pci 0000:00:02.5: bridge window [mem 0xfc600000-0xfc7fffff 64bit pref] Dec 16 09:36:54.890916 kernel: pci 0000:00:02.6: PCI bridge to [bus 07] Dec 16 09:36:54.891022 kernel: pci 0000:00:02.6: bridge window [io 0x1000-0x1fff] Dec 16 09:36:54.891178 kernel: pci 0000:00:02.6: bridge window [mem 0xfdc00000-0xfddfffff] Dec 16 09:36:54.891306 kernel: pci 0000:00:02.6: bridge window [mem 0xfc400000-0xfc5fffff 64bit pref] Dec 16 09:36:54.891421 kernel: pci 0000:00:02.7: PCI bridge to [bus 08] Dec 16 09:36:54.892629 kernel: pci 0000:00:02.7: bridge window [io 0x2000-0x2fff] Dec 16 09:36:54.892756 kernel: pci 0000:00:02.7: bridge window [mem 0xfda00000-0xfdbfffff] Dec 16 09:36:54.892862 kernel: pci 0000:00:02.7: bridge window [mem 0xfc200000-0xfc3fffff 64bit pref] Dec 16 09:36:54.892965 kernel: pci 0000:00:03.0: PCI bridge to [bus 09] Dec 16 09:36:54.893088 kernel: pci 0000:00:03.0: bridge window [io 0x3000-0x3fff] Dec 16 09:36:54.893194 kernel: pci 0000:00:03.0: bridge window [mem 0xfd800000-0xfd9fffff] Dec 16 09:36:54.893304 kernel: pci 0000:00:03.0: bridge window [mem 0xfc000000-0xfc1fffff 64bit pref] Dec 16 09:36:54.893402 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Dec 16 09:36:54.893506 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Dec 16 09:36:54.894689 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Dec 16 09:36:54.894799 kernel: pci_bus 0000:00: resource 7 [mem 0x7d000000-0xafffffff window] Dec 16 09:36:54.894894 kernel: pci_bus 0000:00: resource 8 [mem 0xc0000000-0xfebfffff window] Dec 16 09:36:54.894987 kernel: pci_bus 0000:00: resource 9 [mem 0x100000000-0x8ffffffff window] Dec 16 09:36:54.895098 kernel: pci_bus 0000:01: resource 1 [mem 0xfe800000-0xfe9fffff] Dec 16 09:36:54.895197 kernel: pci_bus 0000:01: resource 2 [mem 0xfd000000-0xfd1fffff 64bit pref] Dec 16 09:36:54.895301 kernel: pci_bus 0000:02: resource 1 [mem 0xfe600000-0xfe7fffff] Dec 16 09:36:54.895406 kernel: pci_bus 0000:02: resource 2 [mem 0xfce00000-0xfcffffff 64bit pref] Dec 16 09:36:54.895520 kernel: pci_bus 0000:03: resource 1 [mem 0xfe400000-0xfe5fffff] Dec 16 09:36:54.896709 kernel: pci_bus 0000:03: resource 2 [mem 0xfcc00000-0xfcdfffff 64bit pref] Dec 16 09:36:54.896827 kernel: pci_bus 0000:04: resource 1 [mem 0xfe200000-0xfe3fffff] Dec 16 09:36:54.896928 kernel: pci_bus 0000:04: resource 2 [mem 0xfca00000-0xfcbfffff 64bit pref] Dec 16 09:36:54.897033 kernel: pci_bus 0000:05: resource 1 [mem 0xfe000000-0xfe1fffff] Dec 16 09:36:54.897161 kernel: pci_bus 0000:05: resource 2 [mem 0xfc800000-0xfc9fffff 64bit pref] Dec 16 09:36:54.897318 kernel: pci_bus 0000:06: resource 1 [mem 0xfde00000-0xfdffffff] Dec 16 09:36:54.897464 kernel: pci_bus 0000:06: resource 2 [mem 0xfc600000-0xfc7fffff 64bit pref] Dec 16 09:36:54.897637 kernel: pci_bus 0000:07: resource 0 [io 0x1000-0x1fff] Dec 16 09:36:54.897751 kernel: pci_bus 0000:07: resource 1 [mem 0xfdc00000-0xfddfffff] Dec 16 09:36:54.897855 kernel: pci_bus 0000:07: resource 2 [mem 0xfc400000-0xfc5fffff 64bit pref] Dec 16 09:36:54.897965 kernel: pci_bus 0000:08: resource 0 [io 0x2000-0x2fff] Dec 16 09:36:54.898074 kernel: pci_bus 0000:08: resource 1 [mem 0xfda00000-0xfdbfffff] Dec 16 09:36:54.898174 kernel: pci_bus 0000:08: resource 2 [mem 0xfc200000-0xfc3fffff 64bit pref] Dec 16 09:36:54.898287 kernel: pci_bus 0000:09: resource 0 [io 0x3000-0x3fff] Dec 16 09:36:54.898390 kernel: pci_bus 0000:09: resource 1 [mem 0xfd800000-0xfd9fffff] Dec 16 09:36:54.898501 kernel: pci_bus 0000:09: resource 2 [mem 0xfc000000-0xfc1fffff 64bit pref] Dec 16 09:36:54.898520 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Dec 16 09:36:54.898533 kernel: PCI: CLS 0 bytes, default 64 Dec 16 09:36:54.898547 kernel: Initialise system trusted keyrings Dec 16 09:36:54.898553 kernel: workingset: timestamp_bits=39 max_order=19 bucket_order=0 Dec 16 09:36:54.898560 kernel: Key type asymmetric registered Dec 16 09:36:54.900629 kernel: Asymmetric key parser 'x509' registered Dec 16 09:36:54.900637 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Dec 16 09:36:54.900644 kernel: io scheduler mq-deadline registered Dec 16 09:36:54.900651 kernel: io scheduler kyber registered Dec 16 09:36:54.900657 kernel: io scheduler bfq registered Dec 16 09:36:54.900782 kernel: pcieport 0000:00:02.0: PME: Signaling with IRQ 24 Dec 16 09:36:54.900897 kernel: pcieport 0000:00:02.0: AER: enabled with IRQ 24 Dec 16 09:36:54.901001 kernel: pcieport 0000:00:02.1: PME: Signaling with IRQ 25 Dec 16 09:36:54.901120 kernel: pcieport 0000:00:02.1: AER: enabled with IRQ 25 Dec 16 09:36:54.901224 kernel: pcieport 0000:00:02.2: PME: Signaling with IRQ 26 Dec 16 09:36:54.901326 kernel: pcieport 0000:00:02.2: AER: enabled with IRQ 26 Dec 16 09:36:54.901428 kernel: pcieport 0000:00:02.3: PME: Signaling with IRQ 27 Dec 16 09:36:54.901550 kernel: pcieport 0000:00:02.3: AER: enabled with IRQ 27 Dec 16 09:36:54.901686 kernel: pcieport 0000:00:02.4: PME: Signaling with IRQ 28 Dec 16 09:36:54.901792 kernel: pcieport 0000:00:02.4: AER: enabled with IRQ 28 Dec 16 09:36:54.901901 kernel: pcieport 0000:00:02.5: PME: Signaling with IRQ 29 Dec 16 09:36:54.902003 kernel: pcieport 0000:00:02.5: AER: enabled with IRQ 29 Dec 16 09:36:54.902104 kernel: pcieport 0000:00:02.6: PME: Signaling with IRQ 30 Dec 16 09:36:54.902205 kernel: pcieport 0000:00:02.6: AER: enabled with IRQ 30 Dec 16 09:36:54.902307 kernel: pcieport 0000:00:02.7: PME: Signaling with IRQ 31 Dec 16 09:36:54.902410 kernel: pcieport 0000:00:02.7: AER: enabled with IRQ 31 Dec 16 09:36:54.902420 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Dec 16 09:36:54.902549 kernel: pcieport 0000:00:03.0: PME: Signaling with IRQ 32 Dec 16 09:36:54.904712 kernel: pcieport 0000:00:03.0: AER: enabled with IRQ 32 Dec 16 09:36:54.904725 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Dec 16 09:36:54.904737 kernel: ACPI: \_SB_.GSIF: Enabled at IRQ 21 Dec 16 09:36:54.904744 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Dec 16 09:36:54.904750 kernel: 00:00: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Dec 16 09:36:54.904757 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Dec 16 09:36:54.904764 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Dec 16 09:36:54.904770 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Dec 16 09:36:54.904780 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Dec 16 09:36:54.904894 kernel: rtc_cmos 00:03: RTC can wake from S4 Dec 16 09:36:54.904994 kernel: rtc_cmos 00:03: registered as rtc0 Dec 16 09:36:54.905115 kernel: rtc_cmos 00:03: setting system clock to 2024-12-16T09:36:54 UTC (1734341814) Dec 16 09:36:54.905246 kernel: rtc_cmos 00:03: alarms up to one day, y3k, 242 bytes nvram, hpet irqs Dec 16 09:36:54.905257 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Dec 16 09:36:54.905264 kernel: NET: Registered PF_INET6 protocol family Dec 16 09:36:54.905270 kernel: Segment Routing with IPv6 Dec 16 09:36:54.905281 kernel: In-situ OAM (IOAM) with IPv6 Dec 16 09:36:54.905287 kernel: NET: Registered PF_PACKET protocol family Dec 16 09:36:54.905294 kernel: Key type dns_resolver registered Dec 16 09:36:54.905300 kernel: IPI shorthand broadcast: enabled Dec 16 09:36:54.905307 kernel: sched_clock: Marking stable (1044006975, 131070647)->(1182051890, -6974268) Dec 16 09:36:54.905314 kernel: registered taskstats version 1 Dec 16 09:36:54.905320 kernel: Loading compiled-in X.509 certificates Dec 16 09:36:54.905327 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.65-flatcar: c82d546f528d79a5758dcebbc47fb6daf92836a0' Dec 16 09:36:54.905333 kernel: Key type .fscrypt registered Dec 16 09:36:54.905340 kernel: Key type fscrypt-provisioning registered Dec 16 09:36:54.905349 kernel: ima: No TPM chip found, activating TPM-bypass! Dec 16 09:36:54.905355 kernel: ima: Allocated hash algorithm: sha1 Dec 16 09:36:54.905362 kernel: ima: No architecture policies found Dec 16 09:36:54.905369 kernel: clk: Disabling unused clocks Dec 16 09:36:54.905375 kernel: Freeing unused kernel image (initmem) memory: 42844K Dec 16 09:36:54.905382 kernel: Write protecting the kernel read-only data: 36864k Dec 16 09:36:54.905388 kernel: Freeing unused kernel image (rodata/data gap) memory: 1852K Dec 16 09:36:54.905395 kernel: Run /init as init process Dec 16 09:36:54.905403 kernel: with arguments: Dec 16 09:36:54.905410 kernel: /init Dec 16 09:36:54.905416 kernel: with environment: Dec 16 09:36:54.905423 kernel: HOME=/ Dec 16 09:36:54.905429 kernel: TERM=linux Dec 16 09:36:54.905436 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Dec 16 09:36:54.905444 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Dec 16 09:36:54.905453 systemd[1]: Detected virtualization kvm. Dec 16 09:36:54.905463 systemd[1]: Detected architecture x86-64. Dec 16 09:36:54.905469 systemd[1]: Running in initrd. Dec 16 09:36:54.905481 systemd[1]: No hostname configured, using default hostname. Dec 16 09:36:54.905494 systemd[1]: Hostname set to . Dec 16 09:36:54.905508 systemd[1]: Initializing machine ID from VM UUID. Dec 16 09:36:54.905520 systemd[1]: Queued start job for default target initrd.target. Dec 16 09:36:54.905533 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Dec 16 09:36:54.905543 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Dec 16 09:36:54.905553 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Dec 16 09:36:54.906588 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Dec 16 09:36:54.906601 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Dec 16 09:36:54.906608 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Dec 16 09:36:54.906617 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Dec 16 09:36:54.906624 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Dec 16 09:36:54.906631 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Dec 16 09:36:54.906642 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Dec 16 09:36:54.906649 systemd[1]: Reached target paths.target - Path Units. Dec 16 09:36:54.906656 systemd[1]: Reached target slices.target - Slice Units. Dec 16 09:36:54.906663 systemd[1]: Reached target swap.target - Swaps. Dec 16 09:36:54.906670 systemd[1]: Reached target timers.target - Timer Units. Dec 16 09:36:54.906676 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Dec 16 09:36:54.906683 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Dec 16 09:36:54.906690 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Dec 16 09:36:54.906700 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Dec 16 09:36:54.906707 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Dec 16 09:36:54.906714 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Dec 16 09:36:54.906721 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Dec 16 09:36:54.906728 systemd[1]: Reached target sockets.target - Socket Units. Dec 16 09:36:54.906735 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Dec 16 09:36:54.906742 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Dec 16 09:36:54.906749 systemd[1]: Finished network-cleanup.service - Network Cleanup. Dec 16 09:36:54.906756 systemd[1]: Starting systemd-fsck-usr.service... Dec 16 09:36:54.906765 systemd[1]: Starting systemd-journald.service - Journal Service... Dec 16 09:36:54.906772 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Dec 16 09:36:54.906779 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Dec 16 09:36:54.906807 systemd-journald[187]: Collecting audit messages is disabled. Dec 16 09:36:54.906827 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Dec 16 09:36:54.906834 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Dec 16 09:36:54.906841 systemd[1]: Finished systemd-fsck-usr.service. Dec 16 09:36:54.906849 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Dec 16 09:36:54.906858 systemd-journald[187]: Journal started Dec 16 09:36:54.906874 systemd-journald[187]: Runtime Journal (/run/log/journal/df949b230dfa435a85dd75356b28f8be) is 4.8M, max 38.4M, 33.6M free. Dec 16 09:36:54.888222 systemd-modules-load[188]: Inserted module 'overlay' Dec 16 09:36:54.944380 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Dec 16 09:36:54.944400 kernel: Bridge firewalling registered Dec 16 09:36:54.944410 systemd[1]: Started systemd-journald.service - Journal Service. Dec 16 09:36:54.920209 systemd-modules-load[188]: Inserted module 'br_netfilter' Dec 16 09:36:54.949984 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Dec 16 09:36:54.951359 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Dec 16 09:36:54.968734 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Dec 16 09:36:54.970699 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Dec 16 09:36:54.972175 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Dec 16 09:36:54.975615 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Dec 16 09:36:54.983406 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Dec 16 09:36:54.992986 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Dec 16 09:36:54.994362 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Dec 16 09:36:55.000703 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Dec 16 09:36:55.001809 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Dec 16 09:36:55.004183 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Dec 16 09:36:55.012060 dracut-cmdline[219]: dracut-dracut-053 Dec 16 09:36:55.012728 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Dec 16 09:36:55.014734 dracut-cmdline[219]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=hetzner verity.usrhash=2fdbba50b59d8c8a9877a81151806ddc16f473fe99b9ba0d8825997d654583ff Dec 16 09:36:55.039291 systemd-resolved[226]: Positive Trust Anchors: Dec 16 09:36:55.039303 systemd-resolved[226]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Dec 16 09:36:55.039327 systemd-resolved[226]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Dec 16 09:36:55.045190 systemd-resolved[226]: Defaulting to hostname 'linux'. Dec 16 09:36:55.046857 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Dec 16 09:36:55.049472 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Dec 16 09:36:55.079605 kernel: SCSI subsystem initialized Dec 16 09:36:55.087587 kernel: Loading iSCSI transport class v2.0-870. Dec 16 09:36:55.096586 kernel: iscsi: registered transport (tcp) Dec 16 09:36:55.114914 kernel: iscsi: registered transport (qla4xxx) Dec 16 09:36:55.114967 kernel: QLogic iSCSI HBA Driver Dec 16 09:36:55.155483 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Dec 16 09:36:55.160711 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Dec 16 09:36:55.182745 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Dec 16 09:36:55.182777 kernel: device-mapper: uevent: version 1.0.3 Dec 16 09:36:55.184305 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Dec 16 09:36:55.223591 kernel: raid6: avx2x4 gen() 35233 MB/s Dec 16 09:36:55.238581 kernel: raid6: avx2x2 gen() 32291 MB/s Dec 16 09:36:55.255645 kernel: raid6: avx2x1 gen() 27036 MB/s Dec 16 09:36:55.255672 kernel: raid6: using algorithm avx2x4 gen() 35233 MB/s Dec 16 09:36:55.273746 kernel: raid6: .... xor() 4311 MB/s, rmw enabled Dec 16 09:36:55.273776 kernel: raid6: using avx2x2 recovery algorithm Dec 16 09:36:55.293600 kernel: xor: automatically using best checksumming function avx Dec 16 09:36:55.419604 kernel: Btrfs loaded, zoned=no, fsverity=no Dec 16 09:36:55.430067 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Dec 16 09:36:55.435695 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Dec 16 09:36:55.447459 systemd-udevd[405]: Using default interface naming scheme 'v255'. Dec 16 09:36:55.451176 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Dec 16 09:36:55.458718 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Dec 16 09:36:55.472366 dracut-pre-trigger[413]: rd.md=0: removing MD RAID activation Dec 16 09:36:55.498682 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Dec 16 09:36:55.505695 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Dec 16 09:36:55.566790 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Dec 16 09:36:55.572696 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Dec 16 09:36:55.596095 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Dec 16 09:36:55.597870 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Dec 16 09:36:55.598316 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Dec 16 09:36:55.600985 systemd[1]: Reached target remote-fs.target - Remote File Systems. Dec 16 09:36:55.606719 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Dec 16 09:36:55.625986 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Dec 16 09:36:55.661622 kernel: cryptd: max_cpu_qlen set to 1000 Dec 16 09:36:55.671625 kernel: scsi host0: Virtio SCSI HBA Dec 16 09:36:55.674462 kernel: AVX2 version of gcm_enc/dec engaged. Dec 16 09:36:55.674485 kernel: AES CTR mode by8 optimization enabled Dec 16 09:36:55.693618 kernel: scsi 0:0:0:0: Direct-Access QEMU QEMU HARDDISK 2.5+ PQ: 0 ANSI: 5 Dec 16 09:36:55.694692 kernel: libata version 3.00 loaded. Dec 16 09:36:55.739093 kernel: ACPI: bus type USB registered Dec 16 09:36:55.739144 kernel: usbcore: registered new interface driver usbfs Dec 16 09:36:55.742344 kernel: usbcore: registered new interface driver hub Dec 16 09:36:55.742398 kernel: usbcore: registered new device driver usb Dec 16 09:36:55.757421 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Dec 16 09:36:55.758200 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Dec 16 09:36:55.759345 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Dec 16 09:36:55.762775 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Dec 16 09:36:55.762831 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Dec 16 09:36:55.763287 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Dec 16 09:36:55.771850 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Dec 16 09:36:55.783673 kernel: ahci 0000:00:1f.2: version 3.0 Dec 16 09:36:55.799812 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Dec 16 09:36:55.799827 kernel: ahci 0000:00:1f.2: AHCI 0001.0000 32 slots 6 ports 1.5 Gbps 0x3f impl SATA mode Dec 16 09:36:55.800814 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Dec 16 09:36:55.800947 kernel: scsi host1: ahci Dec 16 09:36:55.801109 kernel: sd 0:0:0:0: Power-on or device reset occurred Dec 16 09:36:55.801265 kernel: scsi host2: ahci Dec 16 09:36:55.801397 kernel: sd 0:0:0:0: [sda] 80003072 512-byte logical blocks: (41.0 GB/38.1 GiB) Dec 16 09:36:55.801640 kernel: sd 0:0:0:0: [sda] Write Protect is off Dec 16 09:36:55.801835 kernel: sd 0:0:0:0: [sda] Mode Sense: 63 00 00 08 Dec 16 09:36:55.801967 kernel: scsi host3: ahci Dec 16 09:36:55.802091 kernel: sd 0:0:0:0: [sda] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA Dec 16 09:36:55.802221 kernel: scsi host4: ahci Dec 16 09:36:55.802346 kernel: scsi host5: ahci Dec 16 09:36:55.802469 kernel: scsi host6: ahci Dec 16 09:36:55.802614 kernel: ata1: SATA max UDMA/133 abar m4096@0xfea1a000 port 0xfea1a100 irq 46 Dec 16 09:36:55.802625 kernel: ata2: SATA max UDMA/133 abar m4096@0xfea1a000 port 0xfea1a180 irq 46 Dec 16 09:36:55.802633 kernel: ata3: SATA max UDMA/133 abar m4096@0xfea1a000 port 0xfea1a200 irq 46 Dec 16 09:36:55.802641 kernel: ata4: SATA max UDMA/133 abar m4096@0xfea1a000 port 0xfea1a280 irq 46 Dec 16 09:36:55.802648 kernel: ata5: SATA max UDMA/133 abar m4096@0xfea1a000 port 0xfea1a300 irq 46 Dec 16 09:36:55.802656 kernel: ata6: SATA max UDMA/133 abar m4096@0xfea1a000 port 0xfea1a380 irq 46 Dec 16 09:36:55.802668 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Dec 16 09:36:55.802677 kernel: GPT:17805311 != 80003071 Dec 16 09:36:55.802684 kernel: GPT:Alternate GPT header not at the end of the disk. Dec 16 09:36:55.802692 kernel: GPT:17805311 != 80003071 Dec 16 09:36:55.802700 kernel: GPT: Use GNU Parted to correct GPT errors. Dec 16 09:36:55.802708 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Dec 16 09:36:55.802716 kernel: sd 0:0:0:0: [sda] Attached SCSI disk Dec 16 09:36:55.853905 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Dec 16 09:36:55.858711 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Dec 16 09:36:55.877346 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Dec 16 09:36:56.114964 kernel: ata3: SATA link down (SStatus 0 SControl 300) Dec 16 09:36:56.115038 kernel: ata2: SATA link down (SStatus 0 SControl 300) Dec 16 09:36:56.115061 kernel: ata1: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Dec 16 09:36:56.115070 kernel: ata6: SATA link down (SStatus 0 SControl 300) Dec 16 09:36:56.115079 kernel: ata5: SATA link down (SStatus 0 SControl 300) Dec 16 09:36:56.115087 kernel: ata4: SATA link down (SStatus 0 SControl 300) Dec 16 09:36:56.116589 kernel: ata1.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Dec 16 09:36:56.117681 kernel: ata1.00: applying bridge limits Dec 16 09:36:56.119099 kernel: ata1.00: configured for UDMA/100 Dec 16 09:36:56.119784 kernel: scsi 1:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Dec 16 09:36:56.146712 kernel: xhci_hcd 0000:02:00.0: xHCI Host Controller Dec 16 09:36:56.178724 kernel: xhci_hcd 0000:02:00.0: new USB bus registered, assigned bus number 1 Dec 16 09:36:56.178883 kernel: xhci_hcd 0000:02:00.0: hcc params 0x00087001 hci version 0x100 quirks 0x0000000000000010 Dec 16 09:36:56.179013 kernel: xhci_hcd 0000:02:00.0: xHCI Host Controller Dec 16 09:36:56.179135 kernel: xhci_hcd 0000:02:00.0: new USB bus registered, assigned bus number 2 Dec 16 09:36:56.179257 kernel: xhci_hcd 0000:02:00.0: Host supports USB 3.0 SuperSpeed Dec 16 09:36:56.179377 kernel: hub 1-0:1.0: USB hub found Dec 16 09:36:56.179529 kernel: hub 1-0:1.0: 4 ports detected Dec 16 09:36:56.179690 kernel: usb usb2: We don't know the algorithms for LPM for this host, disabling LPM. Dec 16 09:36:56.179831 kernel: sr 1:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Dec 16 09:36:56.183651 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Dec 16 09:36:56.183663 kernel: hub 2-0:1.0: USB hub found Dec 16 09:36:56.183812 kernel: hub 2-0:1.0: 4 ports detected Dec 16 09:36:56.183939 kernel: sr 1:0:0:0: Attached scsi CD-ROM sr0 Dec 16 09:36:56.190466 kernel: BTRFS: device fsid c3b72f8a-27ca-4d37-9d0e-1ec3c4bdc3be devid 1 transid 41 /dev/sda3 scanned by (udev-worker) (451) Dec 16 09:36:56.189992 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - QEMU_HARDDISK EFI-SYSTEM. Dec 16 09:36:56.195143 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - QEMU_HARDDISK ROOT. Dec 16 09:36:56.197628 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/sda6 scanned by (udev-worker) (450) Dec 16 09:36:56.208502 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - QEMU_HARDDISK OEM. Dec 16 09:36:56.212863 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - QEMU_HARDDISK USR-A. Dec 16 09:36:56.213404 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - QEMU_HARDDISK USR-A. Dec 16 09:36:56.217706 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Dec 16 09:36:56.230177 disk-uuid[575]: Primary Header is updated. Dec 16 09:36:56.230177 disk-uuid[575]: Secondary Entries is updated. Dec 16 09:36:56.230177 disk-uuid[575]: Secondary Header is updated. Dec 16 09:36:56.232736 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Dec 16 09:36:56.410599 kernel: usb 1-1: new high-speed USB device number 2 using xhci_hcd Dec 16 09:36:56.555622 kernel: hid: raw HID events driver (C) Jiri Kosina Dec 16 09:36:56.564457 kernel: usbcore: registered new interface driver usbhid Dec 16 09:36:56.564769 kernel: usbhid: USB HID core driver Dec 16 09:36:56.573610 kernel: input: QEMU QEMU USB Tablet as /devices/pci0000:00/0000:00:02.1/0000:02:00.0/usb1/1-1/1-1:1.0/0003:0627:0001.0001/input/input2 Dec 16 09:36:56.578687 kernel: hid-generic 0003:0627:0001.0001: input,hidraw0: USB HID v0.01 Mouse [QEMU QEMU USB Tablet] on usb-0000:02:00.0-1/input0 Dec 16 09:36:57.245598 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Dec 16 09:36:57.248757 disk-uuid[576]: The operation has completed successfully. Dec 16 09:36:57.301758 systemd[1]: disk-uuid.service: Deactivated successfully. Dec 16 09:36:57.301868 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Dec 16 09:36:57.314794 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Dec 16 09:36:57.318455 sh[596]: Success Dec 16 09:36:57.329653 kernel: device-mapper: verity: sha256 using implementation "sha256-ni" Dec 16 09:36:57.374028 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Dec 16 09:36:57.381661 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Dec 16 09:36:57.382791 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Dec 16 09:36:57.397688 kernel: BTRFS info (device dm-0): first mount of filesystem c3b72f8a-27ca-4d37-9d0e-1ec3c4bdc3be Dec 16 09:36:57.397724 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Dec 16 09:36:57.400343 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Dec 16 09:36:57.400359 kernel: BTRFS info (device dm-0): disabling log replay at mount time Dec 16 09:36:57.401833 kernel: BTRFS info (device dm-0): using free space tree Dec 16 09:36:57.409587 kernel: BTRFS info (device dm-0): enabling ssd optimizations Dec 16 09:36:57.410696 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Dec 16 09:36:57.411598 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Dec 16 09:36:57.416706 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Dec 16 09:36:57.420690 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Dec 16 09:36:57.430594 kernel: BTRFS info (device sda6): first mount of filesystem db063747-cac8-4176-8963-c216c1b11dcb Dec 16 09:36:57.434003 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Dec 16 09:36:57.434037 kernel: BTRFS info (device sda6): using free space tree Dec 16 09:36:57.440516 kernel: BTRFS info (device sda6): enabling ssd optimizations Dec 16 09:36:57.440549 kernel: BTRFS info (device sda6): auto enabling async discard Dec 16 09:36:57.451585 kernel: BTRFS info (device sda6): last unmount of filesystem db063747-cac8-4176-8963-c216c1b11dcb Dec 16 09:36:57.451823 systemd[1]: mnt-oem.mount: Deactivated successfully. Dec 16 09:36:57.457024 systemd[1]: Finished ignition-setup.service - Ignition (setup). Dec 16 09:36:57.463707 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Dec 16 09:36:57.527759 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Dec 16 09:36:57.535752 systemd[1]: Starting systemd-networkd.service - Network Configuration... Dec 16 09:36:57.537040 ignition[708]: Ignition 2.19.0 Dec 16 09:36:57.537632 ignition[708]: Stage: fetch-offline Dec 16 09:36:57.537682 ignition[708]: no configs at "/usr/lib/ignition/base.d" Dec 16 09:36:57.537698 ignition[708]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Dec 16 09:36:57.541677 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Dec 16 09:36:57.537797 ignition[708]: parsed url from cmdline: "" Dec 16 09:36:57.537803 ignition[708]: no config URL provided Dec 16 09:36:57.537809 ignition[708]: reading system config file "/usr/lib/ignition/user.ign" Dec 16 09:36:57.537819 ignition[708]: no config at "/usr/lib/ignition/user.ign" Dec 16 09:36:57.537824 ignition[708]: failed to fetch config: resource requires networking Dec 16 09:36:57.538007 ignition[708]: Ignition finished successfully Dec 16 09:36:57.556866 systemd-networkd[781]: lo: Link UP Dec 16 09:36:57.556875 systemd-networkd[781]: lo: Gained carrier Dec 16 09:36:57.559377 systemd-networkd[781]: Enumeration completed Dec 16 09:36:57.559700 systemd[1]: Started systemd-networkd.service - Network Configuration. Dec 16 09:36:57.560370 systemd-networkd[781]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 16 09:36:57.560374 systemd-networkd[781]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Dec 16 09:36:57.561328 systemd[1]: Reached target network.target - Network. Dec 16 09:36:57.561932 systemd-networkd[781]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 16 09:36:57.561937 systemd-networkd[781]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. Dec 16 09:36:57.562754 systemd-networkd[781]: eth0: Link UP Dec 16 09:36:57.562757 systemd-networkd[781]: eth0: Gained carrier Dec 16 09:36:57.562764 systemd-networkd[781]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 16 09:36:57.567953 systemd-networkd[781]: eth1: Link UP Dec 16 09:36:57.567957 systemd-networkd[781]: eth1: Gained carrier Dec 16 09:36:57.567965 systemd-networkd[781]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 16 09:36:57.569690 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Dec 16 09:36:57.580309 ignition[785]: Ignition 2.19.0 Dec 16 09:36:57.580320 ignition[785]: Stage: fetch Dec 16 09:36:57.580446 ignition[785]: no configs at "/usr/lib/ignition/base.d" Dec 16 09:36:57.580456 ignition[785]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Dec 16 09:36:57.580530 ignition[785]: parsed url from cmdline: "" Dec 16 09:36:57.580534 ignition[785]: no config URL provided Dec 16 09:36:57.580539 ignition[785]: reading system config file "/usr/lib/ignition/user.ign" Dec 16 09:36:57.580547 ignition[785]: no config at "/usr/lib/ignition/user.ign" Dec 16 09:36:57.580596 ignition[785]: GET http://169.254.169.254/hetzner/v1/userdata: attempt #1 Dec 16 09:36:57.580725 ignition[785]: GET error: Get "http://169.254.169.254/hetzner/v1/userdata": dial tcp 169.254.169.254:80: connect: network is unreachable Dec 16 09:36:57.608614 systemd-networkd[781]: eth1: DHCPv4 address 10.0.0.3/32, gateway 10.0.0.1 acquired from 10.0.0.1 Dec 16 09:36:57.698635 systemd-networkd[781]: eth0: DHCPv4 address 138.199.145.148/32, gateway 172.31.1.1 acquired from 172.31.1.1 Dec 16 09:36:57.781181 ignition[785]: GET http://169.254.169.254/hetzner/v1/userdata: attempt #2 Dec 16 09:36:57.786557 ignition[785]: GET result: OK Dec 16 09:36:57.786629 ignition[785]: parsing config with SHA512: 251c27ce2416b9f3b41a5b0434e13ecac23c20012035c0b1a613509149ba308ce78c812a01a8832354346fb07d669f31a16616fa987764220ba4f3ce23273261 Dec 16 09:36:57.789835 unknown[785]: fetched base config from "system" Dec 16 09:36:57.789847 unknown[785]: fetched base config from "system" Dec 16 09:36:57.790247 ignition[785]: fetch: fetch complete Dec 16 09:36:57.789857 unknown[785]: fetched user config from "hetzner" Dec 16 09:36:57.790253 ignition[785]: fetch: fetch passed Dec 16 09:36:57.790297 ignition[785]: Ignition finished successfully Dec 16 09:36:57.792976 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Dec 16 09:36:57.802826 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Dec 16 09:36:57.817605 ignition[793]: Ignition 2.19.0 Dec 16 09:36:57.817618 ignition[793]: Stage: kargs Dec 16 09:36:57.817801 ignition[793]: no configs at "/usr/lib/ignition/base.d" Dec 16 09:36:57.817813 ignition[793]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Dec 16 09:36:57.818833 ignition[793]: kargs: kargs passed Dec 16 09:36:57.818885 ignition[793]: Ignition finished successfully Dec 16 09:36:57.820680 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Dec 16 09:36:57.839729 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Dec 16 09:36:57.853983 ignition[800]: Ignition 2.19.0 Dec 16 09:36:57.854002 ignition[800]: Stage: disks Dec 16 09:36:57.854234 ignition[800]: no configs at "/usr/lib/ignition/base.d" Dec 16 09:36:57.854249 ignition[800]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Dec 16 09:36:57.855382 ignition[800]: disks: disks passed Dec 16 09:36:57.857402 systemd[1]: Finished ignition-disks.service - Ignition (disks). Dec 16 09:36:57.855437 ignition[800]: Ignition finished successfully Dec 16 09:36:57.858815 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Dec 16 09:36:57.859941 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Dec 16 09:36:57.861272 systemd[1]: Reached target local-fs.target - Local File Systems. Dec 16 09:36:57.862530 systemd[1]: Reached target sysinit.target - System Initialization. Dec 16 09:36:57.863666 systemd[1]: Reached target basic.target - Basic System. Dec 16 09:36:57.873771 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Dec 16 09:36:57.890885 systemd-fsck[809]: ROOT: clean, 14/1628000 files, 120691/1617920 blocks Dec 16 09:36:57.892872 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Dec 16 09:36:57.897645 systemd[1]: Mounting sysroot.mount - /sysroot... Dec 16 09:36:57.978597 kernel: EXT4-fs (sda9): mounted filesystem 390119fa-ab9c-4f50-b046-3b5c76c46193 r/w with ordered data mode. Quota mode: none. Dec 16 09:36:57.979041 systemd[1]: Mounted sysroot.mount - /sysroot. Dec 16 09:36:57.979990 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Dec 16 09:36:57.985625 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Dec 16 09:36:57.987648 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Dec 16 09:36:57.989695 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Dec 16 09:36:57.990939 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Dec 16 09:36:57.990962 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Dec 16 09:36:57.997597 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/sda6 scanned by mount (817) Dec 16 09:36:57.998154 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Dec 16 09:36:58.003310 kernel: BTRFS info (device sda6): first mount of filesystem db063747-cac8-4176-8963-c216c1b11dcb Dec 16 09:36:58.003332 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Dec 16 09:36:58.003342 kernel: BTRFS info (device sda6): using free space tree Dec 16 09:36:58.005421 kernel: BTRFS info (device sda6): enabling ssd optimizations Dec 16 09:36:58.005482 kernel: BTRFS info (device sda6): auto enabling async discard Dec 16 09:36:58.007477 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Dec 16 09:36:58.009970 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Dec 16 09:36:58.052653 coreos-metadata[819]: Dec 16 09:36:58.052 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/hostname: Attempt #1 Dec 16 09:36:58.053904 coreos-metadata[819]: Dec 16 09:36:58.053 INFO Fetch successful Dec 16 09:36:58.055515 coreos-metadata[819]: Dec 16 09:36:58.054 INFO wrote hostname ci-4081-2-1-7-459a6b42ef to /sysroot/etc/hostname Dec 16 09:36:58.057678 initrd-setup-root[845]: cut: /sysroot/etc/passwd: No such file or directory Dec 16 09:36:58.056240 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Dec 16 09:36:58.062113 initrd-setup-root[853]: cut: /sysroot/etc/group: No such file or directory Dec 16 09:36:58.067333 initrd-setup-root[860]: cut: /sysroot/etc/shadow: No such file or directory Dec 16 09:36:58.073639 initrd-setup-root[867]: cut: /sysroot/etc/gshadow: No such file or directory Dec 16 09:36:58.177645 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Dec 16 09:36:58.183676 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Dec 16 09:36:58.187714 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Dec 16 09:36:58.195604 kernel: BTRFS info (device sda6): last unmount of filesystem db063747-cac8-4176-8963-c216c1b11dcb Dec 16 09:36:58.217480 ignition[934]: INFO : Ignition 2.19.0 Dec 16 09:36:58.219164 ignition[934]: INFO : Stage: mount Dec 16 09:36:58.219164 ignition[934]: INFO : no configs at "/usr/lib/ignition/base.d" Dec 16 09:36:58.219164 ignition[934]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Dec 16 09:36:58.219164 ignition[934]: INFO : mount: mount passed Dec 16 09:36:58.219164 ignition[934]: INFO : Ignition finished successfully Dec 16 09:36:58.220270 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Dec 16 09:36:58.221179 systemd[1]: Finished ignition-mount.service - Ignition (mount). Dec 16 09:36:58.228676 systemd[1]: Starting ignition-files.service - Ignition (files)... Dec 16 09:36:58.397141 systemd[1]: sysroot-oem.mount: Deactivated successfully. Dec 16 09:36:58.403765 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Dec 16 09:36:58.413926 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/sda6 scanned by mount (948) Dec 16 09:36:58.413965 kernel: BTRFS info (device sda6): first mount of filesystem db063747-cac8-4176-8963-c216c1b11dcb Dec 16 09:36:58.415598 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Dec 16 09:36:58.417659 kernel: BTRFS info (device sda6): using free space tree Dec 16 09:36:58.422176 kernel: BTRFS info (device sda6): enabling ssd optimizations Dec 16 09:36:58.422215 kernel: BTRFS info (device sda6): auto enabling async discard Dec 16 09:36:58.424473 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Dec 16 09:36:58.444504 ignition[964]: INFO : Ignition 2.19.0 Dec 16 09:36:58.444504 ignition[964]: INFO : Stage: files Dec 16 09:36:58.445719 ignition[964]: INFO : no configs at "/usr/lib/ignition/base.d" Dec 16 09:36:58.445719 ignition[964]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Dec 16 09:36:58.445719 ignition[964]: DEBUG : files: compiled without relabeling support, skipping Dec 16 09:36:58.447652 ignition[964]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Dec 16 09:36:58.447652 ignition[964]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Dec 16 09:36:58.449518 ignition[964]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Dec 16 09:36:58.450317 ignition[964]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Dec 16 09:36:58.450317 ignition[964]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Dec 16 09:36:58.449994 unknown[964]: wrote ssh authorized keys file for user: core Dec 16 09:36:58.452468 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Dec 16 09:36:58.452468 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-amd64.tar.gz: attempt #1 Dec 16 09:36:58.551446 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Dec 16 09:36:58.782756 systemd-networkd[781]: eth1: Gained IPv6LL Dec 16 09:36:58.858042 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Dec 16 09:36:58.858042 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Dec 16 09:36:58.859754 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-amd64.tar.gz: attempt #1 Dec 16 09:36:59.038763 systemd-networkd[781]: eth0: Gained IPv6LL Dec 16 09:36:59.397528 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Dec 16 09:36:59.485683 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Dec 16 09:36:59.488159 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Dec 16 09:36:59.488159 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Dec 16 09:36:59.488159 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Dec 16 09:36:59.488159 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Dec 16 09:36:59.488159 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Dec 16 09:36:59.488159 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Dec 16 09:36:59.488159 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Dec 16 09:36:59.488159 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Dec 16 09:36:59.488159 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Dec 16 09:36:59.488159 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Dec 16 09:36:59.488159 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.0-x86-64.raw" Dec 16 09:36:59.488159 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.0-x86-64.raw" Dec 16 09:36:59.488159 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.0-x86-64.raw" Dec 16 09:36:59.488159 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.31.0-x86-64.raw: attempt #1 Dec 16 09:37:00.013698 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Dec 16 09:37:00.266577 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.0-x86-64.raw" Dec 16 09:37:00.266577 ignition[964]: INFO : files: op(c): [started] processing unit "prepare-helm.service" Dec 16 09:37:00.268163 ignition[964]: INFO : files: op(c): op(d): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Dec 16 09:37:00.268163 ignition[964]: INFO : files: op(c): op(d): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Dec 16 09:37:00.268163 ignition[964]: INFO : files: op(c): [finished] processing unit "prepare-helm.service" Dec 16 09:37:00.268163 ignition[964]: INFO : files: op(e): [started] processing unit "coreos-metadata.service" Dec 16 09:37:00.268163 ignition[964]: INFO : files: op(e): op(f): [started] writing systemd drop-in "00-custom-metadata.conf" at "/sysroot/etc/systemd/system/coreos-metadata.service.d/00-custom-metadata.conf" Dec 16 09:37:00.272194 ignition[964]: INFO : files: op(e): op(f): [finished] writing systemd drop-in "00-custom-metadata.conf" at "/sysroot/etc/systemd/system/coreos-metadata.service.d/00-custom-metadata.conf" Dec 16 09:37:00.272194 ignition[964]: INFO : files: op(e): [finished] processing unit "coreos-metadata.service" Dec 16 09:37:00.272194 ignition[964]: INFO : files: op(10): [started] setting preset to enabled for "prepare-helm.service" Dec 16 09:37:00.272194 ignition[964]: INFO : files: op(10): [finished] setting preset to enabled for "prepare-helm.service" Dec 16 09:37:00.272194 ignition[964]: INFO : files: createResultFile: createFiles: op(11): [started] writing file "/sysroot/etc/.ignition-result.json" Dec 16 09:37:00.272194 ignition[964]: INFO : files: createResultFile: createFiles: op(11): [finished] writing file "/sysroot/etc/.ignition-result.json" Dec 16 09:37:00.272194 ignition[964]: INFO : files: files passed Dec 16 09:37:00.272194 ignition[964]: INFO : Ignition finished successfully Dec 16 09:37:00.271393 systemd[1]: Finished ignition-files.service - Ignition (files). Dec 16 09:37:00.278851 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Dec 16 09:37:00.282816 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Dec 16 09:37:00.285153 systemd[1]: ignition-quench.service: Deactivated successfully. Dec 16 09:37:00.285254 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Dec 16 09:37:00.296645 initrd-setup-root-after-ignition[994]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Dec 16 09:37:00.296645 initrd-setup-root-after-ignition[994]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Dec 16 09:37:00.298903 initrd-setup-root-after-ignition[998]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Dec 16 09:37:00.300396 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Dec 16 09:37:00.301971 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Dec 16 09:37:00.307694 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Dec 16 09:37:00.328658 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Dec 16 09:37:00.328793 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Dec 16 09:37:00.329893 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Dec 16 09:37:00.330784 systemd[1]: Reached target initrd.target - Initrd Default Target. Dec 16 09:37:00.331792 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Dec 16 09:37:00.335727 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Dec 16 09:37:00.349037 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Dec 16 09:37:00.353706 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Dec 16 09:37:00.362463 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Dec 16 09:37:00.363030 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Dec 16 09:37:00.364202 systemd[1]: Stopped target timers.target - Timer Units. Dec 16 09:37:00.365216 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Dec 16 09:37:00.365311 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Dec 16 09:37:00.366347 systemd[1]: Stopped target initrd.target - Initrd Default Target. Dec 16 09:37:00.367425 systemd[1]: Stopped target basic.target - Basic System. Dec 16 09:37:00.368436 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Dec 16 09:37:00.369414 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Dec 16 09:37:00.370399 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Dec 16 09:37:00.371733 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Dec 16 09:37:00.372729 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Dec 16 09:37:00.373870 systemd[1]: Stopped target sysinit.target - System Initialization. Dec 16 09:37:00.374894 systemd[1]: Stopped target local-fs.target - Local File Systems. Dec 16 09:37:00.375909 systemd[1]: Stopped target swap.target - Swaps. Dec 16 09:37:00.376814 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Dec 16 09:37:00.376959 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Dec 16 09:37:00.378014 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Dec 16 09:37:00.378681 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Dec 16 09:37:00.379531 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Dec 16 09:37:00.379659 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Dec 16 09:37:00.380587 systemd[1]: dracut-initqueue.service: Deactivated successfully. Dec 16 09:37:00.380681 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Dec 16 09:37:00.382109 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Dec 16 09:37:00.382212 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Dec 16 09:37:00.382872 systemd[1]: ignition-files.service: Deactivated successfully. Dec 16 09:37:00.383009 systemd[1]: Stopped ignition-files.service - Ignition (files). Dec 16 09:37:00.383701 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Dec 16 09:37:00.383789 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Dec 16 09:37:00.395007 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Dec 16 09:37:00.397751 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Dec 16 09:37:00.398235 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Dec 16 09:37:00.398385 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Dec 16 09:37:00.400292 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Dec 16 09:37:00.400429 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Dec 16 09:37:00.409419 ignition[1018]: INFO : Ignition 2.19.0 Dec 16 09:37:00.409419 ignition[1018]: INFO : Stage: umount Dec 16 09:37:00.416260 ignition[1018]: INFO : no configs at "/usr/lib/ignition/base.d" Dec 16 09:37:00.416260 ignition[1018]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Dec 16 09:37:00.416260 ignition[1018]: INFO : umount: umount passed Dec 16 09:37:00.416260 ignition[1018]: INFO : Ignition finished successfully Dec 16 09:37:00.412211 systemd[1]: initrd-cleanup.service: Deactivated successfully. Dec 16 09:37:00.412319 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Dec 16 09:37:00.414183 systemd[1]: ignition-mount.service: Deactivated successfully. Dec 16 09:37:00.414291 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Dec 16 09:37:00.415467 systemd[1]: ignition-disks.service: Deactivated successfully. Dec 16 09:37:00.415549 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Dec 16 09:37:00.418310 systemd[1]: ignition-kargs.service: Deactivated successfully. Dec 16 09:37:00.418358 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Dec 16 09:37:00.418997 systemd[1]: ignition-fetch.service: Deactivated successfully. Dec 16 09:37:00.419041 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Dec 16 09:37:00.421115 systemd[1]: Stopped target network.target - Network. Dec 16 09:37:00.421515 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Dec 16 09:37:00.421588 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Dec 16 09:37:00.422207 systemd[1]: Stopped target paths.target - Path Units. Dec 16 09:37:00.425627 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Dec 16 09:37:00.427743 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Dec 16 09:37:00.428278 systemd[1]: Stopped target slices.target - Slice Units. Dec 16 09:37:00.428684 systemd[1]: Stopped target sockets.target - Socket Units. Dec 16 09:37:00.429112 systemd[1]: iscsid.socket: Deactivated successfully. Dec 16 09:37:00.429154 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Dec 16 09:37:00.430231 systemd[1]: iscsiuio.socket: Deactivated successfully. Dec 16 09:37:00.430274 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Dec 16 09:37:00.433633 systemd[1]: ignition-setup.service: Deactivated successfully. Dec 16 09:37:00.433682 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Dec 16 09:37:00.435103 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Dec 16 09:37:00.435148 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Dec 16 09:37:00.436275 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Dec 16 09:37:00.437316 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Dec 16 09:37:00.439965 systemd[1]: sysroot-boot.mount: Deactivated successfully. Dec 16 09:37:00.440382 systemd-networkd[781]: eth1: DHCPv6 lease lost Dec 16 09:37:00.445604 systemd-networkd[781]: eth0: DHCPv6 lease lost Dec 16 09:37:00.447216 systemd[1]: systemd-networkd.service: Deactivated successfully. Dec 16 09:37:00.447328 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Dec 16 09:37:00.450316 systemd[1]: systemd-networkd.socket: Deactivated successfully. Dec 16 09:37:00.450367 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Dec 16 09:37:00.462713 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Dec 16 09:37:00.463156 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Dec 16 09:37:00.463207 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Dec 16 09:37:00.472127 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Dec 16 09:37:00.473908 systemd[1]: systemd-resolved.service: Deactivated successfully. Dec 16 09:37:00.474054 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Dec 16 09:37:00.482947 systemd[1]: systemd-udevd.service: Deactivated successfully. Dec 16 09:37:00.483103 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Dec 16 09:37:00.483892 systemd[1]: sysroot-boot.service: Deactivated successfully. Dec 16 09:37:00.483986 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Dec 16 09:37:00.485768 systemd[1]: network-cleanup.service: Deactivated successfully. Dec 16 09:37:00.485875 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Dec 16 09:37:00.488901 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Dec 16 09:37:00.488953 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Dec 16 09:37:00.489456 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Dec 16 09:37:00.489494 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Dec 16 09:37:00.490485 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Dec 16 09:37:00.490534 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Dec 16 09:37:00.491946 systemd[1]: dracut-cmdline.service: Deactivated successfully. Dec 16 09:37:00.491992 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Dec 16 09:37:00.493026 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Dec 16 09:37:00.493089 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Dec 16 09:37:00.494029 systemd[1]: initrd-setup-root.service: Deactivated successfully. Dec 16 09:37:00.494073 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Dec 16 09:37:00.499739 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Dec 16 09:37:00.500441 systemd[1]: systemd-sysctl.service: Deactivated successfully. Dec 16 09:37:00.500493 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Dec 16 09:37:00.500954 systemd[1]: systemd-modules-load.service: Deactivated successfully. Dec 16 09:37:00.500996 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Dec 16 09:37:00.503289 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Dec 16 09:37:00.503334 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Dec 16 09:37:00.506657 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Dec 16 09:37:00.506707 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Dec 16 09:37:00.507209 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Dec 16 09:37:00.507263 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Dec 16 09:37:00.507765 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Dec 16 09:37:00.507810 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Dec 16 09:37:00.508330 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Dec 16 09:37:00.508372 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Dec 16 09:37:00.509858 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Dec 16 09:37:00.509948 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Dec 16 09:37:00.511110 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Dec 16 09:37:00.518080 systemd[1]: Starting initrd-switch-root.service - Switch Root... Dec 16 09:37:00.523243 systemd[1]: Switching root. Dec 16 09:37:00.552914 systemd-journald[187]: Journal stopped Dec 16 09:37:01.529968 systemd-journald[187]: Received SIGTERM from PID 1 (systemd). Dec 16 09:37:01.530041 kernel: SELinux: policy capability network_peer_controls=1 Dec 16 09:37:01.530054 kernel: SELinux: policy capability open_perms=1 Dec 16 09:37:01.530063 kernel: SELinux: policy capability extended_socket_class=1 Dec 16 09:37:01.530073 kernel: SELinux: policy capability always_check_network=0 Dec 16 09:37:01.530087 kernel: SELinux: policy capability cgroup_seclabel=1 Dec 16 09:37:01.530097 kernel: SELinux: policy capability nnp_nosuid_transition=1 Dec 16 09:37:01.530106 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Dec 16 09:37:01.530121 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Dec 16 09:37:01.530130 kernel: audit: type=1403 audit(1734341820.684:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Dec 16 09:37:01.530143 systemd[1]: Successfully loaded SELinux policy in 46.674ms. Dec 16 09:37:01.530174 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 12.489ms. Dec 16 09:37:01.530185 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Dec 16 09:37:01.530195 systemd[1]: Detected virtualization kvm. Dec 16 09:37:01.530210 systemd[1]: Detected architecture x86-64. Dec 16 09:37:01.530220 systemd[1]: Detected first boot. Dec 16 09:37:01.530230 systemd[1]: Hostname set to . Dec 16 09:37:01.530241 systemd[1]: Initializing machine ID from VM UUID. Dec 16 09:37:01.530253 zram_generator::config[1061]: No configuration found. Dec 16 09:37:01.530263 systemd[1]: Populated /etc with preset unit settings. Dec 16 09:37:01.530273 systemd[1]: initrd-switch-root.service: Deactivated successfully. Dec 16 09:37:01.530283 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Dec 16 09:37:01.530293 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Dec 16 09:37:01.530303 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Dec 16 09:37:01.530313 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Dec 16 09:37:01.530323 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Dec 16 09:37:01.530335 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Dec 16 09:37:01.530345 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Dec 16 09:37:01.530355 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Dec 16 09:37:01.530365 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Dec 16 09:37:01.530376 systemd[1]: Created slice user.slice - User and Session Slice. Dec 16 09:37:01.530385 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Dec 16 09:37:01.530397 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Dec 16 09:37:01.530407 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Dec 16 09:37:01.530417 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Dec 16 09:37:01.530429 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Dec 16 09:37:01.530439 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Dec 16 09:37:01.530449 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Dec 16 09:37:01.530459 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Dec 16 09:37:01.530469 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Dec 16 09:37:01.530479 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Dec 16 09:37:01.530491 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Dec 16 09:37:01.530501 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Dec 16 09:37:01.530511 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Dec 16 09:37:01.530521 systemd[1]: Reached target remote-fs.target - Remote File Systems. Dec 16 09:37:01.530531 systemd[1]: Reached target slices.target - Slice Units. Dec 16 09:37:01.530541 systemd[1]: Reached target swap.target - Swaps. Dec 16 09:37:01.530551 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Dec 16 09:37:01.534766 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Dec 16 09:37:01.534791 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Dec 16 09:37:01.534807 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Dec 16 09:37:01.534817 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Dec 16 09:37:01.534829 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Dec 16 09:37:01.534842 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Dec 16 09:37:01.534854 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Dec 16 09:37:01.534864 systemd[1]: Mounting media.mount - External Media Directory... Dec 16 09:37:01.534876 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 16 09:37:01.534886 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Dec 16 09:37:01.534897 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Dec 16 09:37:01.534907 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Dec 16 09:37:01.534917 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Dec 16 09:37:01.534929 systemd[1]: Reached target machines.target - Containers. Dec 16 09:37:01.534939 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Dec 16 09:37:01.534949 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Dec 16 09:37:01.534961 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Dec 16 09:37:01.534971 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Dec 16 09:37:01.534981 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Dec 16 09:37:01.534991 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Dec 16 09:37:01.535001 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Dec 16 09:37:01.535011 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Dec 16 09:37:01.535021 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Dec 16 09:37:01.535031 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Dec 16 09:37:01.535041 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Dec 16 09:37:01.535053 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Dec 16 09:37:01.535063 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Dec 16 09:37:01.535073 systemd[1]: Stopped systemd-fsck-usr.service. Dec 16 09:37:01.535083 systemd[1]: Starting systemd-journald.service - Journal Service... Dec 16 09:37:01.535093 kernel: loop: module loaded Dec 16 09:37:01.535103 kernel: fuse: init (API version 7.39) Dec 16 09:37:01.535113 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Dec 16 09:37:01.535123 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Dec 16 09:37:01.535133 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Dec 16 09:37:01.535145 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Dec 16 09:37:01.535155 systemd[1]: verity-setup.service: Deactivated successfully. Dec 16 09:37:01.535165 systemd[1]: Stopped verity-setup.service. Dec 16 09:37:01.535175 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 16 09:37:01.535203 systemd-journald[1141]: Collecting audit messages is disabled. Dec 16 09:37:01.535223 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Dec 16 09:37:01.535233 systemd-journald[1141]: Journal started Dec 16 09:37:01.535254 systemd-journald[1141]: Runtime Journal (/run/log/journal/df949b230dfa435a85dd75356b28f8be) is 4.8M, max 38.4M, 33.6M free. Dec 16 09:37:01.538200 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Dec 16 09:37:01.266186 systemd[1]: Queued start job for default target multi-user.target. Dec 16 09:37:01.540966 systemd[1]: Started systemd-journald.service - Journal Service. Dec 16 09:37:01.283775 systemd[1]: Unnecessary job was removed for dev-sda6.device - /dev/sda6. Dec 16 09:37:01.284467 systemd[1]: systemd-journald.service: Deactivated successfully. Dec 16 09:37:01.543792 systemd[1]: Mounted media.mount - External Media Directory. Dec 16 09:37:01.544327 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Dec 16 09:37:01.544873 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Dec 16 09:37:01.545590 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Dec 16 09:37:01.546227 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Dec 16 09:37:01.549325 systemd[1]: modprobe@configfs.service: Deactivated successfully. Dec 16 09:37:01.549506 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Dec 16 09:37:01.550434 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Dec 16 09:37:01.550639 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Dec 16 09:37:01.552000 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Dec 16 09:37:01.552139 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Dec 16 09:37:01.554662 systemd[1]: modprobe@fuse.service: Deactivated successfully. Dec 16 09:37:01.554826 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Dec 16 09:37:01.555604 kernel: ACPI: bus type drm_connector registered Dec 16 09:37:01.555950 systemd[1]: modprobe@loop.service: Deactivated successfully. Dec 16 09:37:01.556115 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Dec 16 09:37:01.558056 systemd[1]: modprobe@drm.service: Deactivated successfully. Dec 16 09:37:01.558205 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Dec 16 09:37:01.558936 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Dec 16 09:37:01.559751 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Dec 16 09:37:01.560455 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Dec 16 09:37:01.564792 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Dec 16 09:37:01.577350 systemd[1]: Reached target network-pre.target - Preparation for Network. Dec 16 09:37:01.585629 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Dec 16 09:37:01.588312 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Dec 16 09:37:01.588972 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Dec 16 09:37:01.589058 systemd[1]: Reached target local-fs.target - Local File Systems. Dec 16 09:37:01.590386 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Dec 16 09:37:01.596069 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Dec 16 09:37:01.600659 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Dec 16 09:37:01.601243 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Dec 16 09:37:01.608697 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Dec 16 09:37:01.612273 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Dec 16 09:37:01.614643 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Dec 16 09:37:01.620953 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Dec 16 09:37:01.621850 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Dec 16 09:37:01.625671 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Dec 16 09:37:01.627792 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Dec 16 09:37:01.638450 systemd-journald[1141]: Time spent on flushing to /var/log/journal/df949b230dfa435a85dd75356b28f8be is 26.807ms for 1134 entries. Dec 16 09:37:01.638450 systemd-journald[1141]: System Journal (/var/log/journal/df949b230dfa435a85dd75356b28f8be) is 8.0M, max 584.8M, 576.8M free. Dec 16 09:37:01.694283 systemd-journald[1141]: Received client request to flush runtime journal. Dec 16 09:37:01.694326 kernel: loop0: detected capacity change from 0 to 8 Dec 16 09:37:01.632725 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Dec 16 09:37:01.635874 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Dec 16 09:37:01.636442 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Dec 16 09:37:01.637803 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Dec 16 09:37:01.686956 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Dec 16 09:37:01.687575 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Dec 16 09:37:01.696854 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Dec 16 09:37:01.708059 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Dec 16 09:37:01.706935 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Dec 16 09:37:01.735740 kernel: loop1: detected capacity change from 0 to 142488 Dec 16 09:37:01.747977 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Dec 16 09:37:01.751839 systemd-tmpfiles[1182]: ACLs are not supported, ignoring. Dec 16 09:37:01.751851 systemd-tmpfiles[1182]: ACLs are not supported, ignoring. Dec 16 09:37:01.753717 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Dec 16 09:37:01.758085 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Dec 16 09:37:01.772902 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Dec 16 09:37:01.784758 systemd[1]: Starting systemd-sysusers.service - Create System Users... Dec 16 09:37:01.785553 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Dec 16 09:37:01.790221 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Dec 16 09:37:01.792584 kernel: loop2: detected capacity change from 0 to 140768 Dec 16 09:37:01.807629 udevadm[1201]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in. Dec 16 09:37:01.845962 systemd[1]: Finished systemd-sysusers.service - Create System Users. Dec 16 09:37:01.855373 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Dec 16 09:37:01.855597 kernel: loop3: detected capacity change from 0 to 205544 Dec 16 09:37:01.870222 systemd-tmpfiles[1205]: ACLs are not supported, ignoring. Dec 16 09:37:01.870533 systemd-tmpfiles[1205]: ACLs are not supported, ignoring. Dec 16 09:37:01.875912 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Dec 16 09:37:01.904612 kernel: loop4: detected capacity change from 0 to 8 Dec 16 09:37:01.908919 kernel: loop5: detected capacity change from 0 to 142488 Dec 16 09:37:01.930582 kernel: loop6: detected capacity change from 0 to 140768 Dec 16 09:37:01.952628 kernel: loop7: detected capacity change from 0 to 205544 Dec 16 09:37:01.969894 (sd-merge)[1209]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-hetzner'. Dec 16 09:37:01.972278 (sd-merge)[1209]: Merged extensions into '/usr'. Dec 16 09:37:01.978035 systemd[1]: Reloading requested from client PID 1181 ('systemd-sysext') (unit systemd-sysext.service)... Dec 16 09:37:01.978170 systemd[1]: Reloading... Dec 16 09:37:02.071660 zram_generator::config[1235]: No configuration found. Dec 16 09:37:02.137594 ldconfig[1176]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Dec 16 09:37:02.178048 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Dec 16 09:37:02.220783 systemd[1]: Reloading finished in 242 ms. Dec 16 09:37:02.240398 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Dec 16 09:37:02.241493 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Dec 16 09:37:02.252071 systemd[1]: Starting ensure-sysext.service... Dec 16 09:37:02.254100 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Dec 16 09:37:02.267677 systemd[1]: Reloading requested from client PID 1278 ('systemctl') (unit ensure-sysext.service)... Dec 16 09:37:02.267689 systemd[1]: Reloading... Dec 16 09:37:02.287909 systemd-tmpfiles[1279]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Dec 16 09:37:02.288205 systemd-tmpfiles[1279]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Dec 16 09:37:02.289032 systemd-tmpfiles[1279]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Dec 16 09:37:02.289287 systemd-tmpfiles[1279]: ACLs are not supported, ignoring. Dec 16 09:37:02.289361 systemd-tmpfiles[1279]: ACLs are not supported, ignoring. Dec 16 09:37:02.292589 systemd-tmpfiles[1279]: Detected autofs mount point /boot during canonicalization of boot. Dec 16 09:37:02.292623 systemd-tmpfiles[1279]: Skipping /boot Dec 16 09:37:02.306888 systemd-tmpfiles[1279]: Detected autofs mount point /boot during canonicalization of boot. Dec 16 09:37:02.306899 systemd-tmpfiles[1279]: Skipping /boot Dec 16 09:37:02.338590 zram_generator::config[1305]: No configuration found. Dec 16 09:37:02.432631 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Dec 16 09:37:02.473664 systemd[1]: Reloading finished in 205 ms. Dec 16 09:37:02.489503 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Dec 16 09:37:02.497059 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Dec 16 09:37:02.504765 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Dec 16 09:37:02.508725 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Dec 16 09:37:02.515864 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Dec 16 09:37:02.522725 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Dec 16 09:37:02.528062 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Dec 16 09:37:02.530796 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Dec 16 09:37:02.543492 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 16 09:37:02.543978 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Dec 16 09:37:02.548474 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Dec 16 09:37:02.553820 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Dec 16 09:37:02.557857 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Dec 16 09:37:02.558726 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Dec 16 09:37:02.558869 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 16 09:37:02.562449 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 16 09:37:02.562938 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Dec 16 09:37:02.563152 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Dec 16 09:37:02.563276 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 16 09:37:02.570380 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 16 09:37:02.570780 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Dec 16 09:37:02.579946 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Dec 16 09:37:02.580771 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Dec 16 09:37:02.580929 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 16 09:37:02.595799 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Dec 16 09:37:02.597694 systemd[1]: Finished ensure-sysext.service. Dec 16 09:37:02.598478 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Dec 16 09:37:02.603060 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Dec 16 09:37:02.603204 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Dec 16 09:37:02.605342 systemd[1]: modprobe@drm.service: Deactivated successfully. Dec 16 09:37:02.605963 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Dec 16 09:37:02.607490 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Dec 16 09:37:02.608431 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Dec 16 09:37:02.609652 systemd[1]: modprobe@loop.service: Deactivated successfully. Dec 16 09:37:02.610520 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Dec 16 09:37:02.611648 systemd-udevd[1361]: Using default interface naming scheme 'v255'. Dec 16 09:37:02.616183 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Dec 16 09:37:02.623767 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Dec 16 09:37:02.624063 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Dec 16 09:37:02.632720 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Dec 16 09:37:02.638720 systemd[1]: Starting systemd-update-done.service - Update is Completed... Dec 16 09:37:02.646157 augenrules[1385]: No rules Dec 16 09:37:02.649367 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Dec 16 09:37:02.656052 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Dec 16 09:37:02.663899 systemd[1]: Starting systemd-networkd.service - Network Configuration... Dec 16 09:37:02.670498 systemd[1]: Started systemd-userdbd.service - User Database Manager. Dec 16 09:37:02.672554 systemd[1]: Finished systemd-update-done.service - Update is Completed. Dec 16 09:37:02.681187 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Dec 16 09:37:02.681786 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Dec 16 09:37:02.752601 kernel: BTRFS info: devid 1 device path /dev/mapper/usr changed to /dev/dm-0 scanned by (udev-worker) (1403) Dec 16 09:37:02.760597 kernel: BTRFS info: devid 1 device path /dev/dm-0 changed to /dev/mapper/usr scanned by (udev-worker) (1403) Dec 16 09:37:02.795856 systemd-networkd[1391]: lo: Link UP Dec 16 09:37:02.796180 systemd-networkd[1391]: lo: Gained carrier Dec 16 09:37:02.797206 systemd-networkd[1391]: Enumeration completed Dec 16 09:37:02.797604 systemd[1]: Started systemd-networkd.service - Network Configuration. Dec 16 09:37:02.798720 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Dec 16 09:37:02.806756 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Dec 16 09:37:02.837779 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Dec 16 09:37:02.838780 systemd[1]: Reached target time-set.target - System Time Set. Dec 16 09:37:02.845924 systemd-resolved[1359]: Positive Trust Anchors: Dec 16 09:37:02.845943 systemd-resolved[1359]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Dec 16 09:37:02.845969 systemd-resolved[1359]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Dec 16 09:37:02.852667 systemd-resolved[1359]: Using system hostname 'ci-4081-2-1-7-459a6b42ef'. Dec 16 09:37:02.854367 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Dec 16 09:37:02.854995 systemd[1]: Reached target network.target - Network. Dec 16 09:37:02.855767 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Dec 16 09:37:02.859418 systemd-networkd[1391]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 16 09:37:02.859427 systemd-networkd[1391]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Dec 16 09:37:02.861245 systemd-networkd[1391]: eth0: Link UP Dec 16 09:37:02.861401 systemd-networkd[1391]: eth0: Gained carrier Dec 16 09:37:02.861462 systemd-networkd[1391]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 16 09:37:02.870592 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Dec 16 09:37:02.872528 systemd-networkd[1391]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 16 09:37:02.872540 systemd-networkd[1391]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. Dec 16 09:37:02.873845 systemd-networkd[1391]: eth1: Link UP Dec 16 09:37:02.873856 systemd-networkd[1391]: eth1: Gained carrier Dec 16 09:37:02.873867 systemd-networkd[1391]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 16 09:37:02.881586 kernel: ACPI: button: Power Button [PWRF] Dec 16 09:37:02.887634 kernel: mousedev: PS/2 mouse device common for all mice Dec 16 09:37:02.902650 systemd-networkd[1391]: eth1: DHCPv4 address 10.0.0.3/32, gateway 10.0.0.1 acquired from 10.0.0.1 Dec 16 09:37:02.903970 systemd-timesyncd[1382]: Network configuration changed, trying to establish connection. Dec 16 09:37:02.921955 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 41 scanned by (udev-worker) (1395) Dec 16 09:37:02.917155 systemd[1]: Condition check resulted in dev-virtio\x2dports-org.qemu.guest_agent.0.device - /dev/virtio-ports/org.qemu.guest_agent.0 being skipped. Dec 16 09:37:02.917230 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 16 09:37:02.917329 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Dec 16 09:37:02.922281 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Dec 16 09:37:02.931700 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Dec 16 09:37:02.934704 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Dec 16 09:37:02.935218 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Dec 16 09:37:02.935246 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Dec 16 09:37:02.935258 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 16 09:37:02.956481 kernel: [drm] pci: virtio-vga detected at 0000:00:01.0 Dec 16 09:37:02.956536 kernel: virtio-pci 0000:00:01.0: vgaarb: deactivate vga console Dec 16 09:37:02.954708 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Dec 16 09:37:02.954874 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Dec 16 09:37:02.994647 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Dec 16 09:37:02.994910 kernel: i2c i2c-0: 1/1 memory slots populated (from DMI) Dec 16 09:37:02.995086 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Dec 16 09:37:02.998038 kernel: Console: switching to colour dummy device 80x25 Dec 16 09:37:03.000640 systemd-networkd[1391]: eth0: DHCPv4 address 138.199.145.148/32, gateway 172.31.1.1 acquired from 172.31.1.1 Dec 16 09:37:03.001383 systemd-timesyncd[1382]: Network configuration changed, trying to establish connection. Dec 16 09:37:03.006587 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input4 Dec 16 09:37:03.009589 kernel: EDAC MC: Ver: 3.0.0 Dec 16 09:37:03.018692 kernel: [drm] features: -virgl +edid -resource_blob -host_visible Dec 16 09:37:03.018736 kernel: [drm] features: -context_init Dec 16 09:37:03.019911 kernel: [drm] number of scanouts: 1 Dec 16 09:37:03.019938 kernel: [drm] number of cap sets: 0 Dec 16 09:37:03.023639 kernel: [drm] Initialized virtio_gpu 0.1.0 0 for 0000:00:01.0 on minor 0 Dec 16 09:37:03.033600 kernel: fbcon: virtio_gpudrmfb (fb0) is primary device Dec 16 09:37:03.033650 kernel: Console: switching to colour frame buffer device 160x50 Dec 16 09:37:03.034176 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Dec 16 09:37:03.034413 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Dec 16 09:37:03.048495 kernel: virtio-pci 0000:00:01.0: [drm] fb0: virtio_gpudrmfb frame buffer device Dec 16 09:37:03.048326 systemd[1]: modprobe@loop.service: Deactivated successfully. Dec 16 09:37:03.048479 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Dec 16 09:37:03.060216 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - QEMU_HARDDISK OEM. Dec 16 09:37:03.069107 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Dec 16 09:37:03.069880 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Dec 16 09:37:03.070005 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Dec 16 09:37:03.073068 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Dec 16 09:37:03.075655 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Dec 16 09:37:03.075826 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Dec 16 09:37:03.079205 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Dec 16 09:37:03.095219 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Dec 16 09:37:03.156798 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Dec 16 09:37:03.166482 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Dec 16 09:37:03.176766 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Dec 16 09:37:03.188961 lvm[1458]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Dec 16 09:37:03.218374 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Dec 16 09:37:03.220866 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Dec 16 09:37:03.220960 systemd[1]: Reached target sysinit.target - System Initialization. Dec 16 09:37:03.221137 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Dec 16 09:37:03.221247 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Dec 16 09:37:03.221492 systemd[1]: Started logrotate.timer - Daily rotation of log files. Dec 16 09:37:03.221718 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Dec 16 09:37:03.221790 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Dec 16 09:37:03.221851 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Dec 16 09:37:03.221868 systemd[1]: Reached target paths.target - Path Units. Dec 16 09:37:03.221919 systemd[1]: Reached target timers.target - Timer Units. Dec 16 09:37:03.231268 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Dec 16 09:37:03.234217 systemd[1]: Starting docker.socket - Docker Socket for the API... Dec 16 09:37:03.247267 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Dec 16 09:37:03.252197 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Dec 16 09:37:03.258478 systemd[1]: Listening on docker.socket - Docker Socket for the API. Dec 16 09:37:03.259182 systemd[1]: Reached target sockets.target - Socket Units. Dec 16 09:37:03.262173 lvm[1462]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Dec 16 09:37:03.262838 systemd[1]: Reached target basic.target - Basic System. Dec 16 09:37:03.263515 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Dec 16 09:37:03.263590 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Dec 16 09:37:03.268681 systemd[1]: Starting containerd.service - containerd container runtime... Dec 16 09:37:03.277744 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Dec 16 09:37:03.282065 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Dec 16 09:37:03.293707 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Dec 16 09:37:03.298963 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Dec 16 09:37:03.299791 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Dec 16 09:37:03.301890 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Dec 16 09:37:03.314250 jq[1468]: false Dec 16 09:37:03.309080 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Dec 16 09:37:03.310744 systemd[1]: Started qemu-guest-agent.service - QEMU Guest Agent. Dec 16 09:37:03.321713 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Dec 16 09:37:03.325732 coreos-metadata[1464]: Dec 16 09:37:03.323 INFO Fetching http://169.254.169.254/hetzner/v1/metadata: Attempt #1 Dec 16 09:37:03.326130 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Dec 16 09:37:03.336202 coreos-metadata[1464]: Dec 16 09:37:03.330 INFO Fetch successful Dec 16 09:37:03.336202 coreos-metadata[1464]: Dec 16 09:37:03.331 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/private-networks: Attempt #1 Dec 16 09:37:03.336777 systemd[1]: Starting systemd-logind.service - User Login Management... Dec 16 09:37:03.338200 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Dec 16 09:37:03.338647 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Dec 16 09:37:03.344256 coreos-metadata[1464]: Dec 16 09:37:03.343 INFO Fetch successful Dec 16 09:37:03.342699 systemd[1]: Starting update-engine.service - Update Engine... Dec 16 09:37:03.347939 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Dec 16 09:37:03.351502 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Dec 16 09:37:03.360008 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Dec 16 09:37:03.371253 extend-filesystems[1469]: Found loop4 Dec 16 09:37:03.371253 extend-filesystems[1469]: Found loop5 Dec 16 09:37:03.371253 extend-filesystems[1469]: Found loop6 Dec 16 09:37:03.371253 extend-filesystems[1469]: Found loop7 Dec 16 09:37:03.371253 extend-filesystems[1469]: Found sda Dec 16 09:37:03.371253 extend-filesystems[1469]: Found sda1 Dec 16 09:37:03.371253 extend-filesystems[1469]: Found sda2 Dec 16 09:37:03.371253 extend-filesystems[1469]: Found sda3 Dec 16 09:37:03.371253 extend-filesystems[1469]: Found usr Dec 16 09:37:03.371253 extend-filesystems[1469]: Found sda4 Dec 16 09:37:03.371253 extend-filesystems[1469]: Found sda6 Dec 16 09:37:03.371253 extend-filesystems[1469]: Found sda7 Dec 16 09:37:03.371253 extend-filesystems[1469]: Found sda9 Dec 16 09:37:03.371253 extend-filesystems[1469]: Checking size of /dev/sda9 Dec 16 09:37:03.361055 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Dec 16 09:37:03.442363 extend-filesystems[1469]: Resized partition /dev/sda9 Dec 16 09:37:03.476729 kernel: EXT4-fs (sda9): resizing filesystem from 1617920 to 9393147 blocks Dec 16 09:37:03.476760 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 41 scanned by (udev-worker) (1395) Dec 16 09:37:03.410984 dbus-daemon[1465]: [system] SELinux support is enabled Dec 16 09:37:03.361861 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Dec 16 09:37:03.477190 extend-filesystems[1503]: resize2fs 1.47.1 (20-May-2024) Dec 16 09:37:03.362670 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Dec 16 09:37:03.498543 jq[1484]: true Dec 16 09:37:03.366168 systemd[1]: motdgen.service: Deactivated successfully. Dec 16 09:37:03.498936 update_engine[1481]: I20241216 09:37:03.463788 1481 main.cc:92] Flatcar Update Engine starting Dec 16 09:37:03.498936 update_engine[1481]: I20241216 09:37:03.467300 1481 update_check_scheduler.cc:74] Next update check in 6m23s Dec 16 09:37:03.366333 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Dec 16 09:37:03.499273 tar[1487]: linux-amd64/helm Dec 16 09:37:03.412670 systemd[1]: Started dbus.service - D-Bus System Message Bus. Dec 16 09:37:03.424246 (ntainerd)[1489]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Dec 16 09:37:03.463667 systemd-logind[1477]: New seat seat0. Dec 16 09:37:03.493399 systemd[1]: Started update-engine.service - Update Engine. Dec 16 09:37:03.502625 systemd-logind[1477]: Watching system buttons on /dev/input/event2 (Power Button) Dec 16 09:37:03.502645 systemd-logind[1477]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Dec 16 09:37:03.513801 systemd[1]: Started systemd-logind.service - User Login Management. Dec 16 09:37:03.515333 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Dec 16 09:37:03.515364 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Dec 16 09:37:03.519453 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Dec 16 09:37:03.519476 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Dec 16 09:37:03.529558 systemd[1]: Started locksmithd.service - Cluster reboot manager. Dec 16 09:37:03.568605 jq[1508]: true Dec 16 09:37:03.614447 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Dec 16 09:37:03.619905 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Dec 16 09:37:03.682318 locksmithd[1513]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Dec 16 09:37:03.686197 kernel: EXT4-fs (sda9): resized filesystem to 9393147 Dec 16 09:37:03.686240 sshd_keygen[1506]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Dec 16 09:37:03.702399 extend-filesystems[1503]: Filesystem at /dev/sda9 is mounted on /; on-line resizing required Dec 16 09:37:03.702399 extend-filesystems[1503]: old_desc_blocks = 1, new_desc_blocks = 5 Dec 16 09:37:03.702399 extend-filesystems[1503]: The filesystem on /dev/sda9 is now 9393147 (4k) blocks long. Dec 16 09:37:03.705077 extend-filesystems[1469]: Resized filesystem in /dev/sda9 Dec 16 09:37:03.705077 extend-filesystems[1469]: Found sr0 Dec 16 09:37:03.703603 systemd[1]: extend-filesystems.service: Deactivated successfully. Dec 16 09:37:03.703806 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Dec 16 09:37:03.714813 bash[1539]: Updated "/home/core/.ssh/authorized_keys" Dec 16 09:37:03.716844 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Dec 16 09:37:03.720619 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Dec 16 09:37:03.730989 systemd[1]: Starting issuegen.service - Generate /run/issue... Dec 16 09:37:03.743477 systemd[1]: Starting sshkeys.service... Dec 16 09:37:03.749782 systemd[1]: issuegen.service: Deactivated successfully. Dec 16 09:37:03.750140 systemd[1]: Finished issuegen.service - Generate /run/issue. Dec 16 09:37:03.761911 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Dec 16 09:37:03.771900 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Dec 16 09:37:03.784049 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Dec 16 09:37:03.787824 containerd[1489]: time="2024-12-16T09:37:03.786717243Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Dec 16 09:37:03.804440 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Dec 16 09:37:03.812867 coreos-metadata[1564]: Dec 16 09:37:03.812 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/public-keys: Attempt #1 Dec 16 09:37:03.815258 systemd[1]: Started getty@tty1.service - Getty on tty1. Dec 16 09:37:03.817902 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Dec 16 09:37:03.818518 systemd[1]: Reached target getty.target - Login Prompts. Dec 16 09:37:03.823929 coreos-metadata[1564]: Dec 16 09:37:03.823 INFO Fetch successful Dec 16 09:37:03.827794 unknown[1564]: wrote ssh authorized keys file for user: core Dec 16 09:37:03.830330 containerd[1489]: time="2024-12-16T09:37:03.830204754Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Dec 16 09:37:03.833858 containerd[1489]: time="2024-12-16T09:37:03.832539002Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.65-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Dec 16 09:37:03.833858 containerd[1489]: time="2024-12-16T09:37:03.832577083Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Dec 16 09:37:03.833858 containerd[1489]: time="2024-12-16T09:37:03.832604765Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Dec 16 09:37:03.833858 containerd[1489]: time="2024-12-16T09:37:03.832757993Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Dec 16 09:37:03.833858 containerd[1489]: time="2024-12-16T09:37:03.832771789Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Dec 16 09:37:03.833858 containerd[1489]: time="2024-12-16T09:37:03.832840487Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Dec 16 09:37:03.833858 containerd[1489]: time="2024-12-16T09:37:03.832851518Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Dec 16 09:37:03.833858 containerd[1489]: time="2024-12-16T09:37:03.833664904Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Dec 16 09:37:03.834590 containerd[1489]: time="2024-12-16T09:37:03.833680934Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Dec 16 09:37:03.834590 containerd[1489]: time="2024-12-16T09:37:03.834156445Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Dec 16 09:37:03.834590 containerd[1489]: time="2024-12-16T09:37:03.834169630Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Dec 16 09:37:03.834590 containerd[1489]: time="2024-12-16T09:37:03.834261342Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Dec 16 09:37:03.834590 containerd[1489]: time="2024-12-16T09:37:03.834477347Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Dec 16 09:37:03.835760 containerd[1489]: time="2024-12-16T09:37:03.835729425Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Dec 16 09:37:03.835760 containerd[1489]: time="2024-12-16T09:37:03.835755354Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Dec 16 09:37:03.835872 containerd[1489]: time="2024-12-16T09:37:03.835850262Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Dec 16 09:37:03.836068 containerd[1489]: time="2024-12-16T09:37:03.835909643Z" level=info msg="metadata content store policy set" policy=shared Dec 16 09:37:03.839763 containerd[1489]: time="2024-12-16T09:37:03.839731571Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Dec 16 09:37:03.839808 containerd[1489]: time="2024-12-16T09:37:03.839784190Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Dec 16 09:37:03.839808 containerd[1489]: time="2024-12-16T09:37:03.839803075Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Dec 16 09:37:03.839889 containerd[1489]: time="2024-12-16T09:37:03.839818174Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Dec 16 09:37:03.839889 containerd[1489]: time="2024-12-16T09:37:03.839832341Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Dec 16 09:37:03.840170 containerd[1489]: time="2024-12-16T09:37:03.839949931Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Dec 16 09:37:03.840170 containerd[1489]: time="2024-12-16T09:37:03.840129808Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Dec 16 09:37:03.840254 containerd[1489]: time="2024-12-16T09:37:03.840226139Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Dec 16 09:37:03.840254 containerd[1489]: time="2024-12-16T09:37:03.840247960Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Dec 16 09:37:03.840298 containerd[1489]: time="2024-12-16T09:37:03.840259342Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Dec 16 09:37:03.840298 containerd[1489]: time="2024-12-16T09:37:03.840276694Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Dec 16 09:37:03.840298 containerd[1489]: time="2024-12-16T09:37:03.840287845Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Dec 16 09:37:03.840298 containerd[1489]: time="2024-12-16T09:37:03.840297633Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Dec 16 09:37:03.840362 containerd[1489]: time="2024-12-16T09:37:03.840308974Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Dec 16 09:37:03.840362 containerd[1489]: time="2024-12-16T09:37:03.840320927Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Dec 16 09:37:03.840362 containerd[1489]: time="2024-12-16T09:37:03.840331827Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Dec 16 09:37:03.840362 containerd[1489]: time="2024-12-16T09:37:03.840341546Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Dec 16 09:37:03.840362 containerd[1489]: time="2024-12-16T09:37:03.840350963Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Dec 16 09:37:03.840457 containerd[1489]: time="2024-12-16T09:37:03.840367213Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Dec 16 09:37:03.840457 containerd[1489]: time="2024-12-16T09:37:03.840378776Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Dec 16 09:37:03.840457 containerd[1489]: time="2024-12-16T09:37:03.840391029Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Dec 16 09:37:03.840457 containerd[1489]: time="2024-12-16T09:37:03.840402380Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Dec 16 09:37:03.840457 containerd[1489]: time="2024-12-16T09:37:03.840412889Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Dec 16 09:37:03.840457 containerd[1489]: time="2024-12-16T09:37:03.840423169Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Dec 16 09:37:03.840457 containerd[1489]: time="2024-12-16T09:37:03.840433207Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Dec 16 09:37:03.840457 containerd[1489]: time="2024-12-16T09:37:03.840443046Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Dec 16 09:37:03.840457 containerd[1489]: time="2024-12-16T09:37:03.840452915Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Dec 16 09:37:03.840642 containerd[1489]: time="2024-12-16T09:37:03.840464937Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Dec 16 09:37:03.840642 containerd[1489]: time="2024-12-16T09:37:03.840474946Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Dec 16 09:37:03.840642 containerd[1489]: time="2024-12-16T09:37:03.840484824Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Dec 16 09:37:03.840642 containerd[1489]: time="2024-12-16T09:37:03.840503610Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Dec 16 09:37:03.840642 containerd[1489]: time="2024-12-16T09:37:03.840516524Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Dec 16 09:37:03.840642 containerd[1489]: time="2024-12-16T09:37:03.840532584Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Dec 16 09:37:03.840642 containerd[1489]: time="2024-12-16T09:37:03.840549465Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Dec 16 09:37:03.840642 containerd[1489]: time="2024-12-16T09:37:03.840558142Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Dec 16 09:37:03.842063 containerd[1489]: time="2024-12-16T09:37:03.841493426Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Dec 16 09:37:03.842063 containerd[1489]: time="2024-12-16T09:37:03.841519545Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Dec 16 09:37:03.842063 containerd[1489]: time="2024-12-16T09:37:03.841530015Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Dec 16 09:37:03.842063 containerd[1489]: time="2024-12-16T09:37:03.841635873Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Dec 16 09:37:03.842063 containerd[1489]: time="2024-12-16T09:37:03.841648737Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Dec 16 09:37:03.842063 containerd[1489]: time="2024-12-16T09:37:03.841662904Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Dec 16 09:37:03.842063 containerd[1489]: time="2024-12-16T09:37:03.841672843Z" level=info msg="NRI interface is disabled by configuration." Dec 16 09:37:03.842063 containerd[1489]: time="2024-12-16T09:37:03.841681589Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Dec 16 09:37:03.842207 containerd[1489]: time="2024-12-16T09:37:03.841894999Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Dec 16 09:37:03.842207 containerd[1489]: time="2024-12-16T09:37:03.841942018Z" level=info msg="Connect containerd service" Dec 16 09:37:03.842207 containerd[1489]: time="2024-12-16T09:37:03.841968457Z" level=info msg="using legacy CRI server" Dec 16 09:37:03.842207 containerd[1489]: time="2024-12-16T09:37:03.841974478Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Dec 16 09:37:03.842207 containerd[1489]: time="2024-12-16T09:37:03.842066571Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Dec 16 09:37:03.845245 containerd[1489]: time="2024-12-16T09:37:03.844360583Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Dec 16 09:37:03.845920 containerd[1489]: time="2024-12-16T09:37:03.845880664Z" level=info msg="Start subscribing containerd event" Dec 16 09:37:03.845961 containerd[1489]: time="2024-12-16T09:37:03.845926380Z" level=info msg="Start recovering state" Dec 16 09:37:03.846004 containerd[1489]: time="2024-12-16T09:37:03.845980852Z" level=info msg="Start event monitor" Dec 16 09:37:03.846004 containerd[1489]: time="2024-12-16T09:37:03.846003284Z" level=info msg="Start snapshots syncer" Dec 16 09:37:03.846044 containerd[1489]: time="2024-12-16T09:37:03.846012091Z" level=info msg="Start cni network conf syncer for default" Dec 16 09:37:03.846044 containerd[1489]: time="2024-12-16T09:37:03.846019675Z" level=info msg="Start streaming server" Dec 16 09:37:03.846246 containerd[1489]: time="2024-12-16T09:37:03.846219790Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Dec 16 09:37:03.846298 containerd[1489]: time="2024-12-16T09:37:03.846276116Z" level=info msg=serving... address=/run/containerd/containerd.sock Dec 16 09:37:03.846390 systemd[1]: Started containerd.service - containerd container runtime. Dec 16 09:37:03.850866 containerd[1489]: time="2024-12-16T09:37:03.850837641Z" level=info msg="containerd successfully booted in 0.066224s" Dec 16 09:37:03.863598 update-ssh-keys[1573]: Updated "/home/core/.ssh/authorized_keys" Dec 16 09:37:03.864146 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Dec 16 09:37:03.869728 systemd[1]: Finished sshkeys.service. Dec 16 09:37:04.073452 tar[1487]: linux-amd64/LICENSE Dec 16 09:37:04.073537 tar[1487]: linux-amd64/README.md Dec 16 09:37:04.084697 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Dec 16 09:37:04.478895 systemd-networkd[1391]: eth0: Gained IPv6LL Dec 16 09:37:04.481917 systemd-timesyncd[1382]: Network configuration changed, trying to establish connection. Dec 16 09:37:04.486642 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Dec 16 09:37:04.490800 systemd[1]: Reached target network-online.target - Network is Online. Dec 16 09:37:04.501854 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 16 09:37:04.517167 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Dec 16 09:37:04.558987 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Dec 16 09:37:04.607830 systemd-networkd[1391]: eth1: Gained IPv6LL Dec 16 09:37:04.608655 systemd-timesyncd[1382]: Network configuration changed, trying to establish connection. Dec 16 09:37:05.284200 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 16 09:37:05.285247 systemd[1]: Reached target multi-user.target - Multi-User System. Dec 16 09:37:05.292726 (kubelet)[1595]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 16 09:37:05.292881 systemd[1]: Startup finished in 1.167s (kernel) + 5.994s (initrd) + 4.650s (userspace) = 11.812s. Dec 16 09:37:05.828413 kubelet[1595]: E1216 09:37:05.828345 1595 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 16 09:37:05.831839 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 16 09:37:05.832023 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 16 09:37:15.990592 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Dec 16 09:37:15.996046 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 16 09:37:16.120299 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 16 09:37:16.127841 (kubelet)[1614]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 16 09:37:16.170959 kubelet[1614]: E1216 09:37:16.170889 1614 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 16 09:37:16.177047 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 16 09:37:16.177355 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 16 09:37:26.240332 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Dec 16 09:37:26.245738 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 16 09:37:26.366146 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 16 09:37:26.370013 (kubelet)[1629]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 16 09:37:26.401921 kubelet[1629]: E1216 09:37:26.401856 1629 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 16 09:37:26.405660 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 16 09:37:26.405841 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 16 09:37:34.956322 systemd-timesyncd[1382]: Contacted time server 85.215.189.120:123 (2.flatcar.pool.ntp.org). Dec 16 09:37:34.956424 systemd-timesyncd[1382]: Initial clock synchronization to Mon 2024-12-16 09:37:34.938020 UTC. Dec 16 09:37:36.490316 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Dec 16 09:37:36.495784 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 16 09:37:36.618745 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 16 09:37:36.622261 (kubelet)[1645]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 16 09:37:36.655056 kubelet[1645]: E1216 09:37:36.654997 1645 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 16 09:37:36.658676 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 16 09:37:36.658863 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 16 09:37:46.740509 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 4. Dec 16 09:37:46.746749 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 16 09:37:46.868642 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 16 09:37:46.872517 (kubelet)[1660]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 16 09:37:46.907521 kubelet[1660]: E1216 09:37:46.907470 1660 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 16 09:37:46.911130 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 16 09:37:46.911375 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 16 09:37:48.482580 update_engine[1481]: I20241216 09:37:48.482452 1481 update_attempter.cc:509] Updating boot flags... Dec 16 09:37:48.523608 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 41 scanned by (udev-worker) (1676) Dec 16 09:37:48.568609 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 41 scanned by (udev-worker) (1678) Dec 16 09:37:48.612624 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 41 scanned by (udev-worker) (1678) Dec 16 09:37:56.990731 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 5. Dec 16 09:37:57.002000 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 16 09:37:57.141108 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 16 09:37:57.145252 (kubelet)[1696]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 16 09:37:57.182349 kubelet[1696]: E1216 09:37:57.182262 1696 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 16 09:37:57.185870 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 16 09:37:57.186118 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 16 09:38:07.240804 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 6. Dec 16 09:38:07.250961 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 16 09:38:07.390635 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 16 09:38:07.394684 (kubelet)[1711]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 16 09:38:07.423078 kubelet[1711]: E1216 09:38:07.423017 1711 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 16 09:38:07.425244 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 16 09:38:07.425449 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 16 09:38:17.490464 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 7. Dec 16 09:38:17.496807 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 16 09:38:17.668108 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 16 09:38:17.678809 (kubelet)[1726]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 16 09:38:17.713771 kubelet[1726]: E1216 09:38:17.713713 1726 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 16 09:38:17.717187 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 16 09:38:17.717403 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 16 09:38:27.740306 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 8. Dec 16 09:38:27.745756 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 16 09:38:27.864677 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 16 09:38:27.868527 (kubelet)[1741]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 16 09:38:27.903524 kubelet[1741]: E1216 09:38:27.903462 1741 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 16 09:38:27.906975 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 16 09:38:27.907210 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 16 09:38:37.990282 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 9. Dec 16 09:38:37.996020 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 16 09:38:38.125689 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 16 09:38:38.125897 (kubelet)[1757]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 16 09:38:38.160299 kubelet[1757]: E1216 09:38:38.160196 1757 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 16 09:38:38.163065 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 16 09:38:38.163248 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 16 09:38:48.240526 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 10. Dec 16 09:38:48.247117 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 16 09:38:48.396886 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 16 09:38:48.406934 (kubelet)[1772]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 16 09:38:48.452160 kubelet[1772]: E1216 09:38:48.452098 1772 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 16 09:38:48.455538 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 16 09:38:48.455854 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 16 09:38:58.490439 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 11. Dec 16 09:38:58.496103 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 16 09:38:58.629016 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 16 09:38:58.640946 (kubelet)[1787]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 16 09:38:58.674957 kubelet[1787]: E1216 09:38:58.674850 1787 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 16 09:38:58.678125 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 16 09:38:58.678469 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 16 09:39:03.839995 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Dec 16 09:39:03.844762 systemd[1]: Started sshd@0-138.199.145.148:22-147.75.109.163:56556.service - OpenSSH per-connection server daemon (147.75.109.163:56556). Dec 16 09:39:04.821731 sshd[1795]: Accepted publickey for core from 147.75.109.163 port 56556 ssh2: RSA SHA256:zB/zPQRxUCFkkFdvDftk99JQqA6bP3NHPa7FnaDUxKk Dec 16 09:39:04.824207 sshd[1795]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 09:39:04.833509 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Dec 16 09:39:04.838784 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Dec 16 09:39:04.842206 systemd-logind[1477]: New session 1 of user core. Dec 16 09:39:04.853144 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Dec 16 09:39:04.859832 systemd[1]: Starting user@500.service - User Manager for UID 500... Dec 16 09:39:04.865275 (systemd)[1799]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Dec 16 09:39:04.968142 systemd[1799]: Queued start job for default target default.target. Dec 16 09:39:04.977740 systemd[1799]: Created slice app.slice - User Application Slice. Dec 16 09:39:04.977767 systemd[1799]: Reached target paths.target - Paths. Dec 16 09:39:04.977781 systemd[1799]: Reached target timers.target - Timers. Dec 16 09:39:04.979225 systemd[1799]: Starting dbus.socket - D-Bus User Message Bus Socket... Dec 16 09:39:04.999910 systemd[1799]: Listening on dbus.socket - D-Bus User Message Bus Socket. Dec 16 09:39:05.000089 systemd[1799]: Reached target sockets.target - Sockets. Dec 16 09:39:05.000109 systemd[1799]: Reached target basic.target - Basic System. Dec 16 09:39:05.000160 systemd[1799]: Reached target default.target - Main User Target. Dec 16 09:39:05.000200 systemd[1799]: Startup finished in 127ms. Dec 16 09:39:05.000312 systemd[1]: Started user@500.service - User Manager for UID 500. Dec 16 09:39:05.010757 systemd[1]: Started session-1.scope - Session 1 of User core. Dec 16 09:39:05.699628 systemd[1]: Started sshd@1-138.199.145.148:22-147.75.109.163:56564.service - OpenSSH per-connection server daemon (147.75.109.163:56564). Dec 16 09:39:06.681680 sshd[1810]: Accepted publickey for core from 147.75.109.163 port 56564 ssh2: RSA SHA256:zB/zPQRxUCFkkFdvDftk99JQqA6bP3NHPa7FnaDUxKk Dec 16 09:39:06.683270 sshd[1810]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 09:39:06.687358 systemd-logind[1477]: New session 2 of user core. Dec 16 09:39:06.694676 systemd[1]: Started session-2.scope - Session 2 of User core. Dec 16 09:39:07.368346 sshd[1810]: pam_unix(sshd:session): session closed for user core Dec 16 09:39:07.372207 systemd[1]: sshd@1-138.199.145.148:22-147.75.109.163:56564.service: Deactivated successfully. Dec 16 09:39:07.374113 systemd[1]: session-2.scope: Deactivated successfully. Dec 16 09:39:07.374863 systemd-logind[1477]: Session 2 logged out. Waiting for processes to exit. Dec 16 09:39:07.375871 systemd-logind[1477]: Removed session 2. Dec 16 09:39:07.544944 systemd[1]: Started sshd@2-138.199.145.148:22-147.75.109.163:33236.service - OpenSSH per-connection server daemon (147.75.109.163:33236). Dec 16 09:39:08.543083 sshd[1817]: Accepted publickey for core from 147.75.109.163 port 33236 ssh2: RSA SHA256:zB/zPQRxUCFkkFdvDftk99JQqA6bP3NHPa7FnaDUxKk Dec 16 09:39:08.544862 sshd[1817]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 09:39:08.549099 systemd-logind[1477]: New session 3 of user core. Dec 16 09:39:08.560688 systemd[1]: Started session-3.scope - Session 3 of User core. Dec 16 09:39:08.740402 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 12. Dec 16 09:39:08.745918 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 16 09:39:08.871286 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 16 09:39:08.883825 (kubelet)[1828]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 16 09:39:08.914973 kubelet[1828]: E1216 09:39:08.914919 1828 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 16 09:39:08.918132 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 16 09:39:08.918327 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 16 09:39:09.217248 sshd[1817]: pam_unix(sshd:session): session closed for user core Dec 16 09:39:09.221416 systemd[1]: sshd@2-138.199.145.148:22-147.75.109.163:33236.service: Deactivated successfully. Dec 16 09:39:09.223090 systemd[1]: session-3.scope: Deactivated successfully. Dec 16 09:39:09.224998 systemd-logind[1477]: Session 3 logged out. Waiting for processes to exit. Dec 16 09:39:09.226140 systemd-logind[1477]: Removed session 3. Dec 16 09:39:09.390943 systemd[1]: Started sshd@3-138.199.145.148:22-147.75.109.163:33244.service - OpenSSH per-connection server daemon (147.75.109.163:33244). Dec 16 09:39:10.381634 sshd[1839]: Accepted publickey for core from 147.75.109.163 port 33244 ssh2: RSA SHA256:zB/zPQRxUCFkkFdvDftk99JQqA6bP3NHPa7FnaDUxKk Dec 16 09:39:10.383217 sshd[1839]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 09:39:10.387873 systemd-logind[1477]: New session 4 of user core. Dec 16 09:39:10.400701 systemd[1]: Started session-4.scope - Session 4 of User core. Dec 16 09:39:11.062379 sshd[1839]: pam_unix(sshd:session): session closed for user core Dec 16 09:39:11.065847 systemd[1]: sshd@3-138.199.145.148:22-147.75.109.163:33244.service: Deactivated successfully. Dec 16 09:39:11.068240 systemd[1]: session-4.scope: Deactivated successfully. Dec 16 09:39:11.070195 systemd-logind[1477]: Session 4 logged out. Waiting for processes to exit. Dec 16 09:39:11.071441 systemd-logind[1477]: Removed session 4. Dec 16 09:39:11.237940 systemd[1]: Started sshd@4-138.199.145.148:22-147.75.109.163:33250.service - OpenSSH per-connection server daemon (147.75.109.163:33250). Dec 16 09:39:12.222104 sshd[1846]: Accepted publickey for core from 147.75.109.163 port 33250 ssh2: RSA SHA256:zB/zPQRxUCFkkFdvDftk99JQqA6bP3NHPa7FnaDUxKk Dec 16 09:39:12.224822 sshd[1846]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 09:39:12.231597 systemd-logind[1477]: New session 5 of user core. Dec 16 09:39:12.236714 systemd[1]: Started session-5.scope - Session 5 of User core. Dec 16 09:39:12.759128 sudo[1849]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Dec 16 09:39:12.759749 sudo[1849]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Dec 16 09:39:12.778452 sudo[1849]: pam_unix(sudo:session): session closed for user root Dec 16 09:39:12.939892 sshd[1846]: pam_unix(sshd:session): session closed for user core Dec 16 09:39:12.943749 systemd[1]: sshd@4-138.199.145.148:22-147.75.109.163:33250.service: Deactivated successfully. Dec 16 09:39:12.945776 systemd[1]: session-5.scope: Deactivated successfully. Dec 16 09:39:12.948077 systemd-logind[1477]: Session 5 logged out. Waiting for processes to exit. Dec 16 09:39:12.949405 systemd-logind[1477]: Removed session 5. Dec 16 09:39:13.108808 systemd[1]: Started sshd@5-138.199.145.148:22-147.75.109.163:33254.service - OpenSSH per-connection server daemon (147.75.109.163:33254). Dec 16 09:39:14.086275 sshd[1854]: Accepted publickey for core from 147.75.109.163 port 33254 ssh2: RSA SHA256:zB/zPQRxUCFkkFdvDftk99JQqA6bP3NHPa7FnaDUxKk Dec 16 09:39:14.088284 sshd[1854]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 09:39:14.093130 systemd-logind[1477]: New session 6 of user core. Dec 16 09:39:14.100700 systemd[1]: Started session-6.scope - Session 6 of User core. Dec 16 09:39:14.604774 sudo[1858]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Dec 16 09:39:14.605194 sudo[1858]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Dec 16 09:39:14.609224 sudo[1858]: pam_unix(sudo:session): session closed for user root Dec 16 09:39:14.615887 sudo[1857]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Dec 16 09:39:14.616240 sudo[1857]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Dec 16 09:39:14.635829 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Dec 16 09:39:14.638085 auditctl[1861]: No rules Dec 16 09:39:14.638809 systemd[1]: audit-rules.service: Deactivated successfully. Dec 16 09:39:14.639102 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Dec 16 09:39:14.641278 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Dec 16 09:39:14.667693 augenrules[1879]: No rules Dec 16 09:39:14.669412 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Dec 16 09:39:14.670862 sudo[1857]: pam_unix(sudo:session): session closed for user root Dec 16 09:39:14.829036 sshd[1854]: pam_unix(sshd:session): session closed for user core Dec 16 09:39:14.832585 systemd[1]: sshd@5-138.199.145.148:22-147.75.109.163:33254.service: Deactivated successfully. Dec 16 09:39:14.834668 systemd[1]: session-6.scope: Deactivated successfully. Dec 16 09:39:14.836097 systemd-logind[1477]: Session 6 logged out. Waiting for processes to exit. Dec 16 09:39:14.837341 systemd-logind[1477]: Removed session 6. Dec 16 09:39:14.999466 systemd[1]: Started sshd@6-138.199.145.148:22-147.75.109.163:33270.service - OpenSSH per-connection server daemon (147.75.109.163:33270). Dec 16 09:39:15.975632 sshd[1887]: Accepted publickey for core from 147.75.109.163 port 33270 ssh2: RSA SHA256:zB/zPQRxUCFkkFdvDftk99JQqA6bP3NHPa7FnaDUxKk Dec 16 09:39:15.977207 sshd[1887]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 09:39:15.981971 systemd-logind[1477]: New session 7 of user core. Dec 16 09:39:15.986706 systemd[1]: Started session-7.scope - Session 7 of User core. Dec 16 09:39:16.496344 sudo[1890]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Dec 16 09:39:16.496786 sudo[1890]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Dec 16 09:39:16.755822 systemd[1]: Starting docker.service - Docker Application Container Engine... Dec 16 09:39:16.757222 (dockerd)[1906]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Dec 16 09:39:17.001022 dockerd[1906]: time="2024-12-16T09:39:17.000957867Z" level=info msg="Starting up" Dec 16 09:39:17.074659 systemd[1]: var-lib-docker-metacopy\x2dcheck1803686740-merged.mount: Deactivated successfully. Dec 16 09:39:17.094276 dockerd[1906]: time="2024-12-16T09:39:17.094230734Z" level=info msg="Loading containers: start." Dec 16 09:39:17.198611 kernel: Initializing XFRM netlink socket Dec 16 09:39:17.276866 systemd-networkd[1391]: docker0: Link UP Dec 16 09:39:17.300912 dockerd[1906]: time="2024-12-16T09:39:17.300863850Z" level=info msg="Loading containers: done." Dec 16 09:39:17.315429 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck973344690-merged.mount: Deactivated successfully. Dec 16 09:39:17.316619 dockerd[1906]: time="2024-12-16T09:39:17.316154889Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Dec 16 09:39:17.316619 dockerd[1906]: time="2024-12-16T09:39:17.316247442Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 Dec 16 09:39:17.316619 dockerd[1906]: time="2024-12-16T09:39:17.316353409Z" level=info msg="Daemon has completed initialization" Dec 16 09:39:17.341311 dockerd[1906]: time="2024-12-16T09:39:17.341213207Z" level=info msg="API listen on /run/docker.sock" Dec 16 09:39:17.341862 systemd[1]: Started docker.service - Docker Application Container Engine. Dec 16 09:39:18.301529 containerd[1489]: time="2024-12-16T09:39:18.301455551Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.4\"" Dec 16 09:39:18.890854 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount34479948.mount: Deactivated successfully. Dec 16 09:39:18.990325 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 13. Dec 16 09:39:18.998636 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 16 09:39:19.138555 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 16 09:39:19.142722 (kubelet)[2068]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 16 09:39:19.175584 kubelet[2068]: E1216 09:39:19.173858 2068 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 16 09:39:19.178108 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 16 09:39:19.178384 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 16 09:39:19.684523 containerd[1489]: time="2024-12-16T09:39:19.684468828Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 09:39:19.685581 containerd[1489]: time="2024-12-16T09:39:19.685409782Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.31.4: active requests=0, bytes read=27975575" Dec 16 09:39:19.686403 containerd[1489]: time="2024-12-16T09:39:19.686351309Z" level=info msg="ImageCreate event name:\"sha256:bdc2eadbf366279693097982a31da61cc2f1d90f07ada3f4b3b91251a18f665e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 09:39:19.688896 containerd[1489]: time="2024-12-16T09:39:19.688860047Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:ace6a943b058439bd6daeb74f152e7c36e6fc0b5e481cdff9364cd6ca0473e5e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 09:39:19.690251 containerd[1489]: time="2024-12-16T09:39:19.690023787Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.31.4\" with image id \"sha256:bdc2eadbf366279693097982a31da61cc2f1d90f07ada3f4b3b91251a18f665e\", repo tag \"registry.k8s.io/kube-apiserver:v1.31.4\", repo digest \"registry.k8s.io/kube-apiserver@sha256:ace6a943b058439bd6daeb74f152e7c36e6fc0b5e481cdff9364cd6ca0473e5e\", size \"27972283\" in 1.388520176s" Dec 16 09:39:19.690251 containerd[1489]: time="2024-12-16T09:39:19.690056648Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.4\" returns image reference \"sha256:bdc2eadbf366279693097982a31da61cc2f1d90f07ada3f4b3b91251a18f665e\"" Dec 16 09:39:19.693957 containerd[1489]: time="2024-12-16T09:39:19.693929370Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.4\"" Dec 16 09:39:20.801041 containerd[1489]: time="2024-12-16T09:39:20.800973099Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 09:39:20.801944 containerd[1489]: time="2024-12-16T09:39:20.801895891Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.31.4: active requests=0, bytes read=24702177" Dec 16 09:39:20.802750 containerd[1489]: time="2024-12-16T09:39:20.802710299Z" level=info msg="ImageCreate event name:\"sha256:359b9f2307326a4c66172318ca63ee9792c3146ca57d53329239bd123ea70079\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 09:39:20.804981 containerd[1489]: time="2024-12-16T09:39:20.804940919Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:4bd1d4a449e7a1a4f375bd7c71abf48a95f8949b38f725ded255077329f21f7b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 09:39:20.805971 containerd[1489]: time="2024-12-16T09:39:20.805867367Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.31.4\" with image id \"sha256:359b9f2307326a4c66172318ca63ee9792c3146ca57d53329239bd123ea70079\", repo tag \"registry.k8s.io/kube-controller-manager:v1.31.4\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:4bd1d4a449e7a1a4f375bd7c71abf48a95f8949b38f725ded255077329f21f7b\", size \"26147269\" in 1.111892131s" Dec 16 09:39:20.805971 containerd[1489]: time="2024-12-16T09:39:20.805893897Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.4\" returns image reference \"sha256:359b9f2307326a4c66172318ca63ee9792c3146ca57d53329239bd123ea70079\"" Dec 16 09:39:20.806703 containerd[1489]: time="2024-12-16T09:39:20.806377289Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.4\"" Dec 16 09:39:21.771382 containerd[1489]: time="2024-12-16T09:39:21.771320240Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 09:39:21.772192 containerd[1489]: time="2024-12-16T09:39:21.772146741Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.31.4: active requests=0, bytes read=18652087" Dec 16 09:39:21.773052 containerd[1489]: time="2024-12-16T09:39:21.773001045Z" level=info msg="ImageCreate event name:\"sha256:3a66234066fe10fa299c0a52265f90a107450f0372652867118cd9007940d674\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 09:39:21.775218 containerd[1489]: time="2024-12-16T09:39:21.775184968Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:1a3081cb7d21763d22eb2c0781cc462d89f501ed523ad558dea1226f128fbfdd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 09:39:21.776346 containerd[1489]: time="2024-12-16T09:39:21.776244435Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.31.4\" with image id \"sha256:3a66234066fe10fa299c0a52265f90a107450f0372652867118cd9007940d674\", repo tag \"registry.k8s.io/kube-scheduler:v1.31.4\", repo digest \"registry.k8s.io/kube-scheduler@sha256:1a3081cb7d21763d22eb2c0781cc462d89f501ed523ad558dea1226f128fbfdd\", size \"20097197\" in 969.844354ms" Dec 16 09:39:21.776346 containerd[1489]: time="2024-12-16T09:39:21.776271916Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.4\" returns image reference \"sha256:3a66234066fe10fa299c0a52265f90a107450f0372652867118cd9007940d674\"" Dec 16 09:39:21.777208 containerd[1489]: time="2024-12-16T09:39:21.777184558Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.4\"" Dec 16 09:39:22.726412 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount637814817.mount: Deactivated successfully. Dec 16 09:39:23.011592 containerd[1489]: time="2024-12-16T09:39:23.011343733Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 09:39:23.012348 containerd[1489]: time="2024-12-16T09:39:23.012178762Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.31.4: active requests=0, bytes read=30230269" Dec 16 09:39:23.013072 containerd[1489]: time="2024-12-16T09:39:23.013025361Z" level=info msg="ImageCreate event name:\"sha256:ebf80573666f86f115452db568feb34f6f771c3bdc7bfed14b9577f992cfa300\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 09:39:23.014534 containerd[1489]: time="2024-12-16T09:39:23.014497478Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:1739b3febca392035bf6edfe31efdfa55226be7b57389b2001ae357f7dcb99cf\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 09:39:23.015350 containerd[1489]: time="2024-12-16T09:39:23.015038688Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.31.4\" with image id \"sha256:ebf80573666f86f115452db568feb34f6f771c3bdc7bfed14b9577f992cfa300\", repo tag \"registry.k8s.io/kube-proxy:v1.31.4\", repo digest \"registry.k8s.io/kube-proxy@sha256:1739b3febca392035bf6edfe31efdfa55226be7b57389b2001ae357f7dcb99cf\", size \"30229262\" in 1.23782753s" Dec 16 09:39:23.015350 containerd[1489]: time="2024-12-16T09:39:23.015067702Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.4\" returns image reference \"sha256:ebf80573666f86f115452db568feb34f6f771c3bdc7bfed14b9577f992cfa300\"" Dec 16 09:39:23.015597 containerd[1489]: time="2024-12-16T09:39:23.015542737Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\"" Dec 16 09:39:23.541900 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2995282830.mount: Deactivated successfully. Dec 16 09:39:24.164413 containerd[1489]: time="2024-12-16T09:39:24.164334187Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 09:39:24.165286 containerd[1489]: time="2024-12-16T09:39:24.165249706Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.1: active requests=0, bytes read=18185841" Dec 16 09:39:24.165910 containerd[1489]: time="2024-12-16T09:39:24.165867127Z" level=info msg="ImageCreate event name:\"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 09:39:24.169044 containerd[1489]: time="2024-12-16T09:39:24.168997408Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 09:39:24.169846 containerd[1489]: time="2024-12-16T09:39:24.169731969Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.1\" with image id \"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\", size \"18182961\" in 1.154144047s" Dec 16 09:39:24.169846 containerd[1489]: time="2024-12-16T09:39:24.169758920Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\" returns image reference \"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\"" Dec 16 09:39:24.170322 containerd[1489]: time="2024-12-16T09:39:24.170199282Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Dec 16 09:39:24.661621 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3383936782.mount: Deactivated successfully. Dec 16 09:39:24.667185 containerd[1489]: time="2024-12-16T09:39:24.667128648Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 09:39:24.667925 containerd[1489]: time="2024-12-16T09:39:24.667888646Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321158" Dec 16 09:39:24.668674 containerd[1489]: time="2024-12-16T09:39:24.668623878Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 09:39:24.670334 containerd[1489]: time="2024-12-16T09:39:24.670297821Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 09:39:24.671587 containerd[1489]: time="2024-12-16T09:39:24.671004580Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 500.782537ms" Dec 16 09:39:24.671587 containerd[1489]: time="2024-12-16T09:39:24.671030388Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Dec 16 09:39:24.671587 containerd[1489]: time="2024-12-16T09:39:24.671516035Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\"" Dec 16 09:39:25.182746 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2705601670.mount: Deactivated successfully. Dec 16 09:39:26.499376 containerd[1489]: time="2024-12-16T09:39:26.499321027Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.15-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 09:39:26.500422 containerd[1489]: time="2024-12-16T09:39:26.500375365Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.15-0: active requests=0, bytes read=56780035" Dec 16 09:39:26.501201 containerd[1489]: time="2024-12-16T09:39:26.501149740Z" level=info msg="ImageCreate event name:\"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 09:39:26.503733 containerd[1489]: time="2024-12-16T09:39:26.503699208Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 09:39:26.504884 containerd[1489]: time="2024-12-16T09:39:26.504762583Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.15-0\" with image id \"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\", repo tag \"registry.k8s.io/etcd:3.5.15-0\", repo digest \"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\", size \"56909194\" in 1.833224877s" Dec 16 09:39:26.504884 containerd[1489]: time="2024-12-16T09:39:26.504789574Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\" returns image reference \"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\"" Dec 16 09:39:28.964545 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Dec 16 09:39:28.970797 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 16 09:39:29.007114 systemd[1]: Reloading requested from client PID 2258 ('systemctl') (unit session-7.scope)... Dec 16 09:39:29.007277 systemd[1]: Reloading... Dec 16 09:39:29.141592 zram_generator::config[2298]: No configuration found. Dec 16 09:39:29.238405 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Dec 16 09:39:29.306332 systemd[1]: Reloading finished in 298 ms. Dec 16 09:39:29.363364 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Dec 16 09:39:29.363504 systemd[1]: kubelet.service: Failed with result 'signal'. Dec 16 09:39:29.363871 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Dec 16 09:39:29.368840 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 16 09:39:29.496165 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 16 09:39:29.505908 (kubelet)[2353]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Dec 16 09:39:29.540604 kubelet[2353]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 16 09:39:29.540604 kubelet[2353]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 16 09:39:29.540604 kubelet[2353]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 16 09:39:29.540604 kubelet[2353]: I1216 09:39:29.540156 2353 server.go:206] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 16 09:39:29.999004 kubelet[2353]: I1216 09:39:29.998956 2353 server.go:486] "Kubelet version" kubeletVersion="v1.31.0" Dec 16 09:39:29.999004 kubelet[2353]: I1216 09:39:29.998996 2353 server.go:488] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 16 09:39:29.999323 kubelet[2353]: I1216 09:39:29.999298 2353 server.go:929] "Client rotation is on, will bootstrap in background" Dec 16 09:39:30.027202 kubelet[2353]: I1216 09:39:30.026983 2353 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Dec 16 09:39:30.027202 kubelet[2353]: E1216 09:39:30.027175 2353 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://138.199.145.148:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 138.199.145.148:6443: connect: connection refused" logger="UnhandledError" Dec 16 09:39:30.037487 kubelet[2353]: E1216 09:39:30.037450 2353 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Dec 16 09:39:30.037487 kubelet[2353]: I1216 09:39:30.037481 2353 server.go:1403] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Dec 16 09:39:30.042777 kubelet[2353]: I1216 09:39:30.042756 2353 server.go:744] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Dec 16 09:39:30.043811 kubelet[2353]: I1216 09:39:30.043785 2353 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 16 09:39:30.043970 kubelet[2353]: I1216 09:39:30.043933 2353 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 16 09:39:30.044100 kubelet[2353]: I1216 09:39:30.043957 2353 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4081-2-1-7-459a6b42ef","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 16 09:39:30.044100 kubelet[2353]: I1216 09:39:30.044097 2353 topology_manager.go:138] "Creating topology manager with none policy" Dec 16 09:39:30.044196 kubelet[2353]: I1216 09:39:30.044105 2353 container_manager_linux.go:300] "Creating device plugin manager" Dec 16 09:39:30.044218 kubelet[2353]: I1216 09:39:30.044209 2353 state_mem.go:36] "Initialized new in-memory state store" Dec 16 09:39:30.045836 kubelet[2353]: I1216 09:39:30.045806 2353 kubelet.go:408] "Attempting to sync node with API server" Dec 16 09:39:30.045836 kubelet[2353]: I1216 09:39:30.045833 2353 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 16 09:39:30.045895 kubelet[2353]: I1216 09:39:30.045861 2353 kubelet.go:314] "Adding apiserver pod source" Dec 16 09:39:30.045895 kubelet[2353]: I1216 09:39:30.045876 2353 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 16 09:39:30.051455 kubelet[2353]: W1216 09:39:30.051237 2353 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://138.199.145.148:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081-2-1-7-459a6b42ef&limit=500&resourceVersion=0": dial tcp 138.199.145.148:6443: connect: connection refused Dec 16 09:39:30.051455 kubelet[2353]: E1216 09:39:30.051281 2353 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://138.199.145.148:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081-2-1-7-459a6b42ef&limit=500&resourceVersion=0\": dial tcp 138.199.145.148:6443: connect: connection refused" logger="UnhandledError" Dec 16 09:39:30.051671 kubelet[2353]: W1216 09:39:30.051634 2353 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://138.199.145.148:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 138.199.145.148:6443: connect: connection refused Dec 16 09:39:30.051671 kubelet[2353]: E1216 09:39:30.051668 2353 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://138.199.145.148:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 138.199.145.148:6443: connect: connection refused" logger="UnhandledError" Dec 16 09:39:30.052459 kubelet[2353]: I1216 09:39:30.051903 2353 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Dec 16 09:39:30.054543 kubelet[2353]: I1216 09:39:30.054512 2353 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 16 09:39:30.055348 kubelet[2353]: W1216 09:39:30.055250 2353 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Dec 16 09:39:30.057282 kubelet[2353]: I1216 09:39:30.057124 2353 server.go:1269] "Started kubelet" Dec 16 09:39:30.057328 kubelet[2353]: I1216 09:39:30.057286 2353 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 16 09:39:30.058208 kubelet[2353]: I1216 09:39:30.058058 2353 server.go:460] "Adding debug handlers to kubelet server" Dec 16 09:39:30.060246 kubelet[2353]: I1216 09:39:30.059813 2353 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 16 09:39:30.060246 kubelet[2353]: I1216 09:39:30.060013 2353 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 16 09:39:30.060246 kubelet[2353]: I1216 09:39:30.060202 2353 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 16 09:39:30.065044 kubelet[2353]: E1216 09:39:30.062037 2353 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://138.199.145.148:6443/api/v1/namespaces/default/events\": dial tcp 138.199.145.148:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4081-2-1-7-459a6b42ef.18119ed5344e18e2 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4081-2-1-7-459a6b42ef,UID:ci-4081-2-1-7-459a6b42ef,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4081-2-1-7-459a6b42ef,},FirstTimestamp:2024-12-16 09:39:30.057107682 +0000 UTC m=+0.547608261,LastTimestamp:2024-12-16 09:39:30.057107682 +0000 UTC m=+0.547608261,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4081-2-1-7-459a6b42ef,}" Dec 16 09:39:30.065507 kubelet[2353]: I1216 09:39:30.065491 2353 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Dec 16 09:39:30.067431 kubelet[2353]: I1216 09:39:30.067410 2353 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 16 09:39:30.067633 kubelet[2353]: E1216 09:39:30.067613 2353 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4081-2-1-7-459a6b42ef\" not found" Dec 16 09:39:30.069414 kubelet[2353]: E1216 09:39:30.069384 2353 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://138.199.145.148:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081-2-1-7-459a6b42ef?timeout=10s\": dial tcp 138.199.145.148:6443: connect: connection refused" interval="200ms" Dec 16 09:39:30.070790 kubelet[2353]: I1216 09:39:30.070776 2353 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 16 09:39:30.071074 kubelet[2353]: I1216 09:39:30.071053 2353 reconciler.go:26] "Reconciler: start to sync state" Dec 16 09:39:30.071203 kubelet[2353]: W1216 09:39:30.071172 2353 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://138.199.145.148:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 138.199.145.148:6443: connect: connection refused Dec 16 09:39:30.071275 kubelet[2353]: E1216 09:39:30.071260 2353 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://138.199.145.148:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 138.199.145.148:6443: connect: connection refused" logger="UnhandledError" Dec 16 09:39:30.071509 kubelet[2353]: I1216 09:39:30.071472 2353 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Dec 16 09:39:30.071962 kubelet[2353]: E1216 09:39:30.071941 2353 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Dec 16 09:39:30.073112 kubelet[2353]: I1216 09:39:30.073098 2353 factory.go:221] Registration of the containerd container factory successfully Dec 16 09:39:30.073194 kubelet[2353]: I1216 09:39:30.073184 2353 factory.go:221] Registration of the systemd container factory successfully Dec 16 09:39:30.085713 kubelet[2353]: I1216 09:39:30.085666 2353 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 16 09:39:30.087942 kubelet[2353]: I1216 09:39:30.087634 2353 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 16 09:39:30.087942 kubelet[2353]: I1216 09:39:30.087663 2353 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 16 09:39:30.087942 kubelet[2353]: I1216 09:39:30.087677 2353 kubelet.go:2321] "Starting kubelet main sync loop" Dec 16 09:39:30.087942 kubelet[2353]: E1216 09:39:30.087717 2353 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 16 09:39:30.097641 kubelet[2353]: W1216 09:39:30.097604 2353 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://138.199.145.148:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 138.199.145.148:6443: connect: connection refused Dec 16 09:39:30.097815 kubelet[2353]: E1216 09:39:30.097788 2353 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://138.199.145.148:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 138.199.145.148:6443: connect: connection refused" logger="UnhandledError" Dec 16 09:39:30.106056 kubelet[2353]: I1216 09:39:30.106041 2353 cpu_manager.go:214] "Starting CPU manager" policy="none" Dec 16 09:39:30.106126 kubelet[2353]: I1216 09:39:30.106116 2353 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Dec 16 09:39:30.106195 kubelet[2353]: I1216 09:39:30.106187 2353 state_mem.go:36] "Initialized new in-memory state store" Dec 16 09:39:30.108100 kubelet[2353]: I1216 09:39:30.108086 2353 policy_none.go:49] "None policy: Start" Dec 16 09:39:30.108632 kubelet[2353]: I1216 09:39:30.108619 2353 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 16 09:39:30.108951 kubelet[2353]: I1216 09:39:30.108707 2353 state_mem.go:35] "Initializing new in-memory state store" Dec 16 09:39:30.115766 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Dec 16 09:39:30.129314 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Dec 16 09:39:30.133348 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Dec 16 09:39:30.151309 kubelet[2353]: I1216 09:39:30.150817 2353 manager.go:510] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 16 09:39:30.151309 kubelet[2353]: I1216 09:39:30.151088 2353 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 16 09:39:30.151309 kubelet[2353]: I1216 09:39:30.151104 2353 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 16 09:39:30.151309 kubelet[2353]: I1216 09:39:30.151313 2353 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 16 09:39:30.153631 kubelet[2353]: E1216 09:39:30.153609 2353 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4081-2-1-7-459a6b42ef\" not found" Dec 16 09:39:30.199206 systemd[1]: Created slice kubepods-burstable-pod051a64e47cbe729b1fee0d93ec1faff5.slice - libcontainer container kubepods-burstable-pod051a64e47cbe729b1fee0d93ec1faff5.slice. Dec 16 09:39:30.208274 systemd[1]: Created slice kubepods-burstable-pod15c858ba29bf2c6e9d716382efcb2dbf.slice - libcontainer container kubepods-burstable-pod15c858ba29bf2c6e9d716382efcb2dbf.slice. Dec 16 09:39:30.228151 systemd[1]: Created slice kubepods-burstable-pod24f5c188f00bb4e278359e82a7e1823f.slice - libcontainer container kubepods-burstable-pod24f5c188f00bb4e278359e82a7e1823f.slice. Dec 16 09:39:30.253972 kubelet[2353]: I1216 09:39:30.253876 2353 kubelet_node_status.go:72] "Attempting to register node" node="ci-4081-2-1-7-459a6b42ef" Dec 16 09:39:30.254501 kubelet[2353]: E1216 09:39:30.254474 2353 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://138.199.145.148:6443/api/v1/nodes\": dial tcp 138.199.145.148:6443: connect: connection refused" node="ci-4081-2-1-7-459a6b42ef" Dec 16 09:39:30.271047 kubelet[2353]: E1216 09:39:30.271005 2353 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://138.199.145.148:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081-2-1-7-459a6b42ef?timeout=10s\": dial tcp 138.199.145.148:6443: connect: connection refused" interval="400ms" Dec 16 09:39:30.272229 kubelet[2353]: I1216 09:39:30.272186 2353 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/15c858ba29bf2c6e9d716382efcb2dbf-flexvolume-dir\") pod \"kube-controller-manager-ci-4081-2-1-7-459a6b42ef\" (UID: \"15c858ba29bf2c6e9d716382efcb2dbf\") " pod="kube-system/kube-controller-manager-ci-4081-2-1-7-459a6b42ef" Dec 16 09:39:30.272229 kubelet[2353]: I1216 09:39:30.272214 2353 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/15c858ba29bf2c6e9d716382efcb2dbf-kubeconfig\") pod \"kube-controller-manager-ci-4081-2-1-7-459a6b42ef\" (UID: \"15c858ba29bf2c6e9d716382efcb2dbf\") " pod="kube-system/kube-controller-manager-ci-4081-2-1-7-459a6b42ef" Dec 16 09:39:30.272333 kubelet[2353]: I1216 09:39:30.272234 2353 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/15c858ba29bf2c6e9d716382efcb2dbf-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4081-2-1-7-459a6b42ef\" (UID: \"15c858ba29bf2c6e9d716382efcb2dbf\") " pod="kube-system/kube-controller-manager-ci-4081-2-1-7-459a6b42ef" Dec 16 09:39:30.272333 kubelet[2353]: I1216 09:39:30.272250 2353 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/24f5c188f00bb4e278359e82a7e1823f-kubeconfig\") pod \"kube-scheduler-ci-4081-2-1-7-459a6b42ef\" (UID: \"24f5c188f00bb4e278359e82a7e1823f\") " pod="kube-system/kube-scheduler-ci-4081-2-1-7-459a6b42ef" Dec 16 09:39:30.272333 kubelet[2353]: I1216 09:39:30.272265 2353 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/051a64e47cbe729b1fee0d93ec1faff5-ca-certs\") pod \"kube-apiserver-ci-4081-2-1-7-459a6b42ef\" (UID: \"051a64e47cbe729b1fee0d93ec1faff5\") " pod="kube-system/kube-apiserver-ci-4081-2-1-7-459a6b42ef" Dec 16 09:39:30.272333 kubelet[2353]: I1216 09:39:30.272282 2353 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/051a64e47cbe729b1fee0d93ec1faff5-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4081-2-1-7-459a6b42ef\" (UID: \"051a64e47cbe729b1fee0d93ec1faff5\") " pod="kube-system/kube-apiserver-ci-4081-2-1-7-459a6b42ef" Dec 16 09:39:30.272333 kubelet[2353]: I1216 09:39:30.272296 2353 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/15c858ba29bf2c6e9d716382efcb2dbf-ca-certs\") pod \"kube-controller-manager-ci-4081-2-1-7-459a6b42ef\" (UID: \"15c858ba29bf2c6e9d716382efcb2dbf\") " pod="kube-system/kube-controller-manager-ci-4081-2-1-7-459a6b42ef" Dec 16 09:39:30.272510 kubelet[2353]: I1216 09:39:30.272310 2353 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/051a64e47cbe729b1fee0d93ec1faff5-k8s-certs\") pod \"kube-apiserver-ci-4081-2-1-7-459a6b42ef\" (UID: \"051a64e47cbe729b1fee0d93ec1faff5\") " pod="kube-system/kube-apiserver-ci-4081-2-1-7-459a6b42ef" Dec 16 09:39:30.272510 kubelet[2353]: I1216 09:39:30.272333 2353 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/15c858ba29bf2c6e9d716382efcb2dbf-k8s-certs\") pod \"kube-controller-manager-ci-4081-2-1-7-459a6b42ef\" (UID: \"15c858ba29bf2c6e9d716382efcb2dbf\") " pod="kube-system/kube-controller-manager-ci-4081-2-1-7-459a6b42ef" Dec 16 09:39:30.456894 kubelet[2353]: I1216 09:39:30.456716 2353 kubelet_node_status.go:72] "Attempting to register node" node="ci-4081-2-1-7-459a6b42ef" Dec 16 09:39:30.457093 kubelet[2353]: E1216 09:39:30.457055 2353 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://138.199.145.148:6443/api/v1/nodes\": dial tcp 138.199.145.148:6443: connect: connection refused" node="ci-4081-2-1-7-459a6b42ef" Dec 16 09:39:30.506738 containerd[1489]: time="2024-12-16T09:39:30.506602691Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4081-2-1-7-459a6b42ef,Uid:051a64e47cbe729b1fee0d93ec1faff5,Namespace:kube-system,Attempt:0,}" Dec 16 09:39:30.526421 containerd[1489]: time="2024-12-16T09:39:30.526376479Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4081-2-1-7-459a6b42ef,Uid:15c858ba29bf2c6e9d716382efcb2dbf,Namespace:kube-system,Attempt:0,}" Dec 16 09:39:30.533836 containerd[1489]: time="2024-12-16T09:39:30.533724609Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4081-2-1-7-459a6b42ef,Uid:24f5c188f00bb4e278359e82a7e1823f,Namespace:kube-system,Attempt:0,}" Dec 16 09:39:30.671438 kubelet[2353]: E1216 09:39:30.671377 2353 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://138.199.145.148:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081-2-1-7-459a6b42ef?timeout=10s\": dial tcp 138.199.145.148:6443: connect: connection refused" interval="800ms" Dec 16 09:39:30.859738 kubelet[2353]: I1216 09:39:30.859611 2353 kubelet_node_status.go:72] "Attempting to register node" node="ci-4081-2-1-7-459a6b42ef" Dec 16 09:39:30.860172 kubelet[2353]: E1216 09:39:30.860131 2353 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://138.199.145.148:6443/api/v1/nodes\": dial tcp 138.199.145.148:6443: connect: connection refused" node="ci-4081-2-1-7-459a6b42ef" Dec 16 09:39:31.004958 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1496754021.mount: Deactivated successfully. Dec 16 09:39:31.011376 containerd[1489]: time="2024-12-16T09:39:31.011312862Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Dec 16 09:39:31.012271 containerd[1489]: time="2024-12-16T09:39:31.012217070Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312076" Dec 16 09:39:31.013094 containerd[1489]: time="2024-12-16T09:39:31.013055615Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Dec 16 09:39:31.014351 containerd[1489]: time="2024-12-16T09:39:31.014306481Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Dec 16 09:39:31.018373 containerd[1489]: time="2024-12-16T09:39:31.018206784Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Dec 16 09:39:31.019320 containerd[1489]: time="2024-12-16T09:39:31.019276382Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Dec 16 09:39:31.020005 containerd[1489]: time="2024-12-16T09:39:31.019822592Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Dec 16 09:39:31.021609 containerd[1489]: time="2024-12-16T09:39:31.020973841Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Dec 16 09:39:31.023540 containerd[1489]: time="2024-12-16T09:39:31.023394771Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 489.60984ms" Dec 16 09:39:31.025070 containerd[1489]: time="2024-12-16T09:39:31.025025927Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 518.340723ms" Dec 16 09:39:31.026982 containerd[1489]: time="2024-12-16T09:39:31.026880631Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 500.42853ms" Dec 16 09:39:31.162537 containerd[1489]: time="2024-12-16T09:39:31.162334544Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 16 09:39:31.162537 containerd[1489]: time="2024-12-16T09:39:31.162397923Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 16 09:39:31.162537 containerd[1489]: time="2024-12-16T09:39:31.162412450Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 16 09:39:31.163670 containerd[1489]: time="2024-12-16T09:39:31.163464504Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 16 09:39:31.163670 containerd[1489]: time="2024-12-16T09:39:31.162242102Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 16 09:39:31.163670 containerd[1489]: time="2024-12-16T09:39:31.163434889Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 16 09:39:31.163670 containerd[1489]: time="2024-12-16T09:39:31.163446440Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 16 09:39:31.164040 containerd[1489]: time="2024-12-16T09:39:31.163725632Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 16 09:39:31.167096 containerd[1489]: time="2024-12-16T09:39:31.166830660Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 16 09:39:31.167096 containerd[1489]: time="2024-12-16T09:39:31.166886424Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 16 09:39:31.167096 containerd[1489]: time="2024-12-16T09:39:31.166907763Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 16 09:39:31.167096 containerd[1489]: time="2024-12-16T09:39:31.166998393Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 16 09:39:31.188926 systemd[1]: Started cri-containerd-abe0fa2234a416b1b4ef32934846f5a4e4f92162262edaf37a15afa2b81bde98.scope - libcontainer container abe0fa2234a416b1b4ef32934846f5a4e4f92162262edaf37a15afa2b81bde98. Dec 16 09:39:31.201697 systemd[1]: Started cri-containerd-60ec9c0f41808819781fa5bb065aefe9e0371cac551135dce0599cd70fe7e2f1.scope - libcontainer container 60ec9c0f41808819781fa5bb065aefe9e0371cac551135dce0599cd70fe7e2f1. Dec 16 09:39:31.203788 systemd[1]: Started cri-containerd-7e5d76e65b344aad03f37883243005bc96b123137290468a466d701328c33b77.scope - libcontainer container 7e5d76e65b344aad03f37883243005bc96b123137290468a466d701328c33b77. Dec 16 09:39:31.251840 containerd[1489]: time="2024-12-16T09:39:31.251275979Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4081-2-1-7-459a6b42ef,Uid:24f5c188f00bb4e278359e82a7e1823f,Namespace:kube-system,Attempt:0,} returns sandbox id \"abe0fa2234a416b1b4ef32934846f5a4e4f92162262edaf37a15afa2b81bde98\"" Dec 16 09:39:31.262618 containerd[1489]: time="2024-12-16T09:39:31.259279783Z" level=info msg="CreateContainer within sandbox \"abe0fa2234a416b1b4ef32934846f5a4e4f92162262edaf37a15afa2b81bde98\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Dec 16 09:39:31.274204 containerd[1489]: time="2024-12-16T09:39:31.274173895Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4081-2-1-7-459a6b42ef,Uid:051a64e47cbe729b1fee0d93ec1faff5,Namespace:kube-system,Attempt:0,} returns sandbox id \"60ec9c0f41808819781fa5bb065aefe9e0371cac551135dce0599cd70fe7e2f1\"" Dec 16 09:39:31.277713 containerd[1489]: time="2024-12-16T09:39:31.277688557Z" level=info msg="CreateContainer within sandbox \"60ec9c0f41808819781fa5bb065aefe9e0371cac551135dce0599cd70fe7e2f1\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Dec 16 09:39:31.281439 containerd[1489]: time="2024-12-16T09:39:31.281398174Z" level=info msg="CreateContainer within sandbox \"abe0fa2234a416b1b4ef32934846f5a4e4f92162262edaf37a15afa2b81bde98\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"f76f63d0e9ba0d2c306e06eb64ba47d546040ec2c148ce96b0d1450085ea20f2\"" Dec 16 09:39:31.291510 containerd[1489]: time="2024-12-16T09:39:31.291452148Z" level=info msg="CreateContainer within sandbox \"60ec9c0f41808819781fa5bb065aefe9e0371cac551135dce0599cd70fe7e2f1\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"9d4b7924804f45573663bce8118ce94ec25a54867ba30ddacaa9ee74722212cc\"" Dec 16 09:39:31.293443 containerd[1489]: time="2024-12-16T09:39:31.292100487Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4081-2-1-7-459a6b42ef,Uid:15c858ba29bf2c6e9d716382efcb2dbf,Namespace:kube-system,Attempt:0,} returns sandbox id \"7e5d76e65b344aad03f37883243005bc96b123137290468a466d701328c33b77\"" Dec 16 09:39:31.293443 containerd[1489]: time="2024-12-16T09:39:31.292355904Z" level=info msg="StartContainer for \"9d4b7924804f45573663bce8118ce94ec25a54867ba30ddacaa9ee74722212cc\"" Dec 16 09:39:31.293443 containerd[1489]: time="2024-12-16T09:39:31.292436766Z" level=info msg="StartContainer for \"f76f63d0e9ba0d2c306e06eb64ba47d546040ec2c148ce96b0d1450085ea20f2\"" Dec 16 09:39:31.298133 containerd[1489]: time="2024-12-16T09:39:31.298105731Z" level=info msg="CreateContainer within sandbox \"7e5d76e65b344aad03f37883243005bc96b123137290468a466d701328c33b77\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Dec 16 09:39:31.313821 containerd[1489]: time="2024-12-16T09:39:31.313748961Z" level=info msg="CreateContainer within sandbox \"7e5d76e65b344aad03f37883243005bc96b123137290468a466d701328c33b77\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"f6fed61c4091f9c718c7a3e6c3b6b244a4612e44741e3ec8c44aac6c654801a4\"" Dec 16 09:39:31.314471 containerd[1489]: time="2024-12-16T09:39:31.314440081Z" level=info msg="StartContainer for \"f6fed61c4091f9c718c7a3e6c3b6b244a4612e44741e3ec8c44aac6c654801a4\"" Dec 16 09:39:31.318485 kubelet[2353]: W1216 09:39:31.318372 2353 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://138.199.145.148:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 138.199.145.148:6443: connect: connection refused Dec 16 09:39:31.318485 kubelet[2353]: E1216 09:39:31.318452 2353 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://138.199.145.148:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 138.199.145.148:6443: connect: connection refused" logger="UnhandledError" Dec 16 09:39:31.333864 systemd[1]: Started cri-containerd-9d4b7924804f45573663bce8118ce94ec25a54867ba30ddacaa9ee74722212cc.scope - libcontainer container 9d4b7924804f45573663bce8118ce94ec25a54867ba30ddacaa9ee74722212cc. Dec 16 09:39:31.349848 systemd[1]: Started cri-containerd-f76f63d0e9ba0d2c306e06eb64ba47d546040ec2c148ce96b0d1450085ea20f2.scope - libcontainer container f76f63d0e9ba0d2c306e06eb64ba47d546040ec2c148ce96b0d1450085ea20f2. Dec 16 09:39:31.356722 systemd[1]: Started cri-containerd-f6fed61c4091f9c718c7a3e6c3b6b244a4612e44741e3ec8c44aac6c654801a4.scope - libcontainer container f6fed61c4091f9c718c7a3e6c3b6b244a4612e44741e3ec8c44aac6c654801a4. Dec 16 09:39:31.409593 containerd[1489]: time="2024-12-16T09:39:31.409533963Z" level=info msg="StartContainer for \"9d4b7924804f45573663bce8118ce94ec25a54867ba30ddacaa9ee74722212cc\" returns successfully" Dec 16 09:39:31.425404 containerd[1489]: time="2024-12-16T09:39:31.425220152Z" level=info msg="StartContainer for \"f6fed61c4091f9c718c7a3e6c3b6b244a4612e44741e3ec8c44aac6c654801a4\" returns successfully" Dec 16 09:39:31.436804 containerd[1489]: time="2024-12-16T09:39:31.436764710Z" level=info msg="StartContainer for \"f76f63d0e9ba0d2c306e06eb64ba47d546040ec2c148ce96b0d1450085ea20f2\" returns successfully" Dec 16 09:39:31.472046 kubelet[2353]: E1216 09:39:31.471971 2353 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://138.199.145.148:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081-2-1-7-459a6b42ef?timeout=10s\": dial tcp 138.199.145.148:6443: connect: connection refused" interval="1.6s" Dec 16 09:39:31.490195 kubelet[2353]: W1216 09:39:31.489978 2353 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://138.199.145.148:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 138.199.145.148:6443: connect: connection refused Dec 16 09:39:31.490195 kubelet[2353]: E1216 09:39:31.490059 2353 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://138.199.145.148:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 138.199.145.148:6443: connect: connection refused" logger="UnhandledError" Dec 16 09:39:31.568705 kubelet[2353]: W1216 09:39:31.567495 2353 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://138.199.145.148:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081-2-1-7-459a6b42ef&limit=500&resourceVersion=0": dial tcp 138.199.145.148:6443: connect: connection refused Dec 16 09:39:31.568705 kubelet[2353]: E1216 09:39:31.567613 2353 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://138.199.145.148:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081-2-1-7-459a6b42ef&limit=500&resourceVersion=0\": dial tcp 138.199.145.148:6443: connect: connection refused" logger="UnhandledError" Dec 16 09:39:31.591924 kubelet[2353]: W1216 09:39:31.591847 2353 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://138.199.145.148:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 138.199.145.148:6443: connect: connection refused Dec 16 09:39:31.591924 kubelet[2353]: E1216 09:39:31.591887 2353 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://138.199.145.148:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 138.199.145.148:6443: connect: connection refused" logger="UnhandledError" Dec 16 09:39:31.675368 kubelet[2353]: I1216 09:39:31.674991 2353 kubelet_node_status.go:72] "Attempting to register node" node="ci-4081-2-1-7-459a6b42ef" Dec 16 09:39:31.679590 kubelet[2353]: E1216 09:39:31.675931 2353 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://138.199.145.148:6443/api/v1/nodes\": dial tcp 138.199.145.148:6443: connect: connection refused" node="ci-4081-2-1-7-459a6b42ef" Dec 16 09:39:33.159440 kubelet[2353]: E1216 09:39:33.159380 2353 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-4081-2-1-7-459a6b42ef\" not found" node="ci-4081-2-1-7-459a6b42ef" Dec 16 09:39:33.215703 kubelet[2353]: E1216 09:39:33.215589 2353 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{ci-4081-2-1-7-459a6b42ef.18119ed5344e18e2 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4081-2-1-7-459a6b42ef,UID:ci-4081-2-1-7-459a6b42ef,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4081-2-1-7-459a6b42ef,},FirstTimestamp:2024-12-16 09:39:30.057107682 +0000 UTC m=+0.547608261,LastTimestamp:2024-12-16 09:39:30.057107682 +0000 UTC m=+0.547608261,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4081-2-1-7-459a6b42ef,}" Dec 16 09:39:33.272530 kubelet[2353]: E1216 09:39:33.272431 2353 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{ci-4081-2-1-7-459a6b42ef.18119ed5353039dd default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4081-2-1-7-459a6b42ef,UID:ci-4081-2-1-7-459a6b42ef,APIVersion:,ResourceVersion:,FieldPath:,},Reason:InvalidDiskCapacity,Message:invalid capacity 0 on image filesystem,Source:EventSource{Component:kubelet,Host:ci-4081-2-1-7-459a6b42ef,},FirstTimestamp:2024-12-16 09:39:30.071927261 +0000 UTC m=+0.562427840,LastTimestamp:2024-12-16 09:39:30.071927261 +0000 UTC m=+0.562427840,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4081-2-1-7-459a6b42ef,}" Dec 16 09:39:33.279608 kubelet[2353]: I1216 09:39:33.279127 2353 kubelet_node_status.go:72] "Attempting to register node" node="ci-4081-2-1-7-459a6b42ef" Dec 16 09:39:33.289708 kubelet[2353]: I1216 09:39:33.289667 2353 kubelet_node_status.go:75] "Successfully registered node" node="ci-4081-2-1-7-459a6b42ef" Dec 16 09:39:33.289708 kubelet[2353]: E1216 09:39:33.289694 2353 kubelet_node_status.go:535] "Error updating node status, will retry" err="error getting node \"ci-4081-2-1-7-459a6b42ef\": node \"ci-4081-2-1-7-459a6b42ef\" not found" Dec 16 09:39:33.296145 kubelet[2353]: E1216 09:39:33.296122 2353 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4081-2-1-7-459a6b42ef\" not found" Dec 16 09:39:33.325901 kubelet[2353]: E1216 09:39:33.325813 2353 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{ci-4081-2-1-7-459a6b42ef.18119ed53731c0d1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4081-2-1-7-459a6b42ef,UID:ci-4081-2-1-7-459a6b42ef,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node ci-4081-2-1-7-459a6b42ef status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:ci-4081-2-1-7-459a6b42ef,},FirstTimestamp:2024-12-16 09:39:30.105581777 +0000 UTC m=+0.596082356,LastTimestamp:2024-12-16 09:39:30.105581777 +0000 UTC m=+0.596082356,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4081-2-1-7-459a6b42ef,}" Dec 16 09:39:33.378711 kubelet[2353]: E1216 09:39:33.378620 2353 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{ci-4081-2-1-7-459a6b42ef.18119ed53731d17d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4081-2-1-7-459a6b42ef,UID:ci-4081-2-1-7-459a6b42ef,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node ci-4081-2-1-7-459a6b42ef status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:ci-4081-2-1-7-459a6b42ef,},FirstTimestamp:2024-12-16 09:39:30.105586045 +0000 UTC m=+0.596086624,LastTimestamp:2024-12-16 09:39:30.105586045 +0000 UTC m=+0.596086624,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4081-2-1-7-459a6b42ef,}" Dec 16 09:39:33.396708 kubelet[2353]: E1216 09:39:33.396683 2353 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4081-2-1-7-459a6b42ef\" not found" Dec 16 09:39:33.496869 kubelet[2353]: E1216 09:39:33.496795 2353 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4081-2-1-7-459a6b42ef\" not found" Dec 16 09:39:33.597411 kubelet[2353]: E1216 09:39:33.597334 2353 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4081-2-1-7-459a6b42ef\" not found" Dec 16 09:39:33.697976 kubelet[2353]: E1216 09:39:33.697933 2353 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4081-2-1-7-459a6b42ef\" not found" Dec 16 09:39:33.798690 kubelet[2353]: E1216 09:39:33.798552 2353 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4081-2-1-7-459a6b42ef\" not found" Dec 16 09:39:33.899244 kubelet[2353]: E1216 09:39:33.899186 2353 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4081-2-1-7-459a6b42ef\" not found" Dec 16 09:39:33.999884 kubelet[2353]: E1216 09:39:33.999836 2353 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4081-2-1-7-459a6b42ef\" not found" Dec 16 09:39:34.100731 kubelet[2353]: E1216 09:39:34.100541 2353 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4081-2-1-7-459a6b42ef\" not found" Dec 16 09:39:34.948588 systemd[1]: Reloading requested from client PID 2628 ('systemctl') (unit session-7.scope)... Dec 16 09:39:34.948605 systemd[1]: Reloading... Dec 16 09:39:35.044652 zram_generator::config[2671]: No configuration found. Dec 16 09:39:35.055620 kubelet[2353]: I1216 09:39:35.054076 2353 apiserver.go:52] "Watching apiserver" Dec 16 09:39:35.071548 kubelet[2353]: I1216 09:39:35.071511 2353 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 16 09:39:35.149387 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Dec 16 09:39:35.227457 systemd[1]: Reloading finished in 278 ms. Dec 16 09:39:35.272931 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Dec 16 09:39:35.285175 systemd[1]: kubelet.service: Deactivated successfully. Dec 16 09:39:35.285474 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Dec 16 09:39:35.296854 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 16 09:39:35.419148 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 16 09:39:35.423814 (kubelet)[2719]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Dec 16 09:39:35.466876 kubelet[2719]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 16 09:39:35.466876 kubelet[2719]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 16 09:39:35.466876 kubelet[2719]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 16 09:39:35.466876 kubelet[2719]: I1216 09:39:35.466657 2719 server.go:206] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 16 09:39:35.477598 kubelet[2719]: I1216 09:39:35.476709 2719 server.go:486] "Kubelet version" kubeletVersion="v1.31.0" Dec 16 09:39:35.477598 kubelet[2719]: I1216 09:39:35.476738 2719 server.go:488] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 16 09:39:35.477598 kubelet[2719]: I1216 09:39:35.477158 2719 server.go:929] "Client rotation is on, will bootstrap in background" Dec 16 09:39:35.480060 kubelet[2719]: I1216 09:39:35.479508 2719 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 16 09:39:35.484267 kubelet[2719]: I1216 09:39:35.484228 2719 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Dec 16 09:39:35.487712 kubelet[2719]: E1216 09:39:35.487677 2719 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Dec 16 09:39:35.487832 kubelet[2719]: I1216 09:39:35.487819 2719 server.go:1403] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Dec 16 09:39:35.491289 kubelet[2719]: I1216 09:39:35.491274 2719 server.go:744] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Dec 16 09:39:35.491447 kubelet[2719]: I1216 09:39:35.491435 2719 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 16 09:39:35.491734 kubelet[2719]: I1216 09:39:35.491706 2719 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 16 09:39:35.491928 kubelet[2719]: I1216 09:39:35.491786 2719 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4081-2-1-7-459a6b42ef","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 16 09:39:35.492032 kubelet[2719]: I1216 09:39:35.492021 2719 topology_manager.go:138] "Creating topology manager with none policy" Dec 16 09:39:35.492116 kubelet[2719]: I1216 09:39:35.492095 2719 container_manager_linux.go:300] "Creating device plugin manager" Dec 16 09:39:35.492302 kubelet[2719]: I1216 09:39:35.492278 2719 state_mem.go:36] "Initialized new in-memory state store" Dec 16 09:39:35.495200 kubelet[2719]: I1216 09:39:35.495175 2719 kubelet.go:408] "Attempting to sync node with API server" Dec 16 09:39:35.495294 kubelet[2719]: I1216 09:39:35.495284 2719 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 16 09:39:35.495368 kubelet[2719]: I1216 09:39:35.495359 2719 kubelet.go:314] "Adding apiserver pod source" Dec 16 09:39:35.495422 kubelet[2719]: I1216 09:39:35.495414 2719 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 16 09:39:35.499279 kubelet[2719]: I1216 09:39:35.499083 2719 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Dec 16 09:39:35.500918 kubelet[2719]: I1216 09:39:35.500081 2719 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 16 09:39:35.513777 kubelet[2719]: I1216 09:39:35.513718 2719 server.go:1269] "Started kubelet" Dec 16 09:39:35.515159 kubelet[2719]: I1216 09:39:35.515141 2719 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 16 09:39:35.521123 kubelet[2719]: I1216 09:39:35.521107 2719 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 16 09:39:35.521376 kubelet[2719]: I1216 09:39:35.516180 2719 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 16 09:39:35.523969 kubelet[2719]: I1216 09:39:35.523952 2719 server.go:460] "Adding debug handlers to kubelet server" Dec 16 09:39:35.529474 kubelet[2719]: I1216 09:39:35.529417 2719 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 16 09:39:35.529646 kubelet[2719]: I1216 09:39:35.516230 2719 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 16 09:39:35.529933 kubelet[2719]: I1216 09:39:35.529917 2719 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 16 09:39:35.530148 kubelet[2719]: I1216 09:39:35.530135 2719 reconciler.go:26] "Reconciler: start to sync state" Dec 16 09:39:35.530208 kubelet[2719]: I1216 09:39:35.516949 2719 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Dec 16 09:39:35.533909 kubelet[2719]: I1216 09:39:35.533736 2719 factory.go:221] Registration of the systemd container factory successfully Dec 16 09:39:35.533909 kubelet[2719]: I1216 09:39:35.533852 2719 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Dec 16 09:39:35.536948 kubelet[2719]: I1216 09:39:35.535906 2719 factory.go:221] Registration of the containerd container factory successfully Dec 16 09:39:35.542754 kubelet[2719]: I1216 09:39:35.541616 2719 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 16 09:39:35.542969 kubelet[2719]: I1216 09:39:35.542884 2719 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 16 09:39:35.542969 kubelet[2719]: I1216 09:39:35.542911 2719 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 16 09:39:35.542969 kubelet[2719]: I1216 09:39:35.542952 2719 kubelet.go:2321] "Starting kubelet main sync loop" Dec 16 09:39:35.543364 kubelet[2719]: E1216 09:39:35.542987 2719 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 16 09:39:35.551006 kubelet[2719]: E1216 09:39:35.550981 2719 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Dec 16 09:39:35.589003 kubelet[2719]: I1216 09:39:35.588979 2719 cpu_manager.go:214] "Starting CPU manager" policy="none" Dec 16 09:39:35.589400 kubelet[2719]: I1216 09:39:35.589127 2719 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Dec 16 09:39:35.589400 kubelet[2719]: I1216 09:39:35.589151 2719 state_mem.go:36] "Initialized new in-memory state store" Dec 16 09:39:35.589400 kubelet[2719]: I1216 09:39:35.589296 2719 state_mem.go:88] "Updated default CPUSet" cpuSet="" Dec 16 09:39:35.589400 kubelet[2719]: I1216 09:39:35.589308 2719 state_mem.go:96] "Updated CPUSet assignments" assignments={} Dec 16 09:39:35.589400 kubelet[2719]: I1216 09:39:35.589326 2719 policy_none.go:49] "None policy: Start" Dec 16 09:39:35.589843 kubelet[2719]: I1216 09:39:35.589817 2719 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 16 09:39:35.589914 kubelet[2719]: I1216 09:39:35.589903 2719 state_mem.go:35] "Initializing new in-memory state store" Dec 16 09:39:35.590086 kubelet[2719]: I1216 09:39:35.590074 2719 state_mem.go:75] "Updated machine memory state" Dec 16 09:39:35.594325 kubelet[2719]: I1216 09:39:35.594294 2719 manager.go:510] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 16 09:39:35.594643 kubelet[2719]: I1216 09:39:35.594524 2719 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 16 09:39:35.594643 kubelet[2719]: I1216 09:39:35.594549 2719 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 16 09:39:35.595028 kubelet[2719]: I1216 09:39:35.594966 2719 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 16 09:39:35.651620 kubelet[2719]: E1216 09:39:35.651520 2719 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-ci-4081-2-1-7-459a6b42ef\" already exists" pod="kube-system/kube-apiserver-ci-4081-2-1-7-459a6b42ef" Dec 16 09:39:35.707592 kubelet[2719]: I1216 09:39:35.707533 2719 kubelet_node_status.go:72] "Attempting to register node" node="ci-4081-2-1-7-459a6b42ef" Dec 16 09:39:35.715606 kubelet[2719]: I1216 09:39:35.715126 2719 kubelet_node_status.go:111] "Node was previously registered" node="ci-4081-2-1-7-459a6b42ef" Dec 16 09:39:35.715606 kubelet[2719]: I1216 09:39:35.715191 2719 kubelet_node_status.go:75] "Successfully registered node" node="ci-4081-2-1-7-459a6b42ef" Dec 16 09:39:35.730911 kubelet[2719]: I1216 09:39:35.730774 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/24f5c188f00bb4e278359e82a7e1823f-kubeconfig\") pod \"kube-scheduler-ci-4081-2-1-7-459a6b42ef\" (UID: \"24f5c188f00bb4e278359e82a7e1823f\") " pod="kube-system/kube-scheduler-ci-4081-2-1-7-459a6b42ef" Dec 16 09:39:35.831682 kubelet[2719]: I1216 09:39:35.831633 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/051a64e47cbe729b1fee0d93ec1faff5-ca-certs\") pod \"kube-apiserver-ci-4081-2-1-7-459a6b42ef\" (UID: \"051a64e47cbe729b1fee0d93ec1faff5\") " pod="kube-system/kube-apiserver-ci-4081-2-1-7-459a6b42ef" Dec 16 09:39:35.831682 kubelet[2719]: I1216 09:39:35.831672 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/15c858ba29bf2c6e9d716382efcb2dbf-k8s-certs\") pod \"kube-controller-manager-ci-4081-2-1-7-459a6b42ef\" (UID: \"15c858ba29bf2c6e9d716382efcb2dbf\") " pod="kube-system/kube-controller-manager-ci-4081-2-1-7-459a6b42ef" Dec 16 09:39:35.831820 kubelet[2719]: I1216 09:39:35.831717 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/15c858ba29bf2c6e9d716382efcb2dbf-kubeconfig\") pod \"kube-controller-manager-ci-4081-2-1-7-459a6b42ef\" (UID: \"15c858ba29bf2c6e9d716382efcb2dbf\") " pod="kube-system/kube-controller-manager-ci-4081-2-1-7-459a6b42ef" Dec 16 09:39:35.831820 kubelet[2719]: I1216 09:39:35.831741 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/051a64e47cbe729b1fee0d93ec1faff5-k8s-certs\") pod \"kube-apiserver-ci-4081-2-1-7-459a6b42ef\" (UID: \"051a64e47cbe729b1fee0d93ec1faff5\") " pod="kube-system/kube-apiserver-ci-4081-2-1-7-459a6b42ef" Dec 16 09:39:35.831820 kubelet[2719]: I1216 09:39:35.831766 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/051a64e47cbe729b1fee0d93ec1faff5-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4081-2-1-7-459a6b42ef\" (UID: \"051a64e47cbe729b1fee0d93ec1faff5\") " pod="kube-system/kube-apiserver-ci-4081-2-1-7-459a6b42ef" Dec 16 09:39:35.831820 kubelet[2719]: I1216 09:39:35.831786 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/15c858ba29bf2c6e9d716382efcb2dbf-ca-certs\") pod \"kube-controller-manager-ci-4081-2-1-7-459a6b42ef\" (UID: \"15c858ba29bf2c6e9d716382efcb2dbf\") " pod="kube-system/kube-controller-manager-ci-4081-2-1-7-459a6b42ef" Dec 16 09:39:35.831820 kubelet[2719]: I1216 09:39:35.831813 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/15c858ba29bf2c6e9d716382efcb2dbf-flexvolume-dir\") pod \"kube-controller-manager-ci-4081-2-1-7-459a6b42ef\" (UID: \"15c858ba29bf2c6e9d716382efcb2dbf\") " pod="kube-system/kube-controller-manager-ci-4081-2-1-7-459a6b42ef" Dec 16 09:39:35.832006 kubelet[2719]: I1216 09:39:35.831832 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/15c858ba29bf2c6e9d716382efcb2dbf-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4081-2-1-7-459a6b42ef\" (UID: \"15c858ba29bf2c6e9d716382efcb2dbf\") " pod="kube-system/kube-controller-manager-ci-4081-2-1-7-459a6b42ef" Dec 16 09:39:35.957811 sudo[2751]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Dec 16 09:39:35.958476 sudo[2751]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=0) Dec 16 09:39:36.478750 sudo[2751]: pam_unix(sudo:session): session closed for user root Dec 16 09:39:36.496057 kubelet[2719]: I1216 09:39:36.496012 2719 apiserver.go:52] "Watching apiserver" Dec 16 09:39:36.531058 kubelet[2719]: I1216 09:39:36.530145 2719 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 16 09:39:36.580586 kubelet[2719]: E1216 09:39:36.580326 2719 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-ci-4081-2-1-7-459a6b42ef\" already exists" pod="kube-system/kube-apiserver-ci-4081-2-1-7-459a6b42ef" Dec 16 09:39:36.602185 kubelet[2719]: I1216 09:39:36.601823 2719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4081-2-1-7-459a6b42ef" podStartSLOduration=1.601809239 podStartE2EDuration="1.601809239s" podCreationTimestamp="2024-12-16 09:39:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-12-16 09:39:36.594060144 +0000 UTC m=+1.165583517" watchObservedRunningTime="2024-12-16 09:39:36.601809239 +0000 UTC m=+1.173332611" Dec 16 09:39:36.609478 kubelet[2719]: I1216 09:39:36.609431 2719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4081-2-1-7-459a6b42ef" podStartSLOduration=1.609338303 podStartE2EDuration="1.609338303s" podCreationTimestamp="2024-12-16 09:39:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-12-16 09:39:36.601991871 +0000 UTC m=+1.173515263" watchObservedRunningTime="2024-12-16 09:39:36.609338303 +0000 UTC m=+1.180861675" Dec 16 09:39:36.610271 kubelet[2719]: I1216 09:39:36.610180 2719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4081-2-1-7-459a6b42ef" podStartSLOduration=1.6101724210000001 podStartE2EDuration="1.610172421s" podCreationTimestamp="2024-12-16 09:39:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-12-16 09:39:36.608880067 +0000 UTC m=+1.180403440" watchObservedRunningTime="2024-12-16 09:39:36.610172421 +0000 UTC m=+1.181695793" Dec 16 09:39:37.730145 sudo[1890]: pam_unix(sudo:session): session closed for user root Dec 16 09:39:37.889182 sshd[1887]: pam_unix(sshd:session): session closed for user core Dec 16 09:39:37.893378 systemd[1]: sshd@6-138.199.145.148:22-147.75.109.163:33270.service: Deactivated successfully. Dec 16 09:39:37.895861 systemd[1]: session-7.scope: Deactivated successfully. Dec 16 09:39:37.896061 systemd[1]: session-7.scope: Consumed 3.858s CPU time, 143.9M memory peak, 0B memory swap peak. Dec 16 09:39:37.896776 systemd-logind[1477]: Session 7 logged out. Waiting for processes to exit. Dec 16 09:39:37.898297 systemd-logind[1477]: Removed session 7. Dec 16 09:39:42.020602 kubelet[2719]: I1216 09:39:42.020545 2719 kuberuntime_manager.go:1633] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Dec 16 09:39:42.023077 containerd[1489]: time="2024-12-16T09:39:42.023034374Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Dec 16 09:39:42.023865 kubelet[2719]: I1216 09:39:42.023311 2719 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Dec 16 09:39:42.934279 systemd[1]: Created slice kubepods-besteffort-pod3bef8a32_1b17_4c62_b284_964bf10bfca2.slice - libcontainer container kubepods-besteffort-pod3bef8a32_1b17_4c62_b284_964bf10bfca2.slice. Dec 16 09:39:42.945259 systemd[1]: Created slice kubepods-burstable-pod81563a72_3610_425a_9c84_4ec0b9f36bc6.slice - libcontainer container kubepods-burstable-pod81563a72_3610_425a_9c84_4ec0b9f36bc6.slice. Dec 16 09:39:42.980077 kubelet[2719]: I1216 09:39:42.980044 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-hostproc\") pod \"cilium-6j2cb\" (UID: \"81563a72-3610-425a-9c84-4ec0b9f36bc6\") " pod="kube-system/cilium-6j2cb" Dec 16 09:39:42.980548 kubelet[2719]: I1216 09:39:42.980284 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48vfc\" (UniqueName: \"kubernetes.io/projected/81563a72-3610-425a-9c84-4ec0b9f36bc6-kube-api-access-48vfc\") pod \"cilium-6j2cb\" (UID: \"81563a72-3610-425a-9c84-4ec0b9f36bc6\") " pod="kube-system/cilium-6j2cb" Dec 16 09:39:42.980548 kubelet[2719]: I1216 09:39:42.980306 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-bpf-maps\") pod \"cilium-6j2cb\" (UID: \"81563a72-3610-425a-9c84-4ec0b9f36bc6\") " pod="kube-system/cilium-6j2cb" Dec 16 09:39:42.980548 kubelet[2719]: I1216 09:39:42.980320 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-lib-modules\") pod \"cilium-6j2cb\" (UID: \"81563a72-3610-425a-9c84-4ec0b9f36bc6\") " pod="kube-system/cilium-6j2cb" Dec 16 09:39:42.980548 kubelet[2719]: I1216 09:39:42.980333 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-host-proc-sys-net\") pod \"cilium-6j2cb\" (UID: \"81563a72-3610-425a-9c84-4ec0b9f36bc6\") " pod="kube-system/cilium-6j2cb" Dec 16 09:39:42.980548 kubelet[2719]: I1216 09:39:42.980346 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3bef8a32-1b17-4c62-b284-964bf10bfca2-lib-modules\") pod \"kube-proxy-svk9n\" (UID: \"3bef8a32-1b17-4c62-b284-964bf10bfca2\") " pod="kube-system/kube-proxy-svk9n" Dec 16 09:39:42.980548 kubelet[2719]: I1216 09:39:42.980358 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/81563a72-3610-425a-9c84-4ec0b9f36bc6-cilium-config-path\") pod \"cilium-6j2cb\" (UID: \"81563a72-3610-425a-9c84-4ec0b9f36bc6\") " pod="kube-system/cilium-6j2cb" Dec 16 09:39:42.980720 kubelet[2719]: I1216 09:39:42.980370 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-cni-path\") pod \"cilium-6j2cb\" (UID: \"81563a72-3610-425a-9c84-4ec0b9f36bc6\") " pod="kube-system/cilium-6j2cb" Dec 16 09:39:42.980720 kubelet[2719]: I1216 09:39:42.980381 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/81563a72-3610-425a-9c84-4ec0b9f36bc6-hubble-tls\") pod \"cilium-6j2cb\" (UID: \"81563a72-3610-425a-9c84-4ec0b9f36bc6\") " pod="kube-system/cilium-6j2cb" Dec 16 09:39:42.980720 kubelet[2719]: I1216 09:39:42.980395 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/3bef8a32-1b17-4c62-b284-964bf10bfca2-xtables-lock\") pod \"kube-proxy-svk9n\" (UID: \"3bef8a32-1b17-4c62-b284-964bf10bfca2\") " pod="kube-system/kube-proxy-svk9n" Dec 16 09:39:42.980720 kubelet[2719]: I1216 09:39:42.980408 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-cilium-cgroup\") pod \"cilium-6j2cb\" (UID: \"81563a72-3610-425a-9c84-4ec0b9f36bc6\") " pod="kube-system/cilium-6j2cb" Dec 16 09:39:42.980720 kubelet[2719]: I1216 09:39:42.980422 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-host-proc-sys-kernel\") pod \"cilium-6j2cb\" (UID: \"81563a72-3610-425a-9c84-4ec0b9f36bc6\") " pod="kube-system/cilium-6j2cb" Dec 16 09:39:42.980720 kubelet[2719]: I1216 09:39:42.980435 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/81563a72-3610-425a-9c84-4ec0b9f36bc6-clustermesh-secrets\") pod \"cilium-6j2cb\" (UID: \"81563a72-3610-425a-9c84-4ec0b9f36bc6\") " pod="kube-system/cilium-6j2cb" Dec 16 09:39:42.980833 kubelet[2719]: I1216 09:39:42.980449 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/3bef8a32-1b17-4c62-b284-964bf10bfca2-kube-proxy\") pod \"kube-proxy-svk9n\" (UID: \"3bef8a32-1b17-4c62-b284-964bf10bfca2\") " pod="kube-system/kube-proxy-svk9n" Dec 16 09:39:42.980833 kubelet[2719]: I1216 09:39:42.980463 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cr9f8\" (UniqueName: \"kubernetes.io/projected/3bef8a32-1b17-4c62-b284-964bf10bfca2-kube-api-access-cr9f8\") pod \"kube-proxy-svk9n\" (UID: \"3bef8a32-1b17-4c62-b284-964bf10bfca2\") " pod="kube-system/kube-proxy-svk9n" Dec 16 09:39:42.980833 kubelet[2719]: I1216 09:39:42.980478 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-cilium-run\") pod \"cilium-6j2cb\" (UID: \"81563a72-3610-425a-9c84-4ec0b9f36bc6\") " pod="kube-system/cilium-6j2cb" Dec 16 09:39:42.980833 kubelet[2719]: I1216 09:39:42.980494 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-xtables-lock\") pod \"cilium-6j2cb\" (UID: \"81563a72-3610-425a-9c84-4ec0b9f36bc6\") " pod="kube-system/cilium-6j2cb" Dec 16 09:39:42.980833 kubelet[2719]: I1216 09:39:42.980508 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-etc-cni-netd\") pod \"cilium-6j2cb\" (UID: \"81563a72-3610-425a-9c84-4ec0b9f36bc6\") " pod="kube-system/cilium-6j2cb" Dec 16 09:39:43.106551 systemd[1]: Created slice kubepods-besteffort-pod67d9cab9_f389_4c1c_a6e6_319040d73a93.slice - libcontainer container kubepods-besteffort-pod67d9cab9_f389_4c1c_a6e6_319040d73a93.slice. Dec 16 09:39:43.183653 kubelet[2719]: I1216 09:39:43.183547 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/67d9cab9-f389-4c1c-a6e6-319040d73a93-cilium-config-path\") pod \"cilium-operator-5d85765b45-4jzvz\" (UID: \"67d9cab9-f389-4c1c-a6e6-319040d73a93\") " pod="kube-system/cilium-operator-5d85765b45-4jzvz" Dec 16 09:39:43.183653 kubelet[2719]: I1216 09:39:43.183617 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmcrw\" (UniqueName: \"kubernetes.io/projected/67d9cab9-f389-4c1c-a6e6-319040d73a93-kube-api-access-jmcrw\") pod \"cilium-operator-5d85765b45-4jzvz\" (UID: \"67d9cab9-f389-4c1c-a6e6-319040d73a93\") " pod="kube-system/cilium-operator-5d85765b45-4jzvz" Dec 16 09:39:43.249322 containerd[1489]: time="2024-12-16T09:39:43.249290523Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-svk9n,Uid:3bef8a32-1b17-4c62-b284-964bf10bfca2,Namespace:kube-system,Attempt:0,}" Dec 16 09:39:43.258337 containerd[1489]: time="2024-12-16T09:39:43.258278490Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-6j2cb,Uid:81563a72-3610-425a-9c84-4ec0b9f36bc6,Namespace:kube-system,Attempt:0,}" Dec 16 09:39:43.277594 containerd[1489]: time="2024-12-16T09:39:43.277063386Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 16 09:39:43.277594 containerd[1489]: time="2024-12-16T09:39:43.277170756Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 16 09:39:43.277594 containerd[1489]: time="2024-12-16T09:39:43.277223155Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 16 09:39:43.278041 containerd[1489]: time="2024-12-16T09:39:43.277867098Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 16 09:39:43.302065 containerd[1489]: time="2024-12-16T09:39:43.301975623Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 16 09:39:43.303637 containerd[1489]: time="2024-12-16T09:39:43.302087262Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 16 09:39:43.303637 containerd[1489]: time="2024-12-16T09:39:43.302128469Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 16 09:39:43.303637 containerd[1489]: time="2024-12-16T09:39:43.302226372Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 16 09:39:43.307716 systemd[1]: Started cri-containerd-6cd43fe3df065b555c71bb8496fe1e27072a769db0ede1317fb18cc96a81fd3e.scope - libcontainer container 6cd43fe3df065b555c71bb8496fe1e27072a769db0ede1317fb18cc96a81fd3e. Dec 16 09:39:43.326685 systemd[1]: Started cri-containerd-cabc0575ef95bbb1469e40d05f6827c256e2aaae5cb89b9ef5cb1790b3ba4225.scope - libcontainer container cabc0575ef95bbb1469e40d05f6827c256e2aaae5cb89b9ef5cb1790b3ba4225. Dec 16 09:39:43.347092 containerd[1489]: time="2024-12-16T09:39:43.347046413Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-svk9n,Uid:3bef8a32-1b17-4c62-b284-964bf10bfca2,Namespace:kube-system,Attempt:0,} returns sandbox id \"6cd43fe3df065b555c71bb8496fe1e27072a769db0ede1317fb18cc96a81fd3e\"" Dec 16 09:39:43.351731 containerd[1489]: time="2024-12-16T09:39:43.351698797Z" level=info msg="CreateContainer within sandbox \"6cd43fe3df065b555c71bb8496fe1e27072a769db0ede1317fb18cc96a81fd3e\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Dec 16 09:39:43.365469 containerd[1489]: time="2024-12-16T09:39:43.365396887Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-6j2cb,Uid:81563a72-3610-425a-9c84-4ec0b9f36bc6,Namespace:kube-system,Attempt:0,} returns sandbox id \"cabc0575ef95bbb1469e40d05f6827c256e2aaae5cb89b9ef5cb1790b3ba4225\"" Dec 16 09:39:43.369383 containerd[1489]: time="2024-12-16T09:39:43.369338373Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Dec 16 09:39:43.376320 containerd[1489]: time="2024-12-16T09:39:43.376282630Z" level=info msg="CreateContainer within sandbox \"6cd43fe3df065b555c71bb8496fe1e27072a769db0ede1317fb18cc96a81fd3e\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"2e9726278f61d1123570b1d404f78837124be96b7a2e2c2fd227a0a1537d8e98\"" Dec 16 09:39:43.378213 containerd[1489]: time="2024-12-16T09:39:43.376952322Z" level=info msg="StartContainer for \"2e9726278f61d1123570b1d404f78837124be96b7a2e2c2fd227a0a1537d8e98\"" Dec 16 09:39:43.408748 systemd[1]: Started cri-containerd-2e9726278f61d1123570b1d404f78837124be96b7a2e2c2fd227a0a1537d8e98.scope - libcontainer container 2e9726278f61d1123570b1d404f78837124be96b7a2e2c2fd227a0a1537d8e98. Dec 16 09:39:43.413170 containerd[1489]: time="2024-12-16T09:39:43.413133218Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-5d85765b45-4jzvz,Uid:67d9cab9-f389-4c1c-a6e6-319040d73a93,Namespace:kube-system,Attempt:0,}" Dec 16 09:39:43.435877 containerd[1489]: time="2024-12-16T09:39:43.435777348Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 16 09:39:43.435877 containerd[1489]: time="2024-12-16T09:39:43.435836848Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 16 09:39:43.436348 containerd[1489]: time="2024-12-16T09:39:43.436288903Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 16 09:39:43.436608 containerd[1489]: time="2024-12-16T09:39:43.436518092Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 16 09:39:43.452334 containerd[1489]: time="2024-12-16T09:39:43.452243891Z" level=info msg="StartContainer for \"2e9726278f61d1123570b1d404f78837124be96b7a2e2c2fd227a0a1537d8e98\" returns successfully" Dec 16 09:39:43.459759 systemd[1]: Started cri-containerd-3b87f8f1710867438768b3941d4b273a4a8bc234af9b1b40cb15b9c0ba524cc0.scope - libcontainer container 3b87f8f1710867438768b3941d4b273a4a8bc234af9b1b40cb15b9c0ba524cc0. Dec 16 09:39:43.508350 containerd[1489]: time="2024-12-16T09:39:43.507886578Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-5d85765b45-4jzvz,Uid:67d9cab9-f389-4c1c-a6e6-319040d73a93,Namespace:kube-system,Attempt:0,} returns sandbox id \"3b87f8f1710867438768b3941d4b273a4a8bc234af9b1b40cb15b9c0ba524cc0\"" Dec 16 09:39:45.375791 kubelet[2719]: I1216 09:39:45.375719 2719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-svk9n" podStartSLOduration=3.375699872 podStartE2EDuration="3.375699872s" podCreationTimestamp="2024-12-16 09:39:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-12-16 09:39:43.604189442 +0000 UTC m=+8.175712814" watchObservedRunningTime="2024-12-16 09:39:45.375699872 +0000 UTC m=+9.947223254" Dec 16 09:39:50.281957 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4208485850.mount: Deactivated successfully. Dec 16 09:39:52.565707 containerd[1489]: time="2024-12-16T09:39:52.565646687Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 09:39:52.566650 containerd[1489]: time="2024-12-16T09:39:52.566605911Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=166735283" Dec 16 09:39:52.567346 containerd[1489]: time="2024-12-16T09:39:52.567312061Z" level=info msg="ImageCreate event name:\"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 09:39:52.568990 containerd[1489]: time="2024-12-16T09:39:52.568956697Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"166719855\" in 9.199586384s" Dec 16 09:39:52.568990 containerd[1489]: time="2024-12-16T09:39:52.568985569Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"" Dec 16 09:39:52.571781 containerd[1489]: time="2024-12-16T09:39:52.571531160Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Dec 16 09:39:52.573158 containerd[1489]: time="2024-12-16T09:39:52.572881906Z" level=info msg="CreateContainer within sandbox \"cabc0575ef95bbb1469e40d05f6827c256e2aaae5cb89b9ef5cb1790b3ba4225\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Dec 16 09:39:52.676576 containerd[1489]: time="2024-12-16T09:39:52.676521531Z" level=info msg="CreateContainer within sandbox \"cabc0575ef95bbb1469e40d05f6827c256e2aaae5cb89b9ef5cb1790b3ba4225\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"95cea8638f01368f96be90942a0735ec3cc96c43a2e920f16ee973799bff7f2b\"" Dec 16 09:39:52.677732 containerd[1489]: time="2024-12-16T09:39:52.677041313Z" level=info msg="StartContainer for \"95cea8638f01368f96be90942a0735ec3cc96c43a2e920f16ee973799bff7f2b\"" Dec 16 09:39:52.751700 systemd[1]: Started cri-containerd-95cea8638f01368f96be90942a0735ec3cc96c43a2e920f16ee973799bff7f2b.scope - libcontainer container 95cea8638f01368f96be90942a0735ec3cc96c43a2e920f16ee973799bff7f2b. Dec 16 09:39:52.779532 containerd[1489]: time="2024-12-16T09:39:52.779449886Z" level=info msg="StartContainer for \"95cea8638f01368f96be90942a0735ec3cc96c43a2e920f16ee973799bff7f2b\" returns successfully" Dec 16 09:39:52.792546 systemd[1]: cri-containerd-95cea8638f01368f96be90942a0735ec3cc96c43a2e920f16ee973799bff7f2b.scope: Deactivated successfully. Dec 16 09:39:52.855200 containerd[1489]: time="2024-12-16T09:39:52.843230006Z" level=info msg="shim disconnected" id=95cea8638f01368f96be90942a0735ec3cc96c43a2e920f16ee973799bff7f2b namespace=k8s.io Dec 16 09:39:52.855200 containerd[1489]: time="2024-12-16T09:39:52.854758301Z" level=warning msg="cleaning up after shim disconnected" id=95cea8638f01368f96be90942a0735ec3cc96c43a2e920f16ee973799bff7f2b namespace=k8s.io Dec 16 09:39:52.855200 containerd[1489]: time="2024-12-16T09:39:52.854772327Z" level=info msg="cleaning up dead shim" namespace=k8s.io Dec 16 09:39:53.633958 containerd[1489]: time="2024-12-16T09:39:53.633912890Z" level=info msg="CreateContainer within sandbox \"cabc0575ef95bbb1469e40d05f6827c256e2aaae5cb89b9ef5cb1790b3ba4225\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Dec 16 09:39:53.653846 containerd[1489]: time="2024-12-16T09:39:53.653808189Z" level=info msg="CreateContainer within sandbox \"cabc0575ef95bbb1469e40d05f6827c256e2aaae5cb89b9ef5cb1790b3ba4225\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"76b958fd76ab693634764099cd1289981aa90f18dd8bffa90a5c04169b1c3928\"" Dec 16 09:39:53.654682 containerd[1489]: time="2024-12-16T09:39:53.654503119Z" level=info msg="StartContainer for \"76b958fd76ab693634764099cd1289981aa90f18dd8bffa90a5c04169b1c3928\"" Dec 16 09:39:53.668502 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-95cea8638f01368f96be90942a0735ec3cc96c43a2e920f16ee973799bff7f2b-rootfs.mount: Deactivated successfully. Dec 16 09:39:53.693085 systemd[1]: run-containerd-runc-k8s.io-76b958fd76ab693634764099cd1289981aa90f18dd8bffa90a5c04169b1c3928-runc.2hD7gf.mount: Deactivated successfully. Dec 16 09:39:53.702726 systemd[1]: Started cri-containerd-76b958fd76ab693634764099cd1289981aa90f18dd8bffa90a5c04169b1c3928.scope - libcontainer container 76b958fd76ab693634764099cd1289981aa90f18dd8bffa90a5c04169b1c3928. Dec 16 09:39:53.725429 containerd[1489]: time="2024-12-16T09:39:53.725376486Z" level=info msg="StartContainer for \"76b958fd76ab693634764099cd1289981aa90f18dd8bffa90a5c04169b1c3928\" returns successfully" Dec 16 09:39:53.738214 systemd[1]: systemd-sysctl.service: Deactivated successfully. Dec 16 09:39:53.739271 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Dec 16 09:39:53.739429 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables... Dec 16 09:39:53.743822 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Dec 16 09:39:53.744028 systemd[1]: cri-containerd-76b958fd76ab693634764099cd1289981aa90f18dd8bffa90a5c04169b1c3928.scope: Deactivated successfully. Dec 16 09:39:53.769476 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-76b958fd76ab693634764099cd1289981aa90f18dd8bffa90a5c04169b1c3928-rootfs.mount: Deactivated successfully. Dec 16 09:39:53.772783 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Dec 16 09:39:53.782357 containerd[1489]: time="2024-12-16T09:39:53.782305872Z" level=info msg="shim disconnected" id=76b958fd76ab693634764099cd1289981aa90f18dd8bffa90a5c04169b1c3928 namespace=k8s.io Dec 16 09:39:53.782357 containerd[1489]: time="2024-12-16T09:39:53.782356426Z" level=warning msg="cleaning up after shim disconnected" id=76b958fd76ab693634764099cd1289981aa90f18dd8bffa90a5c04169b1c3928 namespace=k8s.io Dec 16 09:39:53.782357 containerd[1489]: time="2024-12-16T09:39:53.782365744Z" level=info msg="cleaning up dead shim" namespace=k8s.io Dec 16 09:39:53.793599 containerd[1489]: time="2024-12-16T09:39:53.792953711Z" level=warning msg="cleanup warnings time=\"2024-12-16T09:39:53Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Dec 16 09:39:54.647771 containerd[1489]: time="2024-12-16T09:39:54.647614684Z" level=info msg="CreateContainer within sandbox \"cabc0575ef95bbb1469e40d05f6827c256e2aaae5cb89b9ef5cb1790b3ba4225\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Dec 16 09:39:54.672623 containerd[1489]: time="2024-12-16T09:39:54.672525845Z" level=info msg="CreateContainer within sandbox \"cabc0575ef95bbb1469e40d05f6827c256e2aaae5cb89b9ef5cb1790b3ba4225\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"7c5bcccdd565e4286227ae13392a9a1674ec0d8be5e1af592e4a3bcb00b300f2\"" Dec 16 09:39:54.673040 containerd[1489]: time="2024-12-16T09:39:54.673005491Z" level=info msg="StartContainer for \"7c5bcccdd565e4286227ae13392a9a1674ec0d8be5e1af592e4a3bcb00b300f2\"" Dec 16 09:39:54.712714 systemd[1]: Started cri-containerd-7c5bcccdd565e4286227ae13392a9a1674ec0d8be5e1af592e4a3bcb00b300f2.scope - libcontainer container 7c5bcccdd565e4286227ae13392a9a1674ec0d8be5e1af592e4a3bcb00b300f2. Dec 16 09:39:54.742433 containerd[1489]: time="2024-12-16T09:39:54.742386734Z" level=info msg="StartContainer for \"7c5bcccdd565e4286227ae13392a9a1674ec0d8be5e1af592e4a3bcb00b300f2\" returns successfully" Dec 16 09:39:54.747250 systemd[1]: cri-containerd-7c5bcccdd565e4286227ae13392a9a1674ec0d8be5e1af592e4a3bcb00b300f2.scope: Deactivated successfully. Dec 16 09:39:54.768664 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-7c5bcccdd565e4286227ae13392a9a1674ec0d8be5e1af592e4a3bcb00b300f2-rootfs.mount: Deactivated successfully. Dec 16 09:39:54.774380 containerd[1489]: time="2024-12-16T09:39:54.774305388Z" level=info msg="shim disconnected" id=7c5bcccdd565e4286227ae13392a9a1674ec0d8be5e1af592e4a3bcb00b300f2 namespace=k8s.io Dec 16 09:39:54.774380 containerd[1489]: time="2024-12-16T09:39:54.774352256Z" level=warning msg="cleaning up after shim disconnected" id=7c5bcccdd565e4286227ae13392a9a1674ec0d8be5e1af592e4a3bcb00b300f2 namespace=k8s.io Dec 16 09:39:54.774380 containerd[1489]: time="2024-12-16T09:39:54.774360782Z" level=info msg="cleaning up dead shim" namespace=k8s.io Dec 16 09:39:55.631458 containerd[1489]: time="2024-12-16T09:39:55.631387542Z" level=info msg="CreateContainer within sandbox \"cabc0575ef95bbb1469e40d05f6827c256e2aaae5cb89b9ef5cb1790b3ba4225\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Dec 16 09:39:55.662275 containerd[1489]: time="2024-12-16T09:39:55.662232551Z" level=info msg="CreateContainer within sandbox \"cabc0575ef95bbb1469e40d05f6827c256e2aaae5cb89b9ef5cb1790b3ba4225\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"2e8b7cafc7c2004533f17577ad79101e0aa4ebfadd6e3c7f4d6ed1ce5b3e7246\"" Dec 16 09:39:55.665542 containerd[1489]: time="2024-12-16T09:39:55.664665721Z" level=info msg="StartContainer for \"2e8b7cafc7c2004533f17577ad79101e0aa4ebfadd6e3c7f4d6ed1ce5b3e7246\"" Dec 16 09:39:55.702696 systemd[1]: Started cri-containerd-2e8b7cafc7c2004533f17577ad79101e0aa4ebfadd6e3c7f4d6ed1ce5b3e7246.scope - libcontainer container 2e8b7cafc7c2004533f17577ad79101e0aa4ebfadd6e3c7f4d6ed1ce5b3e7246. Dec 16 09:39:55.726477 systemd[1]: cri-containerd-2e8b7cafc7c2004533f17577ad79101e0aa4ebfadd6e3c7f4d6ed1ce5b3e7246.scope: Deactivated successfully. Dec 16 09:39:55.729532 containerd[1489]: time="2024-12-16T09:39:55.729450655Z" level=info msg="StartContainer for \"2e8b7cafc7c2004533f17577ad79101e0aa4ebfadd6e3c7f4d6ed1ce5b3e7246\" returns successfully" Dec 16 09:39:55.746374 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-2e8b7cafc7c2004533f17577ad79101e0aa4ebfadd6e3c7f4d6ed1ce5b3e7246-rootfs.mount: Deactivated successfully. Dec 16 09:39:55.750510 containerd[1489]: time="2024-12-16T09:39:55.750447588Z" level=info msg="shim disconnected" id=2e8b7cafc7c2004533f17577ad79101e0aa4ebfadd6e3c7f4d6ed1ce5b3e7246 namespace=k8s.io Dec 16 09:39:55.750510 containerd[1489]: time="2024-12-16T09:39:55.750508162Z" level=warning msg="cleaning up after shim disconnected" id=2e8b7cafc7c2004533f17577ad79101e0aa4ebfadd6e3c7f4d6ed1ce5b3e7246 namespace=k8s.io Dec 16 09:39:55.750510 containerd[1489]: time="2024-12-16T09:39:55.750516878Z" level=info msg="cleaning up dead shim" namespace=k8s.io Dec 16 09:39:56.164258 containerd[1489]: time="2024-12-16T09:39:56.164210301Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 09:39:56.164981 containerd[1489]: time="2024-12-16T09:39:56.164943402Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=18907181" Dec 16 09:39:56.165595 containerd[1489]: time="2024-12-16T09:39:56.165530259Z" level=info msg="ImageCreate event name:\"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 09:39:56.166727 containerd[1489]: time="2024-12-16T09:39:56.166701760Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"18897442\" in 3.595147318s" Dec 16 09:39:56.167080 containerd[1489]: time="2024-12-16T09:39:56.166730033Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"" Dec 16 09:39:56.171344 containerd[1489]: time="2024-12-16T09:39:56.171297705Z" level=info msg="CreateContainer within sandbox \"3b87f8f1710867438768b3941d4b273a4a8bc234af9b1b40cb15b9c0ba524cc0\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Dec 16 09:39:56.200898 containerd[1489]: time="2024-12-16T09:39:56.200859255Z" level=info msg="CreateContainer within sandbox \"3b87f8f1710867438768b3941d4b273a4a8bc234af9b1b40cb15b9c0ba524cc0\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"cbea43b6dc94dbdf299c057b62cb3bf562efd7c51c2518de0491b7378ffa0f61\"" Dec 16 09:39:56.202190 containerd[1489]: time="2024-12-16T09:39:56.202171068Z" level=info msg="StartContainer for \"cbea43b6dc94dbdf299c057b62cb3bf562efd7c51c2518de0491b7378ffa0f61\"" Dec 16 09:39:56.224722 systemd[1]: Started cri-containerd-cbea43b6dc94dbdf299c057b62cb3bf562efd7c51c2518de0491b7378ffa0f61.scope - libcontainer container cbea43b6dc94dbdf299c057b62cb3bf562efd7c51c2518de0491b7378ffa0f61. Dec 16 09:39:56.248475 containerd[1489]: time="2024-12-16T09:39:56.248440533Z" level=info msg="StartContainer for \"cbea43b6dc94dbdf299c057b62cb3bf562efd7c51c2518de0491b7378ffa0f61\" returns successfully" Dec 16 09:39:56.642199 containerd[1489]: time="2024-12-16T09:39:56.642119117Z" level=info msg="CreateContainer within sandbox \"cabc0575ef95bbb1469e40d05f6827c256e2aaae5cb89b9ef5cb1790b3ba4225\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Dec 16 09:39:56.672233 containerd[1489]: time="2024-12-16T09:39:56.670827413Z" level=info msg="CreateContainer within sandbox \"cabc0575ef95bbb1469e40d05f6827c256e2aaae5cb89b9ef5cb1790b3ba4225\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"2afccada9da9b8083d9189601319ae6737a931161aae85b051f870382555e281\"" Dec 16 09:39:56.671590 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1238062176.mount: Deactivated successfully. Dec 16 09:39:56.673989 containerd[1489]: time="2024-12-16T09:39:56.672457751Z" level=info msg="StartContainer for \"2afccada9da9b8083d9189601319ae6737a931161aae85b051f870382555e281\"" Dec 16 09:39:56.683102 kubelet[2719]: I1216 09:39:56.683042 2719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-5d85765b45-4jzvz" podStartSLOduration=1.024699362 podStartE2EDuration="13.683026484s" podCreationTimestamp="2024-12-16 09:39:43 +0000 UTC" firstStartedPulling="2024-12-16 09:39:43.509312813 +0000 UTC m=+8.080836185" lastFinishedPulling="2024-12-16 09:39:56.167639935 +0000 UTC m=+20.739163307" observedRunningTime="2024-12-16 09:39:56.682859632 +0000 UTC m=+21.254383004" watchObservedRunningTime="2024-12-16 09:39:56.683026484 +0000 UTC m=+21.254549856" Dec 16 09:39:56.718672 systemd[1]: run-containerd-runc-k8s.io-2afccada9da9b8083d9189601319ae6737a931161aae85b051f870382555e281-runc.wuYIn4.mount: Deactivated successfully. Dec 16 09:39:56.732694 systemd[1]: Started cri-containerd-2afccada9da9b8083d9189601319ae6737a931161aae85b051f870382555e281.scope - libcontainer container 2afccada9da9b8083d9189601319ae6737a931161aae85b051f870382555e281. Dec 16 09:39:56.800183 containerd[1489]: time="2024-12-16T09:39:56.800144215Z" level=info msg="StartContainer for \"2afccada9da9b8083d9189601319ae6737a931161aae85b051f870382555e281\" returns successfully" Dec 16 09:39:57.099035 kubelet[2719]: I1216 09:39:57.098522 2719 kubelet_node_status.go:488] "Fast updating node status as it just became ready" Dec 16 09:39:57.150616 systemd[1]: Created slice kubepods-burstable-pod0058c53d_6810_460b_be34_d644c721a145.slice - libcontainer container kubepods-burstable-pod0058c53d_6810_460b_be34_d644c721a145.slice. Dec 16 09:39:57.162728 systemd[1]: Created slice kubepods-burstable-podfae5c553_44f3_4fb1_a3ab_917443101ce4.slice - libcontainer container kubepods-burstable-podfae5c553_44f3_4fb1_a3ab_917443101ce4.slice. Dec 16 09:39:57.187487 kubelet[2719]: I1216 09:39:57.187354 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0058c53d-6810-460b-be34-d644c721a145-config-volume\") pod \"coredns-6f6b679f8f-srs4l\" (UID: \"0058c53d-6810-460b-be34-d644c721a145\") " pod="kube-system/coredns-6f6b679f8f-srs4l" Dec 16 09:39:57.187487 kubelet[2719]: I1216 09:39:57.187394 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2fmw\" (UniqueName: \"kubernetes.io/projected/fae5c553-44f3-4fb1-a3ab-917443101ce4-kube-api-access-j2fmw\") pod \"coredns-6f6b679f8f-m2f8q\" (UID: \"fae5c553-44f3-4fb1-a3ab-917443101ce4\") " pod="kube-system/coredns-6f6b679f8f-m2f8q" Dec 16 09:39:57.187487 kubelet[2719]: I1216 09:39:57.187414 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpldh\" (UniqueName: \"kubernetes.io/projected/0058c53d-6810-460b-be34-d644c721a145-kube-api-access-zpldh\") pod \"coredns-6f6b679f8f-srs4l\" (UID: \"0058c53d-6810-460b-be34-d644c721a145\") " pod="kube-system/coredns-6f6b679f8f-srs4l" Dec 16 09:39:57.187487 kubelet[2719]: I1216 09:39:57.187428 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fae5c553-44f3-4fb1-a3ab-917443101ce4-config-volume\") pod \"coredns-6f6b679f8f-m2f8q\" (UID: \"fae5c553-44f3-4fb1-a3ab-917443101ce4\") " pod="kube-system/coredns-6f6b679f8f-m2f8q" Dec 16 09:39:57.458003 containerd[1489]: time="2024-12-16T09:39:57.457405116Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-srs4l,Uid:0058c53d-6810-460b-be34-d644c721a145,Namespace:kube-system,Attempt:0,}" Dec 16 09:39:57.469514 containerd[1489]: time="2024-12-16T09:39:57.469303456Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-m2f8q,Uid:fae5c553-44f3-4fb1-a3ab-917443101ce4,Namespace:kube-system,Attempt:0,}" Dec 16 09:39:57.657169 kubelet[2719]: I1216 09:39:57.657066 2719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-6j2cb" podStartSLOduration=6.453331692 podStartE2EDuration="15.657052423s" podCreationTimestamp="2024-12-16 09:39:42 +0000 UTC" firstStartedPulling="2024-12-16 09:39:43.367662412 +0000 UTC m=+7.939185783" lastFinishedPulling="2024-12-16 09:39:52.571383143 +0000 UTC m=+17.142906514" observedRunningTime="2024-12-16 09:39:57.654139025 +0000 UTC m=+22.225662396" watchObservedRunningTime="2024-12-16 09:39:57.657052423 +0000 UTC m=+22.228575795" Dec 16 09:39:59.364363 systemd-networkd[1391]: cilium_host: Link UP Dec 16 09:39:59.369967 systemd-networkd[1391]: cilium_net: Link UP Dec 16 09:39:59.370426 systemd-networkd[1391]: cilium_net: Gained carrier Dec 16 09:39:59.377898 systemd-networkd[1391]: cilium_host: Gained carrier Dec 16 09:39:59.493982 systemd-networkd[1391]: cilium_vxlan: Link UP Dec 16 09:39:59.494161 systemd-networkd[1391]: cilium_vxlan: Gained carrier Dec 16 09:39:59.606710 systemd-networkd[1391]: cilium_net: Gained IPv6LL Dec 16 09:39:59.695256 systemd-networkd[1391]: cilium_host: Gained IPv6LL Dec 16 09:39:59.814591 kernel: NET: Registered PF_ALG protocol family Dec 16 09:40:00.453678 systemd-networkd[1391]: lxc_health: Link UP Dec 16 09:40:00.466836 systemd-networkd[1391]: lxc_health: Gained carrier Dec 16 09:40:00.606752 systemd-networkd[1391]: cilium_vxlan: Gained IPv6LL Dec 16 09:40:01.020961 systemd-networkd[1391]: lxcce470a9af226: Link UP Dec 16 09:40:01.029597 kernel: eth0: renamed from tmpd653d Dec 16 09:40:01.033994 systemd-networkd[1391]: lxcce470a9af226: Gained carrier Dec 16 09:40:01.047470 systemd-networkd[1391]: lxcc73a2f70ef3c: Link UP Dec 16 09:40:01.053597 kernel: eth0: renamed from tmp86aa8 Dec 16 09:40:01.065886 systemd-networkd[1391]: lxcc73a2f70ef3c: Gained carrier Dec 16 09:40:01.886726 systemd-networkd[1391]: lxc_health: Gained IPv6LL Dec 16 09:40:02.270808 systemd-networkd[1391]: lxcc73a2f70ef3c: Gained IPv6LL Dec 16 09:40:02.463303 systemd-networkd[1391]: lxcce470a9af226: Gained IPv6LL Dec 16 09:40:04.252654 containerd[1489]: time="2024-12-16T09:40:04.251135655Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 16 09:40:04.254594 containerd[1489]: time="2024-12-16T09:40:04.253368112Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 16 09:40:04.254594 containerd[1489]: time="2024-12-16T09:40:04.253492074Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 16 09:40:04.254594 containerd[1489]: time="2024-12-16T09:40:04.253596830Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 16 09:40:04.292002 systemd[1]: Started cri-containerd-d653d600f50f28b128a39bc88511788c7abdcc43adfc713c1ff8e5b97a8c85b6.scope - libcontainer container d653d600f50f28b128a39bc88511788c7abdcc43adfc713c1ff8e5b97a8c85b6. Dec 16 09:40:04.317707 containerd[1489]: time="2024-12-16T09:40:04.317634113Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 16 09:40:04.318715 containerd[1489]: time="2024-12-16T09:40:04.318594169Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 16 09:40:04.318715 containerd[1489]: time="2024-12-16T09:40:04.318609247Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 16 09:40:04.318715 containerd[1489]: time="2024-12-16T09:40:04.318669911Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 16 09:40:04.351717 systemd[1]: Started cri-containerd-86aa8b37f65b65eaecb05d3fc12a17333a8beeffddfd861a212ac2e864522ac6.scope - libcontainer container 86aa8b37f65b65eaecb05d3fc12a17333a8beeffddfd861a212ac2e864522ac6. Dec 16 09:40:04.382611 containerd[1489]: time="2024-12-16T09:40:04.381439563Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-srs4l,Uid:0058c53d-6810-460b-be34-d644c721a145,Namespace:kube-system,Attempt:0,} returns sandbox id \"d653d600f50f28b128a39bc88511788c7abdcc43adfc713c1ff8e5b97a8c85b6\"" Dec 16 09:40:04.386871 containerd[1489]: time="2024-12-16T09:40:04.386765135Z" level=info msg="CreateContainer within sandbox \"d653d600f50f28b128a39bc88511788c7abdcc43adfc713c1ff8e5b97a8c85b6\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Dec 16 09:40:04.405670 containerd[1489]: time="2024-12-16T09:40:04.405630473Z" level=info msg="CreateContainer within sandbox \"d653d600f50f28b128a39bc88511788c7abdcc43adfc713c1ff8e5b97a8c85b6\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"70f019d94b71a368f607b9036fcd2c4c8667c4ca5bff9d9b1ae39937f5dffb30\"" Dec 16 09:40:04.407671 containerd[1489]: time="2024-12-16T09:40:04.406854973Z" level=info msg="StartContainer for \"70f019d94b71a368f607b9036fcd2c4c8667c4ca5bff9d9b1ae39937f5dffb30\"" Dec 16 09:40:04.427144 containerd[1489]: time="2024-12-16T09:40:04.427105752Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-m2f8q,Uid:fae5c553-44f3-4fb1-a3ab-917443101ce4,Namespace:kube-system,Attempt:0,} returns sandbox id \"86aa8b37f65b65eaecb05d3fc12a17333a8beeffddfd861a212ac2e864522ac6\"" Dec 16 09:40:04.431095 containerd[1489]: time="2024-12-16T09:40:04.431056263Z" level=info msg="CreateContainer within sandbox \"86aa8b37f65b65eaecb05d3fc12a17333a8beeffddfd861a212ac2e864522ac6\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Dec 16 09:40:04.455242 containerd[1489]: time="2024-12-16T09:40:04.455147736Z" level=info msg="CreateContainer within sandbox \"86aa8b37f65b65eaecb05d3fc12a17333a8beeffddfd861a212ac2e864522ac6\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"dbbc72a3a1b12831a3539984fae2944b4d63b89f51cfbd2e33571e2629de325f\"" Dec 16 09:40:04.457982 containerd[1489]: time="2024-12-16T09:40:04.457860170Z" level=info msg="StartContainer for \"dbbc72a3a1b12831a3539984fae2944b4d63b89f51cfbd2e33571e2629de325f\"" Dec 16 09:40:04.459705 systemd[1]: Started cri-containerd-70f019d94b71a368f607b9036fcd2c4c8667c4ca5bff9d9b1ae39937f5dffb30.scope - libcontainer container 70f019d94b71a368f607b9036fcd2c4c8667c4ca5bff9d9b1ae39937f5dffb30. Dec 16 09:40:04.509058 containerd[1489]: time="2024-12-16T09:40:04.508109282Z" level=info msg="StartContainer for \"70f019d94b71a368f607b9036fcd2c4c8667c4ca5bff9d9b1ae39937f5dffb30\" returns successfully" Dec 16 09:40:04.509687 systemd[1]: Started cri-containerd-dbbc72a3a1b12831a3539984fae2944b4d63b89f51cfbd2e33571e2629de325f.scope - libcontainer container dbbc72a3a1b12831a3539984fae2944b4d63b89f51cfbd2e33571e2629de325f. Dec 16 09:40:04.543613 containerd[1489]: time="2024-12-16T09:40:04.543137433Z" level=info msg="StartContainer for \"dbbc72a3a1b12831a3539984fae2944b4d63b89f51cfbd2e33571e2629de325f\" returns successfully" Dec 16 09:40:04.671797 kubelet[2719]: I1216 09:40:04.671723 2719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-6f6b679f8f-m2f8q" podStartSLOduration=21.671705544 podStartE2EDuration="21.671705544s" podCreationTimestamp="2024-12-16 09:39:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-12-16 09:40:04.667798986 +0000 UTC m=+29.239322358" watchObservedRunningTime="2024-12-16 09:40:04.671705544 +0000 UTC m=+29.243228916" Dec 16 09:40:05.259598 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2336202734.mount: Deactivated successfully. Dec 16 09:40:05.669031 kubelet[2719]: I1216 09:40:05.668689 2719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-6f6b679f8f-srs4l" podStartSLOduration=22.668528666 podStartE2EDuration="22.668528666s" podCreationTimestamp="2024-12-16 09:39:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-12-16 09:40:04.685622415 +0000 UTC m=+29.257145787" watchObservedRunningTime="2024-12-16 09:40:05.668528666 +0000 UTC m=+30.240052048" Dec 16 09:43:26.599445 update_engine[1481]: I20241216 09:43:26.599368 1481 prefs.cc:52] certificate-report-to-send-update not present in /var/lib/update_engine/prefs Dec 16 09:43:26.599445 update_engine[1481]: I20241216 09:43:26.599429 1481 prefs.cc:52] certificate-report-to-send-download not present in /var/lib/update_engine/prefs Dec 16 09:43:26.601985 update_engine[1481]: I20241216 09:43:26.599700 1481 prefs.cc:52] aleph-version not present in /var/lib/update_engine/prefs Dec 16 09:43:26.601985 update_engine[1481]: I20241216 09:43:26.600443 1481 omaha_request_params.cc:62] Current group set to stable Dec 16 09:43:26.601985 update_engine[1481]: I20241216 09:43:26.601904 1481 update_attempter.cc:499] Already updated boot flags. Skipping. Dec 16 09:43:26.601985 update_engine[1481]: I20241216 09:43:26.601920 1481 update_attempter.cc:643] Scheduling an action processor start. Dec 16 09:43:26.601985 update_engine[1481]: I20241216 09:43:26.601943 1481 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction Dec 16 09:43:26.601985 update_engine[1481]: I20241216 09:43:26.601981 1481 prefs.cc:52] previous-version not present in /var/lib/update_engine/prefs Dec 16 09:43:26.602134 update_engine[1481]: I20241216 09:43:26.602051 1481 omaha_request_action.cc:271] Posting an Omaha request to disabled Dec 16 09:43:26.602134 update_engine[1481]: I20241216 09:43:26.602066 1481 omaha_request_action.cc:272] Request: Dec 16 09:43:26.602134 update_engine[1481]: Dec 16 09:43:26.602134 update_engine[1481]: Dec 16 09:43:26.602134 update_engine[1481]: Dec 16 09:43:26.602134 update_engine[1481]: Dec 16 09:43:26.602134 update_engine[1481]: Dec 16 09:43:26.602134 update_engine[1481]: Dec 16 09:43:26.602134 update_engine[1481]: Dec 16 09:43:26.602134 update_engine[1481]: Dec 16 09:43:26.602134 update_engine[1481]: I20241216 09:43:26.602077 1481 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Dec 16 09:43:26.610492 update_engine[1481]: I20241216 09:43:26.610420 1481 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Dec 16 09:43:26.610557 locksmithd[1513]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_CHECKING_FOR_UPDATE" NewVersion=0.0.0 NewSize=0 Dec 16 09:43:26.610814 update_engine[1481]: I20241216 09:43:26.610733 1481 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Dec 16 09:43:26.611743 update_engine[1481]: E20241216 09:43:26.611709 1481 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Dec 16 09:43:26.611794 update_engine[1481]: I20241216 09:43:26.611773 1481 libcurl_http_fetcher.cc:283] No HTTP response, retry 1 Dec 16 09:43:36.487488 update_engine[1481]: I20241216 09:43:36.487406 1481 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Dec 16 09:43:36.487916 update_engine[1481]: I20241216 09:43:36.487702 1481 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Dec 16 09:43:36.487945 update_engine[1481]: I20241216 09:43:36.487912 1481 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Dec 16 09:43:36.488599 update_engine[1481]: E20241216 09:43:36.488571 1481 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Dec 16 09:43:36.488659 update_engine[1481]: I20241216 09:43:36.488615 1481 libcurl_http_fetcher.cc:283] No HTTP response, retry 2 Dec 16 09:43:46.486298 update_engine[1481]: I20241216 09:43:46.486199 1481 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Dec 16 09:43:46.486801 update_engine[1481]: I20241216 09:43:46.486490 1481 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Dec 16 09:43:46.486801 update_engine[1481]: I20241216 09:43:46.486742 1481 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Dec 16 09:43:46.487368 update_engine[1481]: E20241216 09:43:46.487330 1481 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Dec 16 09:43:46.487419 update_engine[1481]: I20241216 09:43:46.487398 1481 libcurl_http_fetcher.cc:283] No HTTP response, retry 3 Dec 16 09:43:56.491423 update_engine[1481]: I20241216 09:43:56.491328 1481 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Dec 16 09:43:56.491894 update_engine[1481]: I20241216 09:43:56.491700 1481 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Dec 16 09:43:56.491973 update_engine[1481]: I20241216 09:43:56.491925 1481 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Dec 16 09:43:56.492605 update_engine[1481]: E20241216 09:43:56.492540 1481 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Dec 16 09:43:56.492665 update_engine[1481]: I20241216 09:43:56.492633 1481 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded Dec 16 09:43:56.492665 update_engine[1481]: I20241216 09:43:56.492655 1481 omaha_request_action.cc:617] Omaha request response: Dec 16 09:43:56.492772 update_engine[1481]: E20241216 09:43:56.492743 1481 omaha_request_action.cc:636] Omaha request network transfer failed. Dec 16 09:43:56.492816 update_engine[1481]: I20241216 09:43:56.492769 1481 action_processor.cc:68] ActionProcessor::ActionComplete: OmahaRequestAction action failed. Aborting processing. Dec 16 09:43:56.492816 update_engine[1481]: I20241216 09:43:56.492777 1481 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Dec 16 09:43:56.492816 update_engine[1481]: I20241216 09:43:56.492784 1481 update_attempter.cc:306] Processing Done. Dec 16 09:43:56.492816 update_engine[1481]: E20241216 09:43:56.492803 1481 update_attempter.cc:619] Update failed. Dec 16 09:43:56.492816 update_engine[1481]: I20241216 09:43:56.492810 1481 utils.cc:600] Converting error code 2000 to kActionCodeOmahaErrorInHTTPResponse Dec 16 09:43:56.492924 update_engine[1481]: I20241216 09:43:56.492818 1481 payload_state.cc:97] Updating payload state for error code: 37 (kActionCodeOmahaErrorInHTTPResponse) Dec 16 09:43:56.492924 update_engine[1481]: I20241216 09:43:56.492825 1481 payload_state.cc:103] Ignoring failures until we get a valid Omaha response. Dec 16 09:43:56.492965 update_engine[1481]: I20241216 09:43:56.492914 1481 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction Dec 16 09:43:56.492965 update_engine[1481]: I20241216 09:43:56.492949 1481 omaha_request_action.cc:271] Posting an Omaha request to disabled Dec 16 09:43:56.493016 update_engine[1481]: I20241216 09:43:56.492964 1481 omaha_request_action.cc:272] Request: Dec 16 09:43:56.493016 update_engine[1481]: Dec 16 09:43:56.493016 update_engine[1481]: Dec 16 09:43:56.493016 update_engine[1481]: Dec 16 09:43:56.493016 update_engine[1481]: Dec 16 09:43:56.493016 update_engine[1481]: Dec 16 09:43:56.493016 update_engine[1481]: Dec 16 09:43:56.493016 update_engine[1481]: I20241216 09:43:56.492974 1481 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Dec 16 09:43:56.493162 update_engine[1481]: I20241216 09:43:56.493117 1481 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Dec 16 09:43:56.493361 update_engine[1481]: I20241216 09:43:56.493257 1481 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Dec 16 09:43:56.493607 locksmithd[1513]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_REPORTING_ERROR_EVENT" NewVersion=0.0.0 NewSize=0 Dec 16 09:43:56.494090 update_engine[1481]: E20241216 09:43:56.494057 1481 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Dec 16 09:43:56.494127 update_engine[1481]: I20241216 09:43:56.494102 1481 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded Dec 16 09:43:56.494127 update_engine[1481]: I20241216 09:43:56.494112 1481 omaha_request_action.cc:617] Omaha request response: Dec 16 09:43:56.494127 update_engine[1481]: I20241216 09:43:56.494122 1481 action_processor.cc:65] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Dec 16 09:43:56.494198 update_engine[1481]: I20241216 09:43:56.494128 1481 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Dec 16 09:43:56.494198 update_engine[1481]: I20241216 09:43:56.494135 1481 update_attempter.cc:306] Processing Done. Dec 16 09:43:56.494198 update_engine[1481]: I20241216 09:43:56.494143 1481 update_attempter.cc:310] Error event sent. Dec 16 09:43:56.494198 update_engine[1481]: I20241216 09:43:56.494153 1481 update_check_scheduler.cc:74] Next update check in 40m10s Dec 16 09:43:56.494423 locksmithd[1513]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_IDLE" NewVersion=0.0.0 NewSize=0 Dec 16 09:44:23.025776 systemd[1]: Started sshd@7-138.199.145.148:22-147.75.109.163:38660.service - OpenSSH per-connection server daemon (147.75.109.163:38660). Dec 16 09:44:24.015017 sshd[4124]: Accepted publickey for core from 147.75.109.163 port 38660 ssh2: RSA SHA256:zB/zPQRxUCFkkFdvDftk99JQqA6bP3NHPa7FnaDUxKk Dec 16 09:44:24.018826 sshd[4124]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 09:44:24.029717 systemd-logind[1477]: New session 8 of user core. Dec 16 09:44:24.033770 systemd[1]: Started session-8.scope - Session 8 of User core. Dec 16 09:44:25.108090 sshd[4124]: pam_unix(sshd:session): session closed for user core Dec 16 09:44:25.112078 systemd-logind[1477]: Session 8 logged out. Waiting for processes to exit. Dec 16 09:44:25.112653 systemd[1]: sshd@7-138.199.145.148:22-147.75.109.163:38660.service: Deactivated successfully. Dec 16 09:44:25.116205 systemd[1]: session-8.scope: Deactivated successfully. Dec 16 09:44:25.117794 systemd-logind[1477]: Removed session 8. Dec 16 09:44:30.277101 systemd[1]: Started sshd@8-138.199.145.148:22-147.75.109.163:54054.service - OpenSSH per-connection server daemon (147.75.109.163:54054). Dec 16 09:44:31.261779 sshd[4137]: Accepted publickey for core from 147.75.109.163 port 54054 ssh2: RSA SHA256:zB/zPQRxUCFkkFdvDftk99JQqA6bP3NHPa7FnaDUxKk Dec 16 09:44:31.263905 sshd[4137]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 09:44:31.269634 systemd-logind[1477]: New session 9 of user core. Dec 16 09:44:31.277757 systemd[1]: Started session-9.scope - Session 9 of User core. Dec 16 09:44:32.008102 sshd[4137]: pam_unix(sshd:session): session closed for user core Dec 16 09:44:32.011789 systemd-logind[1477]: Session 9 logged out. Waiting for processes to exit. Dec 16 09:44:32.012490 systemd[1]: sshd@8-138.199.145.148:22-147.75.109.163:54054.service: Deactivated successfully. Dec 16 09:44:32.014350 systemd[1]: session-9.scope: Deactivated successfully. Dec 16 09:44:32.015486 systemd-logind[1477]: Removed session 9. Dec 16 09:44:37.178935 systemd[1]: Started sshd@9-138.199.145.148:22-147.75.109.163:56032.service - OpenSSH per-connection server daemon (147.75.109.163:56032). Dec 16 09:44:38.178222 sshd[4152]: Accepted publickey for core from 147.75.109.163 port 56032 ssh2: RSA SHA256:zB/zPQRxUCFkkFdvDftk99JQqA6bP3NHPa7FnaDUxKk Dec 16 09:44:38.180198 sshd[4152]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 09:44:38.185269 systemd-logind[1477]: New session 10 of user core. Dec 16 09:44:38.194724 systemd[1]: Started session-10.scope - Session 10 of User core. Dec 16 09:44:38.952958 sshd[4152]: pam_unix(sshd:session): session closed for user core Dec 16 09:44:38.958879 systemd[1]: sshd@9-138.199.145.148:22-147.75.109.163:56032.service: Deactivated successfully. Dec 16 09:44:38.962917 systemd[1]: session-10.scope: Deactivated successfully. Dec 16 09:44:38.964359 systemd-logind[1477]: Session 10 logged out. Waiting for processes to exit. Dec 16 09:44:38.966701 systemd-logind[1477]: Removed session 10. Dec 16 09:44:39.127088 systemd[1]: Started sshd@10-138.199.145.148:22-147.75.109.163:56036.service - OpenSSH per-connection server daemon (147.75.109.163:56036). Dec 16 09:44:40.115009 sshd[4166]: Accepted publickey for core from 147.75.109.163 port 56036 ssh2: RSA SHA256:zB/zPQRxUCFkkFdvDftk99JQqA6bP3NHPa7FnaDUxKk Dec 16 09:44:40.117791 sshd[4166]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 09:44:40.125685 systemd-logind[1477]: New session 11 of user core. Dec 16 09:44:40.132857 systemd[1]: Started session-11.scope - Session 11 of User core. Dec 16 09:44:40.961342 sshd[4166]: pam_unix(sshd:session): session closed for user core Dec 16 09:44:40.967847 systemd[1]: sshd@10-138.199.145.148:22-147.75.109.163:56036.service: Deactivated successfully. Dec 16 09:44:40.973917 systemd[1]: session-11.scope: Deactivated successfully. Dec 16 09:44:40.976684 systemd-logind[1477]: Session 11 logged out. Waiting for processes to exit. Dec 16 09:44:40.979269 systemd-logind[1477]: Removed session 11. Dec 16 09:44:41.135960 systemd[1]: Started sshd@11-138.199.145.148:22-147.75.109.163:56038.service - OpenSSH per-connection server daemon (147.75.109.163:56038). Dec 16 09:44:42.120421 sshd[4177]: Accepted publickey for core from 147.75.109.163 port 56038 ssh2: RSA SHA256:zB/zPQRxUCFkkFdvDftk99JQqA6bP3NHPa7FnaDUxKk Dec 16 09:44:42.122285 sshd[4177]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 09:44:42.127759 systemd-logind[1477]: New session 12 of user core. Dec 16 09:44:42.133768 systemd[1]: Started session-12.scope - Session 12 of User core. Dec 16 09:44:42.855599 sshd[4177]: pam_unix(sshd:session): session closed for user core Dec 16 09:44:42.860260 systemd-logind[1477]: Session 12 logged out. Waiting for processes to exit. Dec 16 09:44:42.860900 systemd[1]: sshd@11-138.199.145.148:22-147.75.109.163:56038.service: Deactivated successfully. Dec 16 09:44:42.863172 systemd[1]: session-12.scope: Deactivated successfully. Dec 16 09:44:42.864835 systemd-logind[1477]: Removed session 12. Dec 16 09:44:48.028616 systemd[1]: Started sshd@12-138.199.145.148:22-147.75.109.163:43560.service - OpenSSH per-connection server daemon (147.75.109.163:43560). Dec 16 09:44:49.021635 sshd[4192]: Accepted publickey for core from 147.75.109.163 port 43560 ssh2: RSA SHA256:zB/zPQRxUCFkkFdvDftk99JQqA6bP3NHPa7FnaDUxKk Dec 16 09:44:49.023323 sshd[4192]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 09:44:49.027802 systemd-logind[1477]: New session 13 of user core. Dec 16 09:44:49.040741 systemd[1]: Started session-13.scope - Session 13 of User core. Dec 16 09:44:49.759504 sshd[4192]: pam_unix(sshd:session): session closed for user core Dec 16 09:44:49.763401 systemd-logind[1477]: Session 13 logged out. Waiting for processes to exit. Dec 16 09:44:49.764398 systemd[1]: sshd@12-138.199.145.148:22-147.75.109.163:43560.service: Deactivated successfully. Dec 16 09:44:49.766931 systemd[1]: session-13.scope: Deactivated successfully. Dec 16 09:44:49.768096 systemd-logind[1477]: Removed session 13. Dec 16 09:44:49.930883 systemd[1]: Started sshd@13-138.199.145.148:22-147.75.109.163:43562.service - OpenSSH per-connection server daemon (147.75.109.163:43562). Dec 16 09:44:50.909261 sshd[4205]: Accepted publickey for core from 147.75.109.163 port 43562 ssh2: RSA SHA256:zB/zPQRxUCFkkFdvDftk99JQqA6bP3NHPa7FnaDUxKk Dec 16 09:44:50.911133 sshd[4205]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 09:44:50.915770 systemd-logind[1477]: New session 14 of user core. Dec 16 09:44:50.924709 systemd[1]: Started session-14.scope - Session 14 of User core. Dec 16 09:44:51.856433 sshd[4205]: pam_unix(sshd:session): session closed for user core Dec 16 09:44:51.865277 systemd[1]: sshd@13-138.199.145.148:22-147.75.109.163:43562.service: Deactivated successfully. Dec 16 09:44:51.869320 systemd[1]: session-14.scope: Deactivated successfully. Dec 16 09:44:51.870496 systemd-logind[1477]: Session 14 logged out. Waiting for processes to exit. Dec 16 09:44:51.873140 systemd-logind[1477]: Removed session 14. Dec 16 09:44:52.027896 systemd[1]: Started sshd@14-138.199.145.148:22-147.75.109.163:43570.service - OpenSSH per-connection server daemon (147.75.109.163:43570). Dec 16 09:44:53.021355 sshd[4216]: Accepted publickey for core from 147.75.109.163 port 43570 ssh2: RSA SHA256:zB/zPQRxUCFkkFdvDftk99JQqA6bP3NHPa7FnaDUxKk Dec 16 09:44:53.022933 sshd[4216]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 09:44:53.027823 systemd-logind[1477]: New session 15 of user core. Dec 16 09:44:53.039695 systemd[1]: Started session-15.scope - Session 15 of User core. Dec 16 09:44:55.160381 sshd[4216]: pam_unix(sshd:session): session closed for user core Dec 16 09:44:55.171966 systemd[1]: sshd@14-138.199.145.148:22-147.75.109.163:43570.service: Deactivated successfully. Dec 16 09:44:55.175985 systemd[1]: session-15.scope: Deactivated successfully. Dec 16 09:44:55.177619 systemd-logind[1477]: Session 15 logged out. Waiting for processes to exit. Dec 16 09:44:55.179471 systemd-logind[1477]: Removed session 15. Dec 16 09:44:55.334155 systemd[1]: Started sshd@15-138.199.145.148:22-147.75.109.163:43582.service - OpenSSH per-connection server daemon (147.75.109.163:43582). Dec 16 09:44:56.327437 sshd[4234]: Accepted publickey for core from 147.75.109.163 port 43582 ssh2: RSA SHA256:zB/zPQRxUCFkkFdvDftk99JQqA6bP3NHPa7FnaDUxKk Dec 16 09:44:56.330196 sshd[4234]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 09:44:56.338092 systemd-logind[1477]: New session 16 of user core. Dec 16 09:44:56.347832 systemd[1]: Started session-16.scope - Session 16 of User core. Dec 16 09:44:57.262437 sshd[4234]: pam_unix(sshd:session): session closed for user core Dec 16 09:44:57.266700 systemd[1]: sshd@15-138.199.145.148:22-147.75.109.163:43582.service: Deactivated successfully. Dec 16 09:44:57.269340 systemd[1]: session-16.scope: Deactivated successfully. Dec 16 09:44:57.270078 systemd-logind[1477]: Session 16 logged out. Waiting for processes to exit. Dec 16 09:44:57.271483 systemd-logind[1477]: Removed session 16. Dec 16 09:44:57.434812 systemd[1]: Started sshd@16-138.199.145.148:22-147.75.109.163:56718.service - OpenSSH per-connection server daemon (147.75.109.163:56718). Dec 16 09:44:58.407176 sshd[4245]: Accepted publickey for core from 147.75.109.163 port 56718 ssh2: RSA SHA256:zB/zPQRxUCFkkFdvDftk99JQqA6bP3NHPa7FnaDUxKk Dec 16 09:44:58.409120 sshd[4245]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 09:44:58.416469 systemd-logind[1477]: New session 17 of user core. Dec 16 09:44:58.422784 systemd[1]: Started session-17.scope - Session 17 of User core. Dec 16 09:44:59.176350 sshd[4245]: pam_unix(sshd:session): session closed for user core Dec 16 09:44:59.181821 systemd[1]: sshd@16-138.199.145.148:22-147.75.109.163:56718.service: Deactivated successfully. Dec 16 09:44:59.182288 systemd-logind[1477]: Session 17 logged out. Waiting for processes to exit. Dec 16 09:44:59.185491 systemd[1]: session-17.scope: Deactivated successfully. Dec 16 09:44:59.187203 systemd-logind[1477]: Removed session 17. Dec 16 09:45:04.351836 systemd[1]: Started sshd@17-138.199.145.148:22-147.75.109.163:56724.service - OpenSSH per-connection server daemon (147.75.109.163:56724). Dec 16 09:45:05.326154 sshd[4262]: Accepted publickey for core from 147.75.109.163 port 56724 ssh2: RSA SHA256:zB/zPQRxUCFkkFdvDftk99JQqA6bP3NHPa7FnaDUxKk Dec 16 09:45:05.327691 sshd[4262]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 09:45:05.332126 systemd-logind[1477]: New session 18 of user core. Dec 16 09:45:05.337727 systemd[1]: Started session-18.scope - Session 18 of User core. Dec 16 09:45:06.057771 sshd[4262]: pam_unix(sshd:session): session closed for user core Dec 16 09:45:06.062729 systemd-logind[1477]: Session 18 logged out. Waiting for processes to exit. Dec 16 09:45:06.063639 systemd[1]: sshd@17-138.199.145.148:22-147.75.109.163:56724.service: Deactivated successfully. Dec 16 09:45:06.066543 systemd[1]: session-18.scope: Deactivated successfully. Dec 16 09:45:06.068043 systemd-logind[1477]: Removed session 18. Dec 16 09:45:11.232921 systemd[1]: Started sshd@18-138.199.145.148:22-147.75.109.163:54128.service - OpenSSH per-connection server daemon (147.75.109.163:54128). Dec 16 09:45:12.214393 sshd[4276]: Accepted publickey for core from 147.75.109.163 port 54128 ssh2: RSA SHA256:zB/zPQRxUCFkkFdvDftk99JQqA6bP3NHPa7FnaDUxKk Dec 16 09:45:12.216330 sshd[4276]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 09:45:12.221679 systemd-logind[1477]: New session 19 of user core. Dec 16 09:45:12.226737 systemd[1]: Started session-19.scope - Session 19 of User core. Dec 16 09:45:12.948105 sshd[4276]: pam_unix(sshd:session): session closed for user core Dec 16 09:45:12.952354 systemd[1]: sshd@18-138.199.145.148:22-147.75.109.163:54128.service: Deactivated successfully. Dec 16 09:45:12.954702 systemd[1]: session-19.scope: Deactivated successfully. Dec 16 09:45:12.955508 systemd-logind[1477]: Session 19 logged out. Waiting for processes to exit. Dec 16 09:45:12.956958 systemd-logind[1477]: Removed session 19. Dec 16 09:45:13.119874 systemd[1]: Started sshd@19-138.199.145.148:22-147.75.109.163:54144.service - OpenSSH per-connection server daemon (147.75.109.163:54144). Dec 16 09:45:14.094519 sshd[4289]: Accepted publickey for core from 147.75.109.163 port 54144 ssh2: RSA SHA256:zB/zPQRxUCFkkFdvDftk99JQqA6bP3NHPa7FnaDUxKk Dec 16 09:45:14.096689 sshd[4289]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 09:45:14.101397 systemd-logind[1477]: New session 20 of user core. Dec 16 09:45:14.104718 systemd[1]: Started session-20.scope - Session 20 of User core. Dec 16 09:45:15.976927 systemd[1]: run-containerd-runc-k8s.io-2afccada9da9b8083d9189601319ae6737a931161aae85b051f870382555e281-runc.MK8lSe.mount: Deactivated successfully. Dec 16 09:45:16.006886 containerd[1489]: time="2024-12-16T09:45:16.006809871Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Dec 16 09:45:16.044967 containerd[1489]: time="2024-12-16T09:45:16.044923445Z" level=info msg="StopContainer for \"2afccada9da9b8083d9189601319ae6737a931161aae85b051f870382555e281\" with timeout 2 (s)" Dec 16 09:45:16.045142 containerd[1489]: time="2024-12-16T09:45:16.044935929Z" level=info msg="StopContainer for \"cbea43b6dc94dbdf299c057b62cb3bf562efd7c51c2518de0491b7378ffa0f61\" with timeout 30 (s)" Dec 16 09:45:16.047495 containerd[1489]: time="2024-12-16T09:45:16.047455278Z" level=info msg="Stop container \"2afccada9da9b8083d9189601319ae6737a931161aae85b051f870382555e281\" with signal terminated" Dec 16 09:45:16.047715 containerd[1489]: time="2024-12-16T09:45:16.047460398Z" level=info msg="Stop container \"cbea43b6dc94dbdf299c057b62cb3bf562efd7c51c2518de0491b7378ffa0f61\" with signal terminated" Dec 16 09:45:16.056220 systemd-networkd[1391]: lxc_health: Link DOWN Dec 16 09:45:16.056228 systemd-networkd[1391]: lxc_health: Lost carrier Dec 16 09:45:16.069459 systemd[1]: cri-containerd-cbea43b6dc94dbdf299c057b62cb3bf562efd7c51c2518de0491b7378ffa0f61.scope: Deactivated successfully. Dec 16 09:45:16.086913 systemd[1]: cri-containerd-2afccada9da9b8083d9189601319ae6737a931161aae85b051f870382555e281.scope: Deactivated successfully. Dec 16 09:45:16.087620 systemd[1]: cri-containerd-2afccada9da9b8083d9189601319ae6737a931161aae85b051f870382555e281.scope: Consumed 7.330s CPU time. Dec 16 09:45:16.104449 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-cbea43b6dc94dbdf299c057b62cb3bf562efd7c51c2518de0491b7378ffa0f61-rootfs.mount: Deactivated successfully. Dec 16 09:45:16.109210 containerd[1489]: time="2024-12-16T09:45:16.109149367Z" level=info msg="shim disconnected" id=cbea43b6dc94dbdf299c057b62cb3bf562efd7c51c2518de0491b7378ffa0f61 namespace=k8s.io Dec 16 09:45:16.109376 containerd[1489]: time="2024-12-16T09:45:16.109359027Z" level=warning msg="cleaning up after shim disconnected" id=cbea43b6dc94dbdf299c057b62cb3bf562efd7c51c2518de0491b7378ffa0f61 namespace=k8s.io Dec 16 09:45:16.109548 containerd[1489]: time="2024-12-16T09:45:16.109531550Z" level=info msg="cleaning up dead shim" namespace=k8s.io Dec 16 09:45:16.117239 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-2afccada9da9b8083d9189601319ae6737a931161aae85b051f870382555e281-rootfs.mount: Deactivated successfully. Dec 16 09:45:16.123235 containerd[1489]: time="2024-12-16T09:45:16.123112861Z" level=info msg="shim disconnected" id=2afccada9da9b8083d9189601319ae6737a931161aae85b051f870382555e281 namespace=k8s.io Dec 16 09:45:16.123235 containerd[1489]: time="2024-12-16T09:45:16.123178213Z" level=warning msg="cleaning up after shim disconnected" id=2afccada9da9b8083d9189601319ae6737a931161aae85b051f870382555e281 namespace=k8s.io Dec 16 09:45:16.123235 containerd[1489]: time="2024-12-16T09:45:16.123190806Z" level=info msg="cleaning up dead shim" namespace=k8s.io Dec 16 09:45:16.127174 containerd[1489]: time="2024-12-16T09:45:16.127075659Z" level=info msg="StopContainer for \"cbea43b6dc94dbdf299c057b62cb3bf562efd7c51c2518de0491b7378ffa0f61\" returns successfully" Dec 16 09:45:16.134278 containerd[1489]: time="2024-12-16T09:45:16.134216640Z" level=info msg="StopPodSandbox for \"3b87f8f1710867438768b3941d4b273a4a8bc234af9b1b40cb15b9c0ba524cc0\"" Dec 16 09:45:16.134278 containerd[1489]: time="2024-12-16T09:45:16.134260300Z" level=info msg="Container to stop \"cbea43b6dc94dbdf299c057b62cb3bf562efd7c51c2518de0491b7378ffa0f61\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Dec 16 09:45:16.136583 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-3b87f8f1710867438768b3941d4b273a4a8bc234af9b1b40cb15b9c0ba524cc0-shm.mount: Deactivated successfully. Dec 16 09:45:16.142361 systemd[1]: cri-containerd-3b87f8f1710867438768b3941d4b273a4a8bc234af9b1b40cb15b9c0ba524cc0.scope: Deactivated successfully. Dec 16 09:45:16.143728 containerd[1489]: time="2024-12-16T09:45:16.143698717Z" level=info msg="StopContainer for \"2afccada9da9b8083d9189601319ae6737a931161aae85b051f870382555e281\" returns successfully" Dec 16 09:45:16.144103 containerd[1489]: time="2024-12-16T09:45:16.144084277Z" level=info msg="StopPodSandbox for \"cabc0575ef95bbb1469e40d05f6827c256e2aaae5cb89b9ef5cb1790b3ba4225\"" Dec 16 09:45:16.144187 containerd[1489]: time="2024-12-16T09:45:16.144171731Z" level=info msg="Container to stop \"7c5bcccdd565e4286227ae13392a9a1674ec0d8be5e1af592e4a3bcb00b300f2\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Dec 16 09:45:16.144265 containerd[1489]: time="2024-12-16T09:45:16.144251670Z" level=info msg="Container to stop \"2e8b7cafc7c2004533f17577ad79101e0aa4ebfadd6e3c7f4d6ed1ce5b3e7246\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Dec 16 09:45:16.144348 containerd[1489]: time="2024-12-16T09:45:16.144332942Z" level=info msg="Container to stop \"2afccada9da9b8083d9189601319ae6737a931161aae85b051f870382555e281\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Dec 16 09:45:16.144449 containerd[1489]: time="2024-12-16T09:45:16.144435013Z" level=info msg="Container to stop \"95cea8638f01368f96be90942a0735ec3cc96c43a2e920f16ee973799bff7f2b\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Dec 16 09:45:16.144517 containerd[1489]: time="2024-12-16T09:45:16.144485067Z" level=info msg="Container to stop \"76b958fd76ab693634764099cd1289981aa90f18dd8bffa90a5c04169b1c3928\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Dec 16 09:45:16.152914 systemd[1]: cri-containerd-cabc0575ef95bbb1469e40d05f6827c256e2aaae5cb89b9ef5cb1790b3ba4225.scope: Deactivated successfully. Dec 16 09:45:16.170052 containerd[1489]: time="2024-12-16T09:45:16.169994266Z" level=info msg="shim disconnected" id=3b87f8f1710867438768b3941d4b273a4a8bc234af9b1b40cb15b9c0ba524cc0 namespace=k8s.io Dec 16 09:45:16.170052 containerd[1489]: time="2024-12-16T09:45:16.170042967Z" level=warning msg="cleaning up after shim disconnected" id=3b87f8f1710867438768b3941d4b273a4a8bc234af9b1b40cb15b9c0ba524cc0 namespace=k8s.io Dec 16 09:45:16.170052 containerd[1489]: time="2024-12-16T09:45:16.170051713Z" level=info msg="cleaning up dead shim" namespace=k8s.io Dec 16 09:45:16.185854 containerd[1489]: time="2024-12-16T09:45:16.185808530Z" level=warning msg="cleanup warnings time=\"2024-12-16T09:45:16Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Dec 16 09:45:16.190474 containerd[1489]: time="2024-12-16T09:45:16.190415414Z" level=info msg="shim disconnected" id=cabc0575ef95bbb1469e40d05f6827c256e2aaae5cb89b9ef5cb1790b3ba4225 namespace=k8s.io Dec 16 09:45:16.191632 containerd[1489]: time="2024-12-16T09:45:16.190716717Z" level=warning msg="cleaning up after shim disconnected" id=cabc0575ef95bbb1469e40d05f6827c256e2aaae5cb89b9ef5cb1790b3ba4225 namespace=k8s.io Dec 16 09:45:16.191632 containerd[1489]: time="2024-12-16T09:45:16.190738527Z" level=info msg="cleaning up dead shim" namespace=k8s.io Dec 16 09:45:16.196172 containerd[1489]: time="2024-12-16T09:45:16.196138112Z" level=info msg="TearDown network for sandbox \"3b87f8f1710867438768b3941d4b273a4a8bc234af9b1b40cb15b9c0ba524cc0\" successfully" Dec 16 09:45:16.196172 containerd[1489]: time="2024-12-16T09:45:16.196165864Z" level=info msg="StopPodSandbox for \"3b87f8f1710867438768b3941d4b273a4a8bc234af9b1b40cb15b9c0ba524cc0\" returns successfully" Dec 16 09:45:16.212668 containerd[1489]: time="2024-12-16T09:45:16.212609084Z" level=info msg="TearDown network for sandbox \"cabc0575ef95bbb1469e40d05f6827c256e2aaae5cb89b9ef5cb1790b3ba4225\" successfully" Dec 16 09:45:16.212668 containerd[1489]: time="2024-12-16T09:45:16.212656684Z" level=info msg="StopPodSandbox for \"cabc0575ef95bbb1469e40d05f6827c256e2aaae5cb89b9ef5cb1790b3ba4225\" returns successfully" Dec 16 09:45:16.253080 kubelet[2719]: I1216 09:45:16.253027 2719 scope.go:117] "RemoveContainer" containerID="cbea43b6dc94dbdf299c057b62cb3bf562efd7c51c2518de0491b7378ffa0f61" Dec 16 09:45:16.256168 containerd[1489]: time="2024-12-16T09:45:16.256085985Z" level=info msg="RemoveContainer for \"cbea43b6dc94dbdf299c057b62cb3bf562efd7c51c2518de0491b7378ffa0f61\"" Dec 16 09:45:16.259777 containerd[1489]: time="2024-12-16T09:45:16.259733735Z" level=info msg="RemoveContainer for \"cbea43b6dc94dbdf299c057b62cb3bf562efd7c51c2518de0491b7378ffa0f61\" returns successfully" Dec 16 09:45:16.259981 kubelet[2719]: I1216 09:45:16.259935 2719 scope.go:117] "RemoveContainer" containerID="cbea43b6dc94dbdf299c057b62cb3bf562efd7c51c2518de0491b7378ffa0f61" Dec 16 09:45:16.267667 containerd[1489]: time="2024-12-16T09:45:16.261638766Z" level=error msg="ContainerStatus for \"cbea43b6dc94dbdf299c057b62cb3bf562efd7c51c2518de0491b7378ffa0f61\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"cbea43b6dc94dbdf299c057b62cb3bf562efd7c51c2518de0491b7378ffa0f61\": not found" Dec 16 09:45:16.276059 kubelet[2719]: E1216 09:45:16.275859 2719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"cbea43b6dc94dbdf299c057b62cb3bf562efd7c51c2518de0491b7378ffa0f61\": not found" containerID="cbea43b6dc94dbdf299c057b62cb3bf562efd7c51c2518de0491b7378ffa0f61" Dec 16 09:45:16.276059 kubelet[2719]: I1216 09:45:16.275907 2719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"cbea43b6dc94dbdf299c057b62cb3bf562efd7c51c2518de0491b7378ffa0f61"} err="failed to get container status \"cbea43b6dc94dbdf299c057b62cb3bf562efd7c51c2518de0491b7378ffa0f61\": rpc error: code = NotFound desc = an error occurred when try to find container \"cbea43b6dc94dbdf299c057b62cb3bf562efd7c51c2518de0491b7378ffa0f61\": not found" Dec 16 09:45:16.276059 kubelet[2719]: I1216 09:45:16.275985 2719 scope.go:117] "RemoveContainer" containerID="2afccada9da9b8083d9189601319ae6737a931161aae85b051f870382555e281" Dec 16 09:45:16.276960 containerd[1489]: time="2024-12-16T09:45:16.276931976Z" level=info msg="RemoveContainer for \"2afccada9da9b8083d9189601319ae6737a931161aae85b051f870382555e281\"" Dec 16 09:45:16.279785 containerd[1489]: time="2024-12-16T09:45:16.279752800Z" level=info msg="RemoveContainer for \"2afccada9da9b8083d9189601319ae6737a931161aae85b051f870382555e281\" returns successfully" Dec 16 09:45:16.280038 kubelet[2719]: I1216 09:45:16.279986 2719 scope.go:117] "RemoveContainer" containerID="2e8b7cafc7c2004533f17577ad79101e0aa4ebfadd6e3c7f4d6ed1ce5b3e7246" Dec 16 09:45:16.281045 containerd[1489]: time="2024-12-16T09:45:16.281024999Z" level=info msg="RemoveContainer for \"2e8b7cafc7c2004533f17577ad79101e0aa4ebfadd6e3c7f4d6ed1ce5b3e7246\"" Dec 16 09:45:16.283829 containerd[1489]: time="2024-12-16T09:45:16.283796199Z" level=info msg="RemoveContainer for \"2e8b7cafc7c2004533f17577ad79101e0aa4ebfadd6e3c7f4d6ed1ce5b3e7246\" returns successfully" Dec 16 09:45:16.283991 kubelet[2719]: I1216 09:45:16.283939 2719 scope.go:117] "RemoveContainer" containerID="7c5bcccdd565e4286227ae13392a9a1674ec0d8be5e1af592e4a3bcb00b300f2" Dec 16 09:45:16.284985 containerd[1489]: time="2024-12-16T09:45:16.284921553Z" level=info msg="RemoveContainer for \"7c5bcccdd565e4286227ae13392a9a1674ec0d8be5e1af592e4a3bcb00b300f2\"" Dec 16 09:45:16.287698 containerd[1489]: time="2024-12-16T09:45:16.287639615Z" level=info msg="RemoveContainer for \"7c5bcccdd565e4286227ae13392a9a1674ec0d8be5e1af592e4a3bcb00b300f2\" returns successfully" Dec 16 09:45:16.287886 kubelet[2719]: I1216 09:45:16.287797 2719 scope.go:117] "RemoveContainer" containerID="76b958fd76ab693634764099cd1289981aa90f18dd8bffa90a5c04169b1c3928" Dec 16 09:45:16.288987 containerd[1489]: time="2024-12-16T09:45:16.288782592Z" level=info msg="RemoveContainer for \"76b958fd76ab693634764099cd1289981aa90f18dd8bffa90a5c04169b1c3928\"" Dec 16 09:45:16.291373 containerd[1489]: time="2024-12-16T09:45:16.291324724Z" level=info msg="RemoveContainer for \"76b958fd76ab693634764099cd1289981aa90f18dd8bffa90a5c04169b1c3928\" returns successfully" Dec 16 09:45:16.291471 kubelet[2719]: I1216 09:45:16.291444 2719 scope.go:117] "RemoveContainer" containerID="95cea8638f01368f96be90942a0735ec3cc96c43a2e920f16ee973799bff7f2b" Dec 16 09:45:16.292317 containerd[1489]: time="2024-12-16T09:45:16.292287875Z" level=info msg="RemoveContainer for \"95cea8638f01368f96be90942a0735ec3cc96c43a2e920f16ee973799bff7f2b\"" Dec 16 09:45:16.294535 containerd[1489]: time="2024-12-16T09:45:16.294478179Z" level=info msg="RemoveContainer for \"95cea8638f01368f96be90942a0735ec3cc96c43a2e920f16ee973799bff7f2b\" returns successfully" Dec 16 09:45:16.294693 kubelet[2719]: I1216 09:45:16.294677 2719 scope.go:117] "RemoveContainer" containerID="2afccada9da9b8083d9189601319ae6737a931161aae85b051f870382555e281" Dec 16 09:45:16.294936 containerd[1489]: time="2024-12-16T09:45:16.294841369Z" level=error msg="ContainerStatus for \"2afccada9da9b8083d9189601319ae6737a931161aae85b051f870382555e281\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"2afccada9da9b8083d9189601319ae6737a931161aae85b051f870382555e281\": not found" Dec 16 09:45:16.295053 kubelet[2719]: E1216 09:45:16.294984 2719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"2afccada9da9b8083d9189601319ae6737a931161aae85b051f870382555e281\": not found" containerID="2afccada9da9b8083d9189601319ae6737a931161aae85b051f870382555e281" Dec 16 09:45:16.295053 kubelet[2719]: I1216 09:45:16.295035 2719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"2afccada9da9b8083d9189601319ae6737a931161aae85b051f870382555e281"} err="failed to get container status \"2afccada9da9b8083d9189601319ae6737a931161aae85b051f870382555e281\": rpc error: code = NotFound desc = an error occurred when try to find container \"2afccada9da9b8083d9189601319ae6737a931161aae85b051f870382555e281\": not found" Dec 16 09:45:16.295053 kubelet[2719]: I1216 09:45:16.295054 2719 scope.go:117] "RemoveContainer" containerID="2e8b7cafc7c2004533f17577ad79101e0aa4ebfadd6e3c7f4d6ed1ce5b3e7246" Dec 16 09:45:16.295370 containerd[1489]: time="2024-12-16T09:45:16.295319642Z" level=error msg="ContainerStatus for \"2e8b7cafc7c2004533f17577ad79101e0aa4ebfadd6e3c7f4d6ed1ce5b3e7246\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"2e8b7cafc7c2004533f17577ad79101e0aa4ebfadd6e3c7f4d6ed1ce5b3e7246\": not found" Dec 16 09:45:16.295430 kubelet[2719]: E1216 09:45:16.295417 2719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"2e8b7cafc7c2004533f17577ad79101e0aa4ebfadd6e3c7f4d6ed1ce5b3e7246\": not found" containerID="2e8b7cafc7c2004533f17577ad79101e0aa4ebfadd6e3c7f4d6ed1ce5b3e7246" Dec 16 09:45:16.295472 kubelet[2719]: I1216 09:45:16.295435 2719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"2e8b7cafc7c2004533f17577ad79101e0aa4ebfadd6e3c7f4d6ed1ce5b3e7246"} err="failed to get container status \"2e8b7cafc7c2004533f17577ad79101e0aa4ebfadd6e3c7f4d6ed1ce5b3e7246\": rpc error: code = NotFound desc = an error occurred when try to find container \"2e8b7cafc7c2004533f17577ad79101e0aa4ebfadd6e3c7f4d6ed1ce5b3e7246\": not found" Dec 16 09:45:16.295472 kubelet[2719]: I1216 09:45:16.295451 2719 scope.go:117] "RemoveContainer" containerID="7c5bcccdd565e4286227ae13392a9a1674ec0d8be5e1af592e4a3bcb00b300f2" Dec 16 09:45:16.295840 containerd[1489]: time="2024-12-16T09:45:16.295641835Z" level=error msg="ContainerStatus for \"7c5bcccdd565e4286227ae13392a9a1674ec0d8be5e1af592e4a3bcb00b300f2\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"7c5bcccdd565e4286227ae13392a9a1674ec0d8be5e1af592e4a3bcb00b300f2\": not found" Dec 16 09:45:16.295967 kubelet[2719]: E1216 09:45:16.295938 2719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"7c5bcccdd565e4286227ae13392a9a1674ec0d8be5e1af592e4a3bcb00b300f2\": not found" containerID="7c5bcccdd565e4286227ae13392a9a1674ec0d8be5e1af592e4a3bcb00b300f2" Dec 16 09:45:16.296024 kubelet[2719]: I1216 09:45:16.295971 2719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"7c5bcccdd565e4286227ae13392a9a1674ec0d8be5e1af592e4a3bcb00b300f2"} err="failed to get container status \"7c5bcccdd565e4286227ae13392a9a1674ec0d8be5e1af592e4a3bcb00b300f2\": rpc error: code = NotFound desc = an error occurred when try to find container \"7c5bcccdd565e4286227ae13392a9a1674ec0d8be5e1af592e4a3bcb00b300f2\": not found" Dec 16 09:45:16.296024 kubelet[2719]: I1216 09:45:16.295992 2719 scope.go:117] "RemoveContainer" containerID="76b958fd76ab693634764099cd1289981aa90f18dd8bffa90a5c04169b1c3928" Dec 16 09:45:16.296262 containerd[1489]: time="2024-12-16T09:45:16.296209477Z" level=error msg="ContainerStatus for \"76b958fd76ab693634764099cd1289981aa90f18dd8bffa90a5c04169b1c3928\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"76b958fd76ab693634764099cd1289981aa90f18dd8bffa90a5c04169b1c3928\": not found" Dec 16 09:45:16.296355 kubelet[2719]: E1216 09:45:16.296327 2719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"76b958fd76ab693634764099cd1289981aa90f18dd8bffa90a5c04169b1c3928\": not found" containerID="76b958fd76ab693634764099cd1289981aa90f18dd8bffa90a5c04169b1c3928" Dec 16 09:45:16.296409 kubelet[2719]: I1216 09:45:16.296357 2719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"76b958fd76ab693634764099cd1289981aa90f18dd8bffa90a5c04169b1c3928"} err="failed to get container status \"76b958fd76ab693634764099cd1289981aa90f18dd8bffa90a5c04169b1c3928\": rpc error: code = NotFound desc = an error occurred when try to find container \"76b958fd76ab693634764099cd1289981aa90f18dd8bffa90a5c04169b1c3928\": not found" Dec 16 09:45:16.296409 kubelet[2719]: I1216 09:45:16.296378 2719 scope.go:117] "RemoveContainer" containerID="95cea8638f01368f96be90942a0735ec3cc96c43a2e920f16ee973799bff7f2b" Dec 16 09:45:16.296666 containerd[1489]: time="2024-12-16T09:45:16.296619242Z" level=error msg="ContainerStatus for \"95cea8638f01368f96be90942a0735ec3cc96c43a2e920f16ee973799bff7f2b\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"95cea8638f01368f96be90942a0735ec3cc96c43a2e920f16ee973799bff7f2b\": not found" Dec 16 09:45:16.296839 kubelet[2719]: E1216 09:45:16.296770 2719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"95cea8638f01368f96be90942a0735ec3cc96c43a2e920f16ee973799bff7f2b\": not found" containerID="95cea8638f01368f96be90942a0735ec3cc96c43a2e920f16ee973799bff7f2b" Dec 16 09:45:16.296839 kubelet[2719]: I1216 09:45:16.296814 2719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"95cea8638f01368f96be90942a0735ec3cc96c43a2e920f16ee973799bff7f2b"} err="failed to get container status \"95cea8638f01368f96be90942a0735ec3cc96c43a2e920f16ee973799bff7f2b\": rpc error: code = NotFound desc = an error occurred when try to find container \"95cea8638f01368f96be90942a0735ec3cc96c43a2e920f16ee973799bff7f2b\": not found" Dec 16 09:45:16.329518 kubelet[2719]: I1216 09:45:16.328559 2719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-lib-modules\") pod \"81563a72-3610-425a-9c84-4ec0b9f36bc6\" (UID: \"81563a72-3610-425a-9c84-4ec0b9f36bc6\") " Dec 16 09:45:16.329518 kubelet[2719]: I1216 09:45:16.329230 2719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/81563a72-3610-425a-9c84-4ec0b9f36bc6-cilium-config-path\") pod \"81563a72-3610-425a-9c84-4ec0b9f36bc6\" (UID: \"81563a72-3610-425a-9c84-4ec0b9f36bc6\") " Dec 16 09:45:16.329518 kubelet[2719]: I1216 09:45:16.329260 2719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-cilium-cgroup\") pod \"81563a72-3610-425a-9c84-4ec0b9f36bc6\" (UID: \"81563a72-3610-425a-9c84-4ec0b9f36bc6\") " Dec 16 09:45:16.329518 kubelet[2719]: I1216 09:45:16.329283 2719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-xtables-lock\") pod \"81563a72-3610-425a-9c84-4ec0b9f36bc6\" (UID: \"81563a72-3610-425a-9c84-4ec0b9f36bc6\") " Dec 16 09:45:16.329518 kubelet[2719]: I1216 09:45:16.329307 2719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmcrw\" (UniqueName: \"kubernetes.io/projected/67d9cab9-f389-4c1c-a6e6-319040d73a93-kube-api-access-jmcrw\") pod \"67d9cab9-f389-4c1c-a6e6-319040d73a93\" (UID: \"67d9cab9-f389-4c1c-a6e6-319040d73a93\") " Dec 16 09:45:16.329518 kubelet[2719]: I1216 09:45:16.329331 2719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/81563a72-3610-425a-9c84-4ec0b9f36bc6-hubble-tls\") pod \"81563a72-3610-425a-9c84-4ec0b9f36bc6\" (UID: \"81563a72-3610-425a-9c84-4ec0b9f36bc6\") " Dec 16 09:45:16.329760 kubelet[2719]: I1216 09:45:16.329352 2719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-bpf-maps\") pod \"81563a72-3610-425a-9c84-4ec0b9f36bc6\" (UID: \"81563a72-3610-425a-9c84-4ec0b9f36bc6\") " Dec 16 09:45:16.329760 kubelet[2719]: I1216 09:45:16.329375 2719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-cni-path\") pod \"81563a72-3610-425a-9c84-4ec0b9f36bc6\" (UID: \"81563a72-3610-425a-9c84-4ec0b9f36bc6\") " Dec 16 09:45:16.329760 kubelet[2719]: I1216 09:45:16.329403 2719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/67d9cab9-f389-4c1c-a6e6-319040d73a93-cilium-config-path\") pod \"67d9cab9-f389-4c1c-a6e6-319040d73a93\" (UID: \"67d9cab9-f389-4c1c-a6e6-319040d73a93\") " Dec 16 09:45:16.329760 kubelet[2719]: I1216 09:45:16.329426 2719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-hostproc\") pod \"81563a72-3610-425a-9c84-4ec0b9f36bc6\" (UID: \"81563a72-3610-425a-9c84-4ec0b9f36bc6\") " Dec 16 09:45:16.329760 kubelet[2719]: I1216 09:45:16.329451 2719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-host-proc-sys-net\") pod \"81563a72-3610-425a-9c84-4ec0b9f36bc6\" (UID: \"81563a72-3610-425a-9c84-4ec0b9f36bc6\") " Dec 16 09:45:16.329760 kubelet[2719]: I1216 09:45:16.329474 2719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-48vfc\" (UniqueName: \"kubernetes.io/projected/81563a72-3610-425a-9c84-4ec0b9f36bc6-kube-api-access-48vfc\") pod \"81563a72-3610-425a-9c84-4ec0b9f36bc6\" (UID: \"81563a72-3610-425a-9c84-4ec0b9f36bc6\") " Dec 16 09:45:16.329911 kubelet[2719]: I1216 09:45:16.329494 2719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-cilium-run\") pod \"81563a72-3610-425a-9c84-4ec0b9f36bc6\" (UID: \"81563a72-3610-425a-9c84-4ec0b9f36bc6\") " Dec 16 09:45:16.329911 kubelet[2719]: I1216 09:45:16.329512 2719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-etc-cni-netd\") pod \"81563a72-3610-425a-9c84-4ec0b9f36bc6\" (UID: \"81563a72-3610-425a-9c84-4ec0b9f36bc6\") " Dec 16 09:45:16.329911 kubelet[2719]: I1216 09:45:16.329535 2719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-host-proc-sys-kernel\") pod \"81563a72-3610-425a-9c84-4ec0b9f36bc6\" (UID: \"81563a72-3610-425a-9c84-4ec0b9f36bc6\") " Dec 16 09:45:16.329911 kubelet[2719]: I1216 09:45:16.329603 2719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/81563a72-3610-425a-9c84-4ec0b9f36bc6-clustermesh-secrets\") pod \"81563a72-3610-425a-9c84-4ec0b9f36bc6\" (UID: \"81563a72-3610-425a-9c84-4ec0b9f36bc6\") " Dec 16 09:45:16.339580 kubelet[2719]: I1216 09:45:16.336742 2719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81563a72-3610-425a-9c84-4ec0b9f36bc6-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "81563a72-3610-425a-9c84-4ec0b9f36bc6" (UID: "81563a72-3610-425a-9c84-4ec0b9f36bc6"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 16 09:45:16.339580 kubelet[2719]: I1216 09:45:16.337004 2719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "81563a72-3610-425a-9c84-4ec0b9f36bc6" (UID: "81563a72-3610-425a-9c84-4ec0b9f36bc6"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 16 09:45:16.339580 kubelet[2719]: I1216 09:45:16.337873 2719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-cni-path" (OuterVolumeSpecName: "cni-path") pod "81563a72-3610-425a-9c84-4ec0b9f36bc6" (UID: "81563a72-3610-425a-9c84-4ec0b9f36bc6"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 16 09:45:16.340882 kubelet[2719]: I1216 09:45:16.340862 2719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67d9cab9-f389-4c1c-a6e6-319040d73a93-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "67d9cab9-f389-4c1c-a6e6-319040d73a93" (UID: "67d9cab9-f389-4c1c-a6e6-319040d73a93"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 16 09:45:16.340974 kubelet[2719]: I1216 09:45:16.340958 2719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-hostproc" (OuterVolumeSpecName: "hostproc") pod "81563a72-3610-425a-9c84-4ec0b9f36bc6" (UID: "81563a72-3610-425a-9c84-4ec0b9f36bc6"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 16 09:45:16.341039 kubelet[2719]: I1216 09:45:16.341027 2719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "81563a72-3610-425a-9c84-4ec0b9f36bc6" (UID: "81563a72-3610-425a-9c84-4ec0b9f36bc6"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 16 09:45:16.342158 kubelet[2719]: I1216 09:45:16.342100 2719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81563a72-3610-425a-9c84-4ec0b9f36bc6-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "81563a72-3610-425a-9c84-4ec0b9f36bc6" (UID: "81563a72-3610-425a-9c84-4ec0b9f36bc6"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 16 09:45:16.342158 kubelet[2719]: I1216 09:45:16.342142 2719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "81563a72-3610-425a-9c84-4ec0b9f36bc6" (UID: "81563a72-3610-425a-9c84-4ec0b9f36bc6"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 16 09:45:16.342229 kubelet[2719]: I1216 09:45:16.342166 2719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "81563a72-3610-425a-9c84-4ec0b9f36bc6" (UID: "81563a72-3610-425a-9c84-4ec0b9f36bc6"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 16 09:45:16.343664 kubelet[2719]: I1216 09:45:16.343545 2719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81563a72-3610-425a-9c84-4ec0b9f36bc6-kube-api-access-48vfc" (OuterVolumeSpecName: "kube-api-access-48vfc") pod "81563a72-3610-425a-9c84-4ec0b9f36bc6" (UID: "81563a72-3610-425a-9c84-4ec0b9f36bc6"). InnerVolumeSpecName "kube-api-access-48vfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 16 09:45:16.343769 kubelet[2719]: I1216 09:45:16.343753 2719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "81563a72-3610-425a-9c84-4ec0b9f36bc6" (UID: "81563a72-3610-425a-9c84-4ec0b9f36bc6"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 16 09:45:16.343845 kubelet[2719]: I1216 09:45:16.343831 2719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "81563a72-3610-425a-9c84-4ec0b9f36bc6" (UID: "81563a72-3610-425a-9c84-4ec0b9f36bc6"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 16 09:45:16.343902 kubelet[2719]: I1216 09:45:16.343890 2719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "81563a72-3610-425a-9c84-4ec0b9f36bc6" (UID: "81563a72-3610-425a-9c84-4ec0b9f36bc6"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 16 09:45:16.345392 kubelet[2719]: I1216 09:45:16.345356 2719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67d9cab9-f389-4c1c-a6e6-319040d73a93-kube-api-access-jmcrw" (OuterVolumeSpecName: "kube-api-access-jmcrw") pod "67d9cab9-f389-4c1c-a6e6-319040d73a93" (UID: "67d9cab9-f389-4c1c-a6e6-319040d73a93"). InnerVolumeSpecName "kube-api-access-jmcrw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 16 09:45:16.345440 kubelet[2719]: I1216 09:45:16.345407 2719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "81563a72-3610-425a-9c84-4ec0b9f36bc6" (UID: "81563a72-3610-425a-9c84-4ec0b9f36bc6"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 16 09:45:16.346300 kubelet[2719]: I1216 09:45:16.346272 2719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81563a72-3610-425a-9c84-4ec0b9f36bc6-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "81563a72-3610-425a-9c84-4ec0b9f36bc6" (UID: "81563a72-3610-425a-9c84-4ec0b9f36bc6"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 16 09:45:16.431340 kubelet[2719]: I1216 09:45:16.431297 2719 reconciler_common.go:288] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-lib-modules\") on node \"ci-4081-2-1-7-459a6b42ef\" DevicePath \"\"" Dec 16 09:45:16.431340 kubelet[2719]: I1216 09:45:16.431325 2719 reconciler_common.go:288] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/81563a72-3610-425a-9c84-4ec0b9f36bc6-cilium-config-path\") on node \"ci-4081-2-1-7-459a6b42ef\" DevicePath \"\"" Dec 16 09:45:16.431340 kubelet[2719]: I1216 09:45:16.431335 2719 reconciler_common.go:288] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-cilium-cgroup\") on node \"ci-4081-2-1-7-459a6b42ef\" DevicePath \"\"" Dec 16 09:45:16.431340 kubelet[2719]: I1216 09:45:16.431343 2719 reconciler_common.go:288] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-xtables-lock\") on node \"ci-4081-2-1-7-459a6b42ef\" DevicePath \"\"" Dec 16 09:45:16.431340 kubelet[2719]: I1216 09:45:16.431352 2719 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-jmcrw\" (UniqueName: \"kubernetes.io/projected/67d9cab9-f389-4c1c-a6e6-319040d73a93-kube-api-access-jmcrw\") on node \"ci-4081-2-1-7-459a6b42ef\" DevicePath \"\"" Dec 16 09:45:16.431555 kubelet[2719]: I1216 09:45:16.431361 2719 reconciler_common.go:288] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/81563a72-3610-425a-9c84-4ec0b9f36bc6-hubble-tls\") on node \"ci-4081-2-1-7-459a6b42ef\" DevicePath \"\"" Dec 16 09:45:16.431555 kubelet[2719]: I1216 09:45:16.431370 2719 reconciler_common.go:288] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-bpf-maps\") on node \"ci-4081-2-1-7-459a6b42ef\" DevicePath \"\"" Dec 16 09:45:16.431555 kubelet[2719]: I1216 09:45:16.431377 2719 reconciler_common.go:288] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-cni-path\") on node \"ci-4081-2-1-7-459a6b42ef\" DevicePath \"\"" Dec 16 09:45:16.431555 kubelet[2719]: I1216 09:45:16.431384 2719 reconciler_common.go:288] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/67d9cab9-f389-4c1c-a6e6-319040d73a93-cilium-config-path\") on node \"ci-4081-2-1-7-459a6b42ef\" DevicePath \"\"" Dec 16 09:45:16.431555 kubelet[2719]: I1216 09:45:16.431391 2719 reconciler_common.go:288] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-hostproc\") on node \"ci-4081-2-1-7-459a6b42ef\" DevicePath \"\"" Dec 16 09:45:16.431555 kubelet[2719]: I1216 09:45:16.431400 2719 reconciler_common.go:288] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-host-proc-sys-net\") on node \"ci-4081-2-1-7-459a6b42ef\" DevicePath \"\"" Dec 16 09:45:16.431555 kubelet[2719]: I1216 09:45:16.431409 2719 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-48vfc\" (UniqueName: \"kubernetes.io/projected/81563a72-3610-425a-9c84-4ec0b9f36bc6-kube-api-access-48vfc\") on node \"ci-4081-2-1-7-459a6b42ef\" DevicePath \"\"" Dec 16 09:45:16.431555 kubelet[2719]: I1216 09:45:16.431416 2719 reconciler_common.go:288] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-cilium-run\") on node \"ci-4081-2-1-7-459a6b42ef\" DevicePath \"\"" Dec 16 09:45:16.431765 kubelet[2719]: I1216 09:45:16.431424 2719 reconciler_common.go:288] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-etc-cni-netd\") on node \"ci-4081-2-1-7-459a6b42ef\" DevicePath \"\"" Dec 16 09:45:16.431765 kubelet[2719]: I1216 09:45:16.431431 2719 reconciler_common.go:288] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/81563a72-3610-425a-9c84-4ec0b9f36bc6-host-proc-sys-kernel\") on node \"ci-4081-2-1-7-459a6b42ef\" DevicePath \"\"" Dec 16 09:45:16.431765 kubelet[2719]: I1216 09:45:16.431439 2719 reconciler_common.go:288] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/81563a72-3610-425a-9c84-4ec0b9f36bc6-clustermesh-secrets\") on node \"ci-4081-2-1-7-459a6b42ef\" DevicePath \"\"" Dec 16 09:45:16.546009 systemd[1]: Removed slice kubepods-besteffort-pod67d9cab9_f389_4c1c_a6e6_319040d73a93.slice - libcontainer container kubepods-besteffort-pod67d9cab9_f389_4c1c_a6e6_319040d73a93.slice. Dec 16 09:45:16.558148 systemd[1]: Removed slice kubepods-burstable-pod81563a72_3610_425a_9c84_4ec0b9f36bc6.slice - libcontainer container kubepods-burstable-pod81563a72_3610_425a_9c84_4ec0b9f36bc6.slice. Dec 16 09:45:16.558241 systemd[1]: kubepods-burstable-pod81563a72_3610_425a_9c84_4ec0b9f36bc6.slice: Consumed 7.411s CPU time. Dec 16 09:45:16.970606 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-3b87f8f1710867438768b3941d4b273a4a8bc234af9b1b40cb15b9c0ba524cc0-rootfs.mount: Deactivated successfully. Dec 16 09:45:16.971284 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-cabc0575ef95bbb1469e40d05f6827c256e2aaae5cb89b9ef5cb1790b3ba4225-rootfs.mount: Deactivated successfully. Dec 16 09:45:16.971380 systemd[1]: var-lib-kubelet-pods-67d9cab9\x2df389\x2d4c1c\x2da6e6\x2d319040d73a93-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2djmcrw.mount: Deactivated successfully. Dec 16 09:45:16.971473 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-cabc0575ef95bbb1469e40d05f6827c256e2aaae5cb89b9ef5cb1790b3ba4225-shm.mount: Deactivated successfully. Dec 16 09:45:16.971578 systemd[1]: var-lib-kubelet-pods-81563a72\x2d3610\x2d425a\x2d9c84\x2d4ec0b9f36bc6-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d48vfc.mount: Deactivated successfully. Dec 16 09:45:16.971673 systemd[1]: var-lib-kubelet-pods-81563a72\x2d3610\x2d425a\x2d9c84\x2d4ec0b9f36bc6-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Dec 16 09:45:16.971814 systemd[1]: var-lib-kubelet-pods-81563a72\x2d3610\x2d425a\x2d9c84\x2d4ec0b9f36bc6-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Dec 16 09:45:17.546359 kubelet[2719]: I1216 09:45:17.546312 2719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67d9cab9-f389-4c1c-a6e6-319040d73a93" path="/var/lib/kubelet/pods/67d9cab9-f389-4c1c-a6e6-319040d73a93/volumes" Dec 16 09:45:17.547054 kubelet[2719]: I1216 09:45:17.547020 2719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81563a72-3610-425a-9c84-4ec0b9f36bc6" path="/var/lib/kubelet/pods/81563a72-3610-425a-9c84-4ec0b9f36bc6/volumes" Dec 16 09:45:18.031354 sshd[4289]: pam_unix(sshd:session): session closed for user core Dec 16 09:45:18.037435 systemd[1]: sshd@19-138.199.145.148:22-147.75.109.163:54144.service: Deactivated successfully. Dec 16 09:45:18.041438 systemd[1]: session-20.scope: Deactivated successfully. Dec 16 09:45:18.045239 systemd-logind[1477]: Session 20 logged out. Waiting for processes to exit. Dec 16 09:45:18.047840 systemd-logind[1477]: Removed session 20. Dec 16 09:45:18.209009 systemd[1]: Started sshd@20-138.199.145.148:22-147.75.109.163:36100.service - OpenSSH per-connection server daemon (147.75.109.163:36100). Dec 16 09:45:19.206636 sshd[4455]: Accepted publickey for core from 147.75.109.163 port 36100 ssh2: RSA SHA256:zB/zPQRxUCFkkFdvDftk99JQqA6bP3NHPa7FnaDUxKk Dec 16 09:45:19.208340 sshd[4455]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 09:45:19.213703 systemd-logind[1477]: New session 21 of user core. Dec 16 09:45:19.219682 systemd[1]: Started session-21.scope - Session 21 of User core. Dec 16 09:45:20.048165 kubelet[2719]: E1216 09:45:20.047894 2719 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="81563a72-3610-425a-9c84-4ec0b9f36bc6" containerName="mount-cgroup" Dec 16 09:45:20.048165 kubelet[2719]: E1216 09:45:20.047920 2719 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="81563a72-3610-425a-9c84-4ec0b9f36bc6" containerName="apply-sysctl-overwrites" Dec 16 09:45:20.048165 kubelet[2719]: E1216 09:45:20.047926 2719 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="81563a72-3610-425a-9c84-4ec0b9f36bc6" containerName="mount-bpf-fs" Dec 16 09:45:20.048165 kubelet[2719]: E1216 09:45:20.047933 2719 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="81563a72-3610-425a-9c84-4ec0b9f36bc6" containerName="clean-cilium-state" Dec 16 09:45:20.048165 kubelet[2719]: E1216 09:45:20.047939 2719 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="67d9cab9-f389-4c1c-a6e6-319040d73a93" containerName="cilium-operator" Dec 16 09:45:20.048165 kubelet[2719]: E1216 09:45:20.047945 2719 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="81563a72-3610-425a-9c84-4ec0b9f36bc6" containerName="cilium-agent" Dec 16 09:45:20.056297 kubelet[2719]: I1216 09:45:20.056267 2719 memory_manager.go:354] "RemoveStaleState removing state" podUID="81563a72-3610-425a-9c84-4ec0b9f36bc6" containerName="cilium-agent" Dec 16 09:45:20.056754 kubelet[2719]: I1216 09:45:20.056437 2719 memory_manager.go:354] "RemoveStaleState removing state" podUID="67d9cab9-f389-4c1c-a6e6-319040d73a93" containerName="cilium-operator" Dec 16 09:45:20.137635 systemd[1]: Created slice kubepods-burstable-podba5c2efd_bac8_42a2_8a7e_c0f6548cc142.slice - libcontainer container kubepods-burstable-podba5c2efd_bac8_42a2_8a7e_c0f6548cc142.slice. Dec 16 09:45:20.155655 kubelet[2719]: I1216 09:45:20.155306 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/ba5c2efd-bac8-42a2-8a7e-c0f6548cc142-cilium-cgroup\") pod \"cilium-f622h\" (UID: \"ba5c2efd-bac8-42a2-8a7e-c0f6548cc142\") " pod="kube-system/cilium-f622h" Dec 16 09:45:20.155655 kubelet[2719]: I1216 09:45:20.155340 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/ba5c2efd-bac8-42a2-8a7e-c0f6548cc142-cilium-ipsec-secrets\") pod \"cilium-f622h\" (UID: \"ba5c2efd-bac8-42a2-8a7e-c0f6548cc142\") " pod="kube-system/cilium-f622h" Dec 16 09:45:20.155655 kubelet[2719]: I1216 09:45:20.155362 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/ba5c2efd-bac8-42a2-8a7e-c0f6548cc142-hostproc\") pod \"cilium-f622h\" (UID: \"ba5c2efd-bac8-42a2-8a7e-c0f6548cc142\") " pod="kube-system/cilium-f622h" Dec 16 09:45:20.155655 kubelet[2719]: I1216 09:45:20.155376 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ba5c2efd-bac8-42a2-8a7e-c0f6548cc142-lib-modules\") pod \"cilium-f622h\" (UID: \"ba5c2efd-bac8-42a2-8a7e-c0f6548cc142\") " pod="kube-system/cilium-f622h" Dec 16 09:45:20.155655 kubelet[2719]: I1216 09:45:20.155391 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ba5c2efd-bac8-42a2-8a7e-c0f6548cc142-etc-cni-netd\") pod \"cilium-f622h\" (UID: \"ba5c2efd-bac8-42a2-8a7e-c0f6548cc142\") " pod="kube-system/cilium-f622h" Dec 16 09:45:20.155655 kubelet[2719]: I1216 09:45:20.155404 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/ba5c2efd-bac8-42a2-8a7e-c0f6548cc142-cilium-config-path\") pod \"cilium-f622h\" (UID: \"ba5c2efd-bac8-42a2-8a7e-c0f6548cc142\") " pod="kube-system/cilium-f622h" Dec 16 09:45:20.155921 kubelet[2719]: I1216 09:45:20.155417 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/ba5c2efd-bac8-42a2-8a7e-c0f6548cc142-cni-path\") pod \"cilium-f622h\" (UID: \"ba5c2efd-bac8-42a2-8a7e-c0f6548cc142\") " pod="kube-system/cilium-f622h" Dec 16 09:45:20.155921 kubelet[2719]: I1216 09:45:20.155430 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/ba5c2efd-bac8-42a2-8a7e-c0f6548cc142-clustermesh-secrets\") pod \"cilium-f622h\" (UID: \"ba5c2efd-bac8-42a2-8a7e-c0f6548cc142\") " pod="kube-system/cilium-f622h" Dec 16 09:45:20.155921 kubelet[2719]: I1216 09:45:20.155442 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/ba5c2efd-bac8-42a2-8a7e-c0f6548cc142-host-proc-sys-net\") pod \"cilium-f622h\" (UID: \"ba5c2efd-bac8-42a2-8a7e-c0f6548cc142\") " pod="kube-system/cilium-f622h" Dec 16 09:45:20.155921 kubelet[2719]: I1216 09:45:20.155454 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/ba5c2efd-bac8-42a2-8a7e-c0f6548cc142-host-proc-sys-kernel\") pod \"cilium-f622h\" (UID: \"ba5c2efd-bac8-42a2-8a7e-c0f6548cc142\") " pod="kube-system/cilium-f622h" Dec 16 09:45:20.155921 kubelet[2719]: I1216 09:45:20.155466 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/ba5c2efd-bac8-42a2-8a7e-c0f6548cc142-hubble-tls\") pod \"cilium-f622h\" (UID: \"ba5c2efd-bac8-42a2-8a7e-c0f6548cc142\") " pod="kube-system/cilium-f622h" Dec 16 09:45:20.157170 kubelet[2719]: I1216 09:45:20.155478 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5xzm\" (UniqueName: \"kubernetes.io/projected/ba5c2efd-bac8-42a2-8a7e-c0f6548cc142-kube-api-access-v5xzm\") pod \"cilium-f622h\" (UID: \"ba5c2efd-bac8-42a2-8a7e-c0f6548cc142\") " pod="kube-system/cilium-f622h" Dec 16 09:45:20.157170 kubelet[2719]: I1216 09:45:20.155492 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/ba5c2efd-bac8-42a2-8a7e-c0f6548cc142-cilium-run\") pod \"cilium-f622h\" (UID: \"ba5c2efd-bac8-42a2-8a7e-c0f6548cc142\") " pod="kube-system/cilium-f622h" Dec 16 09:45:20.157170 kubelet[2719]: I1216 09:45:20.155505 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/ba5c2efd-bac8-42a2-8a7e-c0f6548cc142-bpf-maps\") pod \"cilium-f622h\" (UID: \"ba5c2efd-bac8-42a2-8a7e-c0f6548cc142\") " pod="kube-system/cilium-f622h" Dec 16 09:45:20.157170 kubelet[2719]: I1216 09:45:20.155522 2719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/ba5c2efd-bac8-42a2-8a7e-c0f6548cc142-xtables-lock\") pod \"cilium-f622h\" (UID: \"ba5c2efd-bac8-42a2-8a7e-c0f6548cc142\") " pod="kube-system/cilium-f622h" Dec 16 09:45:20.280804 sshd[4455]: pam_unix(sshd:session): session closed for user core Dec 16 09:45:20.285321 systemd[1]: sshd@20-138.199.145.148:22-147.75.109.163:36100.service: Deactivated successfully. Dec 16 09:45:20.287514 systemd[1]: session-21.scope: Deactivated successfully. Dec 16 09:45:20.289336 systemd-logind[1477]: Session 21 logged out. Waiting for processes to exit. Dec 16 09:45:20.290422 systemd-logind[1477]: Removed session 21. Dec 16 09:45:20.443304 containerd[1489]: time="2024-12-16T09:45:20.442650202Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-f622h,Uid:ba5c2efd-bac8-42a2-8a7e-c0f6548cc142,Namespace:kube-system,Attempt:0,}" Dec 16 09:45:20.452387 systemd[1]: Started sshd@21-138.199.145.148:22-147.75.109.163:36106.service - OpenSSH per-connection server daemon (147.75.109.163:36106). Dec 16 09:45:20.487247 containerd[1489]: time="2024-12-16T09:45:20.486586523Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 16 09:45:20.487247 containerd[1489]: time="2024-12-16T09:45:20.486637879Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 16 09:45:20.487247 containerd[1489]: time="2024-12-16T09:45:20.486668477Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 16 09:45:20.487247 containerd[1489]: time="2024-12-16T09:45:20.486945274Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 16 09:45:20.509697 systemd[1]: Started cri-containerd-8e02e7bf7f17bc46ae937fa24f9348934f112958af28e925e6189149055e5e83.scope - libcontainer container 8e02e7bf7f17bc46ae937fa24f9348934f112958af28e925e6189149055e5e83. Dec 16 09:45:20.534217 containerd[1489]: time="2024-12-16T09:45:20.534160065Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-f622h,Uid:ba5c2efd-bac8-42a2-8a7e-c0f6548cc142,Namespace:kube-system,Attempt:0,} returns sandbox id \"8e02e7bf7f17bc46ae937fa24f9348934f112958af28e925e6189149055e5e83\"" Dec 16 09:45:20.537816 containerd[1489]: time="2024-12-16T09:45:20.537779802Z" level=info msg="CreateContainer within sandbox \"8e02e7bf7f17bc46ae937fa24f9348934f112958af28e925e6189149055e5e83\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Dec 16 09:45:20.548162 containerd[1489]: time="2024-12-16T09:45:20.548119863Z" level=info msg="CreateContainer within sandbox \"8e02e7bf7f17bc46ae937fa24f9348934f112958af28e925e6189149055e5e83\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"a73b2a8a284a4ed67a3f5429ef0070d2978b4d8a9b1d6af5f918ae0e8f5c12b1\"" Dec 16 09:45:20.549060 containerd[1489]: time="2024-12-16T09:45:20.549027721Z" level=info msg="StartContainer for \"a73b2a8a284a4ed67a3f5429ef0070d2978b4d8a9b1d6af5f918ae0e8f5c12b1\"" Dec 16 09:45:20.572697 systemd[1]: Started cri-containerd-a73b2a8a284a4ed67a3f5429ef0070d2978b4d8a9b1d6af5f918ae0e8f5c12b1.scope - libcontainer container a73b2a8a284a4ed67a3f5429ef0070d2978b4d8a9b1d6af5f918ae0e8f5c12b1. Dec 16 09:45:20.595238 containerd[1489]: time="2024-12-16T09:45:20.595177600Z" level=info msg="StartContainer for \"a73b2a8a284a4ed67a3f5429ef0070d2978b4d8a9b1d6af5f918ae0e8f5c12b1\" returns successfully" Dec 16 09:45:20.609213 systemd[1]: cri-containerd-a73b2a8a284a4ed67a3f5429ef0070d2978b4d8a9b1d6af5f918ae0e8f5c12b1.scope: Deactivated successfully. Dec 16 09:45:20.641553 containerd[1489]: time="2024-12-16T09:45:20.641493380Z" level=info msg="shim disconnected" id=a73b2a8a284a4ed67a3f5429ef0070d2978b4d8a9b1d6af5f918ae0e8f5c12b1 namespace=k8s.io Dec 16 09:45:20.641553 containerd[1489]: time="2024-12-16T09:45:20.641541170Z" level=warning msg="cleaning up after shim disconnected" id=a73b2a8a284a4ed67a3f5429ef0070d2978b4d8a9b1d6af5f918ae0e8f5c12b1 namespace=k8s.io Dec 16 09:45:20.641553 containerd[1489]: time="2024-12-16T09:45:20.641550047Z" level=info msg="cleaning up dead shim" namespace=k8s.io Dec 16 09:45:20.686598 kubelet[2719]: E1216 09:45:20.686501 2719 kubelet.go:2901] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Dec 16 09:45:21.284461 containerd[1489]: time="2024-12-16T09:45:21.284378103Z" level=info msg="CreateContainer within sandbox \"8e02e7bf7f17bc46ae937fa24f9348934f112958af28e925e6189149055e5e83\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Dec 16 09:45:21.302157 containerd[1489]: time="2024-12-16T09:45:21.301429593Z" level=info msg="CreateContainer within sandbox \"8e02e7bf7f17bc46ae937fa24f9348934f112958af28e925e6189149055e5e83\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"20b8d7a02264d4d1f9b29094063e9322cf41c4ea4359589b55ed0d5adcf07028\"" Dec 16 09:45:21.302696 containerd[1489]: time="2024-12-16T09:45:21.302642891Z" level=info msg="StartContainer for \"20b8d7a02264d4d1f9b29094063e9322cf41c4ea4359589b55ed0d5adcf07028\"" Dec 16 09:45:21.341732 systemd[1]: Started cri-containerd-20b8d7a02264d4d1f9b29094063e9322cf41c4ea4359589b55ed0d5adcf07028.scope - libcontainer container 20b8d7a02264d4d1f9b29094063e9322cf41c4ea4359589b55ed0d5adcf07028. Dec 16 09:45:21.369077 containerd[1489]: time="2024-12-16T09:45:21.368963222Z" level=info msg="StartContainer for \"20b8d7a02264d4d1f9b29094063e9322cf41c4ea4359589b55ed0d5adcf07028\" returns successfully" Dec 16 09:45:21.376695 systemd[1]: cri-containerd-20b8d7a02264d4d1f9b29094063e9322cf41c4ea4359589b55ed0d5adcf07028.scope: Deactivated successfully. Dec 16 09:45:21.398152 containerd[1489]: time="2024-12-16T09:45:21.398081891Z" level=info msg="shim disconnected" id=20b8d7a02264d4d1f9b29094063e9322cf41c4ea4359589b55ed0d5adcf07028 namespace=k8s.io Dec 16 09:45:21.398152 containerd[1489]: time="2024-12-16T09:45:21.398150359Z" level=warning msg="cleaning up after shim disconnected" id=20b8d7a02264d4d1f9b29094063e9322cf41c4ea4359589b55ed0d5adcf07028 namespace=k8s.io Dec 16 09:45:21.398390 containerd[1489]: time="2024-12-16T09:45:21.398159887Z" level=info msg="cleaning up dead shim" namespace=k8s.io Dec 16 09:45:21.433640 sshd[4471]: Accepted publickey for core from 147.75.109.163 port 36106 ssh2: RSA SHA256:zB/zPQRxUCFkkFdvDftk99JQqA6bP3NHPa7FnaDUxKk Dec 16 09:45:21.435094 sshd[4471]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 09:45:21.439622 systemd-logind[1477]: New session 22 of user core. Dec 16 09:45:21.448703 systemd[1]: Started session-22.scope - Session 22 of User core. Dec 16 09:45:22.114515 sshd[4471]: pam_unix(sshd:session): session closed for user core Dec 16 09:45:22.118919 systemd-logind[1477]: Session 22 logged out. Waiting for processes to exit. Dec 16 09:45:22.119325 systemd[1]: sshd@21-138.199.145.148:22-147.75.109.163:36106.service: Deactivated successfully. Dec 16 09:45:22.121537 systemd[1]: session-22.scope: Deactivated successfully. Dec 16 09:45:22.124667 systemd-logind[1477]: Removed session 22. Dec 16 09:45:22.262475 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-20b8d7a02264d4d1f9b29094063e9322cf41c4ea4359589b55ed0d5adcf07028-rootfs.mount: Deactivated successfully. Dec 16 09:45:22.278075 systemd[1]: Started sshd@22-138.199.145.148:22-147.75.109.163:36122.service - OpenSSH per-connection server daemon (147.75.109.163:36122). Dec 16 09:45:22.292074 containerd[1489]: time="2024-12-16T09:45:22.291699764Z" level=info msg="CreateContainer within sandbox \"8e02e7bf7f17bc46ae937fa24f9348934f112958af28e925e6189149055e5e83\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Dec 16 09:45:22.315999 containerd[1489]: time="2024-12-16T09:45:22.315959310Z" level=info msg="CreateContainer within sandbox \"8e02e7bf7f17bc46ae937fa24f9348934f112958af28e925e6189149055e5e83\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"4ee13129c5f9ac53addc2491ac599fe8bbdee5e9ee80a9d5a4bb31860011d2c1\"" Dec 16 09:45:22.316676 containerd[1489]: time="2024-12-16T09:45:22.316488149Z" level=info msg="StartContainer for \"4ee13129c5f9ac53addc2491ac599fe8bbdee5e9ee80a9d5a4bb31860011d2c1\"" Dec 16 09:45:22.343685 systemd[1]: Started cri-containerd-4ee13129c5f9ac53addc2491ac599fe8bbdee5e9ee80a9d5a4bb31860011d2c1.scope - libcontainer container 4ee13129c5f9ac53addc2491ac599fe8bbdee5e9ee80a9d5a4bb31860011d2c1. Dec 16 09:45:22.369801 containerd[1489]: time="2024-12-16T09:45:22.369182455Z" level=info msg="StartContainer for \"4ee13129c5f9ac53addc2491ac599fe8bbdee5e9ee80a9d5a4bb31860011d2c1\" returns successfully" Dec 16 09:45:22.377937 systemd[1]: cri-containerd-4ee13129c5f9ac53addc2491ac599fe8bbdee5e9ee80a9d5a4bb31860011d2c1.scope: Deactivated successfully. Dec 16 09:45:22.404715 containerd[1489]: time="2024-12-16T09:45:22.404647640Z" level=info msg="shim disconnected" id=4ee13129c5f9ac53addc2491ac599fe8bbdee5e9ee80a9d5a4bb31860011d2c1 namespace=k8s.io Dec 16 09:45:22.404715 containerd[1489]: time="2024-12-16T09:45:22.404711489Z" level=warning msg="cleaning up after shim disconnected" id=4ee13129c5f9ac53addc2491ac599fe8bbdee5e9ee80a9d5a4bb31860011d2c1 namespace=k8s.io Dec 16 09:45:22.404972 containerd[1489]: time="2024-12-16T09:45:22.404723452Z" level=info msg="cleaning up dead shim" namespace=k8s.io Dec 16 09:45:23.249404 sshd[4643]: Accepted publickey for core from 147.75.109.163 port 36122 ssh2: RSA SHA256:zB/zPQRxUCFkkFdvDftk99JQqA6bP3NHPa7FnaDUxKk Dec 16 09:45:23.251144 sshd[4643]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 09:45:23.256200 systemd-logind[1477]: New session 23 of user core. Dec 16 09:45:23.262709 systemd[1]: Started session-23.scope - Session 23 of User core. Dec 16 09:45:23.265009 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-4ee13129c5f9ac53addc2491ac599fe8bbdee5e9ee80a9d5a4bb31860011d2c1-rootfs.mount: Deactivated successfully. Dec 16 09:45:23.296604 containerd[1489]: time="2024-12-16T09:45:23.296544818Z" level=info msg="CreateContainer within sandbox \"8e02e7bf7f17bc46ae937fa24f9348934f112958af28e925e6189149055e5e83\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Dec 16 09:45:23.300391 kubelet[2719]: I1216 09:45:23.300341 2719 setters.go:600] "Node became not ready" node="ci-4081-2-1-7-459a6b42ef" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2024-12-16T09:45:23Z","lastTransitionTime":"2024-12-16T09:45:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} Dec 16 09:45:23.315608 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount323038482.mount: Deactivated successfully. Dec 16 09:45:23.316835 containerd[1489]: time="2024-12-16T09:45:23.316773337Z" level=info msg="CreateContainer within sandbox \"8e02e7bf7f17bc46ae937fa24f9348934f112958af28e925e6189149055e5e83\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"a49ff47330b37e80c1166ed659d46f385980ec2e00d627fa6451d6525d3df443\"" Dec 16 09:45:23.317742 containerd[1489]: time="2024-12-16T09:45:23.317718365Z" level=info msg="StartContainer for \"a49ff47330b37e80c1166ed659d46f385980ec2e00d627fa6451d6525d3df443\"" Dec 16 09:45:23.358704 systemd[1]: Started cri-containerd-a49ff47330b37e80c1166ed659d46f385980ec2e00d627fa6451d6525d3df443.scope - libcontainer container a49ff47330b37e80c1166ed659d46f385980ec2e00d627fa6451d6525d3df443. Dec 16 09:45:23.386862 containerd[1489]: time="2024-12-16T09:45:23.386824354Z" level=info msg="StartContainer for \"a49ff47330b37e80c1166ed659d46f385980ec2e00d627fa6451d6525d3df443\" returns successfully" Dec 16 09:45:23.387074 systemd[1]: cri-containerd-a49ff47330b37e80c1166ed659d46f385980ec2e00d627fa6451d6525d3df443.scope: Deactivated successfully. Dec 16 09:45:23.411677 containerd[1489]: time="2024-12-16T09:45:23.411508965Z" level=info msg="shim disconnected" id=a49ff47330b37e80c1166ed659d46f385980ec2e00d627fa6451d6525d3df443 namespace=k8s.io Dec 16 09:45:23.411677 containerd[1489]: time="2024-12-16T09:45:23.411619020Z" level=warning msg="cleaning up after shim disconnected" id=a49ff47330b37e80c1166ed659d46f385980ec2e00d627fa6451d6525d3df443 namespace=k8s.io Dec 16 09:45:23.411677 containerd[1489]: time="2024-12-16T09:45:23.411633147Z" level=info msg="cleaning up dead shim" namespace=k8s.io Dec 16 09:45:24.265393 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a49ff47330b37e80c1166ed659d46f385980ec2e00d627fa6451d6525d3df443-rootfs.mount: Deactivated successfully. Dec 16 09:45:24.303593 containerd[1489]: time="2024-12-16T09:45:24.298742946Z" level=info msg="CreateContainer within sandbox \"8e02e7bf7f17bc46ae937fa24f9348934f112958af28e925e6189149055e5e83\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Dec 16 09:45:24.324556 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1814948706.mount: Deactivated successfully. Dec 16 09:45:24.325852 containerd[1489]: time="2024-12-16T09:45:24.325801745Z" level=info msg="CreateContainer within sandbox \"8e02e7bf7f17bc46ae937fa24f9348934f112958af28e925e6189149055e5e83\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"6ba7f6653e23ee7441605cff2dcbfdb4c3c48e288bd21b194654788f8bf95894\"" Dec 16 09:45:24.326953 containerd[1489]: time="2024-12-16T09:45:24.326928571Z" level=info msg="StartContainer for \"6ba7f6653e23ee7441605cff2dcbfdb4c3c48e288bd21b194654788f8bf95894\"" Dec 16 09:45:24.368741 systemd[1]: Started cri-containerd-6ba7f6653e23ee7441605cff2dcbfdb4c3c48e288bd21b194654788f8bf95894.scope - libcontainer container 6ba7f6653e23ee7441605cff2dcbfdb4c3c48e288bd21b194654788f8bf95894. Dec 16 09:45:24.406888 containerd[1489]: time="2024-12-16T09:45:24.406538130Z" level=info msg="StartContainer for \"6ba7f6653e23ee7441605cff2dcbfdb4c3c48e288bd21b194654788f8bf95894\" returns successfully" Dec 16 09:45:24.939695 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aesni)) Dec 16 09:45:25.320307 kubelet[2719]: I1216 09:45:25.320187 2719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-f622h" podStartSLOduration=5.32017176 podStartE2EDuration="5.32017176s" podCreationTimestamp="2024-12-16 09:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-12-16 09:45:25.31868007 +0000 UTC m=+349.890203452" watchObservedRunningTime="2024-12-16 09:45:25.32017176 +0000 UTC m=+349.891695131" Dec 16 09:45:27.783848 systemd-networkd[1391]: lxc_health: Link UP Dec 16 09:45:27.798095 systemd-networkd[1391]: lxc_health: Gained carrier Dec 16 09:45:28.158722 systemd[1]: run-containerd-runc-k8s.io-6ba7f6653e23ee7441605cff2dcbfdb4c3c48e288bd21b194654788f8bf95894-runc.KJ4Bzv.mount: Deactivated successfully. Dec 16 09:45:29.118867 systemd-networkd[1391]: lxc_health: Gained IPv6LL Dec 16 09:45:32.527360 systemd[1]: run-containerd-runc-k8s.io-6ba7f6653e23ee7441605cff2dcbfdb4c3c48e288bd21b194654788f8bf95894-runc.8B0BhA.mount: Deactivated successfully. Dec 16 09:45:34.901296 sshd[4643]: pam_unix(sshd:session): session closed for user core Dec 16 09:45:34.907252 systemd[1]: sshd@22-138.199.145.148:22-147.75.109.163:36122.service: Deactivated successfully. Dec 16 09:45:34.910427 systemd[1]: session-23.scope: Deactivated successfully. Dec 16 09:45:34.912418 systemd-logind[1477]: Session 23 logged out. Waiting for processes to exit. Dec 16 09:45:34.913971 systemd-logind[1477]: Removed session 23. Dec 16 09:45:35.564508 containerd[1489]: time="2024-12-16T09:45:35.564259189Z" level=info msg="StopPodSandbox for \"3b87f8f1710867438768b3941d4b273a4a8bc234af9b1b40cb15b9c0ba524cc0\"" Dec 16 09:45:35.564508 containerd[1489]: time="2024-12-16T09:45:35.564385505Z" level=info msg="TearDown network for sandbox \"3b87f8f1710867438768b3941d4b273a4a8bc234af9b1b40cb15b9c0ba524cc0\" successfully" Dec 16 09:45:35.564508 containerd[1489]: time="2024-12-16T09:45:35.564405443Z" level=info msg="StopPodSandbox for \"3b87f8f1710867438768b3941d4b273a4a8bc234af9b1b40cb15b9c0ba524cc0\" returns successfully" Dec 16 09:45:35.565614 containerd[1489]: time="2024-12-16T09:45:35.565464723Z" level=info msg="RemovePodSandbox for \"3b87f8f1710867438768b3941d4b273a4a8bc234af9b1b40cb15b9c0ba524cc0\"" Dec 16 09:45:35.567820 containerd[1489]: time="2024-12-16T09:45:35.567776476Z" level=info msg="Forcibly stopping sandbox \"3b87f8f1710867438768b3941d4b273a4a8bc234af9b1b40cb15b9c0ba524cc0\"" Dec 16 09:45:35.567931 containerd[1489]: time="2024-12-16T09:45:35.567871574Z" level=info msg="TearDown network for sandbox \"3b87f8f1710867438768b3941d4b273a4a8bc234af9b1b40cb15b9c0ba524cc0\" successfully" Dec 16 09:45:35.571691 containerd[1489]: time="2024-12-16T09:45:35.571633367Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"3b87f8f1710867438768b3941d4b273a4a8bc234af9b1b40cb15b9c0ba524cc0\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Dec 16 09:45:35.571795 containerd[1489]: time="2024-12-16T09:45:35.571690905Z" level=info msg="RemovePodSandbox \"3b87f8f1710867438768b3941d4b273a4a8bc234af9b1b40cb15b9c0ba524cc0\" returns successfully" Dec 16 09:45:35.572283 containerd[1489]: time="2024-12-16T09:45:35.572252455Z" level=info msg="StopPodSandbox for \"cabc0575ef95bbb1469e40d05f6827c256e2aaae5cb89b9ef5cb1790b3ba4225\"" Dec 16 09:45:35.572378 containerd[1489]: time="2024-12-16T09:45:35.572353343Z" level=info msg="TearDown network for sandbox \"cabc0575ef95bbb1469e40d05f6827c256e2aaae5cb89b9ef5cb1790b3ba4225\" successfully" Dec 16 09:45:35.572430 containerd[1489]: time="2024-12-16T09:45:35.572373561Z" level=info msg="StopPodSandbox for \"cabc0575ef95bbb1469e40d05f6827c256e2aaae5cb89b9ef5cb1790b3ba4225\" returns successfully" Dec 16 09:45:35.572760 containerd[1489]: time="2024-12-16T09:45:35.572718406Z" level=info msg="RemovePodSandbox for \"cabc0575ef95bbb1469e40d05f6827c256e2aaae5cb89b9ef5cb1790b3ba4225\"" Dec 16 09:45:35.572760 containerd[1489]: time="2024-12-16T09:45:35.572750847Z" level=info msg="Forcibly stopping sandbox \"cabc0575ef95bbb1469e40d05f6827c256e2aaae5cb89b9ef5cb1790b3ba4225\"" Dec 16 09:45:35.572869 containerd[1489]: time="2024-12-16T09:45:35.572825206Z" level=info msg="TearDown network for sandbox \"cabc0575ef95bbb1469e40d05f6827c256e2aaae5cb89b9ef5cb1790b3ba4225\" successfully" Dec 16 09:45:35.575819 containerd[1489]: time="2024-12-16T09:45:35.575743583Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"cabc0575ef95bbb1469e40d05f6827c256e2aaae5cb89b9ef5cb1790b3ba4225\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Dec 16 09:45:35.575819 containerd[1489]: time="2024-12-16T09:45:35.575810648Z" level=info msg="RemovePodSandbox \"cabc0575ef95bbb1469e40d05f6827c256e2aaae5cb89b9ef5cb1790b3ba4225\" returns successfully" Dec 16 09:45:50.175206 systemd[1]: cri-containerd-f6fed61c4091f9c718c7a3e6c3b6b244a4612e44741e3ec8c44aac6c654801a4.scope: Deactivated successfully. Dec 16 09:45:50.175467 systemd[1]: cri-containerd-f6fed61c4091f9c718c7a3e6c3b6b244a4612e44741e3ec8c44aac6c654801a4.scope: Consumed 5.498s CPU time, 20.3M memory peak, 0B memory swap peak. Dec 16 09:45:50.194812 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-f6fed61c4091f9c718c7a3e6c3b6b244a4612e44741e3ec8c44aac6c654801a4-rootfs.mount: Deactivated successfully. Dec 16 09:45:50.201409 containerd[1489]: time="2024-12-16T09:45:50.201316096Z" level=info msg="shim disconnected" id=f6fed61c4091f9c718c7a3e6c3b6b244a4612e44741e3ec8c44aac6c654801a4 namespace=k8s.io Dec 16 09:45:50.201409 containerd[1489]: time="2024-12-16T09:45:50.201370468Z" level=warning msg="cleaning up after shim disconnected" id=f6fed61c4091f9c718c7a3e6c3b6b244a4612e44741e3ec8c44aac6c654801a4 namespace=k8s.io Dec 16 09:45:50.201409 containerd[1489]: time="2024-12-16T09:45:50.201378824Z" level=info msg="cleaning up dead shim" namespace=k8s.io Dec 16 09:45:50.354167 kubelet[2719]: I1216 09:45:50.354102 2719 scope.go:117] "RemoveContainer" containerID="f6fed61c4091f9c718c7a3e6c3b6b244a4612e44741e3ec8c44aac6c654801a4" Dec 16 09:45:50.356418 containerd[1489]: time="2024-12-16T09:45:50.356375239Z" level=info msg="CreateContainer within sandbox \"7e5d76e65b344aad03f37883243005bc96b123137290468a466d701328c33b77\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:1,}" Dec 16 09:45:50.373975 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3811042020.mount: Deactivated successfully. Dec 16 09:45:50.375231 containerd[1489]: time="2024-12-16T09:45:50.375159013Z" level=info msg="CreateContainer within sandbox \"7e5d76e65b344aad03f37883243005bc96b123137290468a466d701328c33b77\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:1,} returns container id \"bb15446fa02e7b0ef8801440890c7de8424a5a702ed83732072c197131b11d63\"" Dec 16 09:45:50.375849 containerd[1489]: time="2024-12-16T09:45:50.375638992Z" level=info msg="StartContainer for \"bb15446fa02e7b0ef8801440890c7de8424a5a702ed83732072c197131b11d63\"" Dec 16 09:45:50.409685 systemd[1]: Started cri-containerd-bb15446fa02e7b0ef8801440890c7de8424a5a702ed83732072c197131b11d63.scope - libcontainer container bb15446fa02e7b0ef8801440890c7de8424a5a702ed83732072c197131b11d63. Dec 16 09:45:50.450310 containerd[1489]: time="2024-12-16T09:45:50.450179323Z" level=info msg="StartContainer for \"bb15446fa02e7b0ef8801440890c7de8424a5a702ed83732072c197131b11d63\" returns successfully" Dec 16 09:45:50.569350 kubelet[2719]: E1216 09:45:50.569303 2719 controller.go:195] "Failed to update lease" err="rpc error: code = Unavailable desc = error reading from server: read tcp 10.0.0.3:45052->10.0.0.2:2379: read: connection timed out" Dec 16 09:45:55.564330 kubelet[2719]: E1216 09:45:55.554219 2719 event.go:359] "Server rejected event (will not retry!)" err="rpc error: code = Unavailable desc = error reading from server: read tcp 10.0.0.3:44878->10.0.0.2:2379: read: connection timed out" event="&Event{ObjectMeta:{kube-apiserver-ci-4081-2-1-7-459a6b42ef.18119f2c84e56b3b kube-system 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:kube-system,Name:kube-apiserver-ci-4081-2-1-7-459a6b42ef,UID:051a64e47cbe729b1fee0d93ec1faff5,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Readiness probe failed: HTTP probe failed with statuscode: 500,Source:EventSource{Component:kubelet,Host:ci-4081-2-1-7-459a6b42ef,},FirstTimestamp:2024-12-16 09:45:45.071356731 +0000 UTC m=+369.642880103,LastTimestamp:2024-12-16 09:45:45.071356731 +0000 UTC m=+369.642880103,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4081-2-1-7-459a6b42ef,}" Dec 16 09:45:56.710902 systemd[1]: cri-containerd-f76f63d0e9ba0d2c306e06eb64ba47d546040ec2c148ce96b0d1450085ea20f2.scope: Deactivated successfully. Dec 16 09:45:56.711282 systemd[1]: cri-containerd-f76f63d0e9ba0d2c306e06eb64ba47d546040ec2c148ce96b0d1450085ea20f2.scope: Consumed 1.737s CPU time, 15.7M memory peak, 0B memory swap peak. Dec 16 09:45:56.737889 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-f76f63d0e9ba0d2c306e06eb64ba47d546040ec2c148ce96b0d1450085ea20f2-rootfs.mount: Deactivated successfully. Dec 16 09:45:56.744896 containerd[1489]: time="2024-12-16T09:45:56.744789192Z" level=info msg="shim disconnected" id=f76f63d0e9ba0d2c306e06eb64ba47d546040ec2c148ce96b0d1450085ea20f2 namespace=k8s.io Dec 16 09:45:56.744896 containerd[1489]: time="2024-12-16T09:45:56.744864201Z" level=warning msg="cleaning up after shim disconnected" id=f76f63d0e9ba0d2c306e06eb64ba47d546040ec2c148ce96b0d1450085ea20f2 namespace=k8s.io Dec 16 09:45:56.744896 containerd[1489]: time="2024-12-16T09:45:56.744873750Z" level=info msg="cleaning up dead shim" namespace=k8s.io Dec 16 09:45:57.371144 kubelet[2719]: I1216 09:45:57.371119 2719 scope.go:117] "RemoveContainer" containerID="f76f63d0e9ba0d2c306e06eb64ba47d546040ec2c148ce96b0d1450085ea20f2" Dec 16 09:45:57.372887 containerd[1489]: time="2024-12-16T09:45:57.372846760Z" level=info msg="CreateContainer within sandbox \"abe0fa2234a416b1b4ef32934846f5a4e4f92162262edaf37a15afa2b81bde98\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:1,}" Dec 16 09:45:57.388558 containerd[1489]: time="2024-12-16T09:45:57.388519267Z" level=info msg="CreateContainer within sandbox \"abe0fa2234a416b1b4ef32934846f5a4e4f92162262edaf37a15afa2b81bde98\" for &ContainerMetadata{Name:kube-scheduler,Attempt:1,} returns container id \"d33ec5d5cd82eb610e6f9d51150bb9b6cc23b25091f3889be6b7a17f4c10f93a\"" Dec 16 09:45:57.389601 containerd[1489]: time="2024-12-16T09:45:57.388930637Z" level=info msg="StartContainer for \"d33ec5d5cd82eb610e6f9d51150bb9b6cc23b25091f3889be6b7a17f4c10f93a\"" Dec 16 09:45:57.418706 systemd[1]: Started cri-containerd-d33ec5d5cd82eb610e6f9d51150bb9b6cc23b25091f3889be6b7a17f4c10f93a.scope - libcontainer container d33ec5d5cd82eb610e6f9d51150bb9b6cc23b25091f3889be6b7a17f4c10f93a. Dec 16 09:45:57.453611 containerd[1489]: time="2024-12-16T09:45:57.453426120Z" level=info msg="StartContainer for \"d33ec5d5cd82eb610e6f9d51150bb9b6cc23b25091f3889be6b7a17f4c10f93a\" returns successfully"