Dec 13 13:19:22.887336 kernel: Booting Linux on physical CPU 0x0000000000 [0x413fd0c1] Dec 13 13:19:22.887360 kernel: Linux version 6.6.65-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 14.2.1_p20241116 p3) 14.2.1 20241116, GNU ld (Gentoo 2.42 p6) 2.42.0) #1 SMP PREEMPT Fri Dec 13 11:56:07 -00 2024 Dec 13 13:19:22.887370 kernel: KASLR enabled Dec 13 13:19:22.887376 kernel: efi: EFI v2.7 by EDK II Dec 13 13:19:22.887382 kernel: efi: SMBIOS 3.0=0x135ed0000 MEMATTR=0x133c6b018 ACPI 2.0=0x132430018 RNG=0x13243e918 MEMRESERVE=0x132357218 Dec 13 13:19:22.887387 kernel: random: crng init done Dec 13 13:19:22.887395 kernel: secureboot: Secure boot disabled Dec 13 13:19:22.887400 kernel: ACPI: Early table checksum verification disabled Dec 13 13:19:22.887406 kernel: ACPI: RSDP 0x0000000132430018 000024 (v02 BOCHS ) Dec 13 13:19:22.887412 kernel: ACPI: XSDT 0x000000013243FE98 00006C (v01 BOCHS BXPC 00000001 01000013) Dec 13 13:19:22.887420 kernel: ACPI: FACP 0x000000013243FA98 000114 (v06 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 13:19:22.887426 kernel: ACPI: DSDT 0x0000000132437518 001468 (v02 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 13:19:22.887432 kernel: ACPI: APIC 0x000000013243FC18 000108 (v04 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 13:19:22.887438 kernel: ACPI: PPTT 0x000000013243FD98 000060 (v02 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 13:19:22.887445 kernel: ACPI: GTDT 0x000000013243D898 000060 (v02 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 13:19:22.887453 kernel: ACPI: MCFG 0x000000013243FF98 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 13:19:22.887459 kernel: ACPI: SPCR 0x000000013243E818 000050 (v02 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 13:19:22.887466 kernel: ACPI: DBG2 0x000000013243E898 000057 (v00 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 13:19:22.887472 kernel: ACPI: IORT 0x000000013243E418 000080 (v03 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 13:19:22.887478 kernel: ACPI: BGRT 0x000000013243E798 000038 (v01 INTEL EDK2 00000002 01000013) Dec 13 13:19:22.887484 kernel: ACPI: SPCR: console: pl011,mmio32,0x9000000,9600 Dec 13 13:19:22.887490 kernel: NUMA: Failed to initialise from firmware Dec 13 13:19:22.887497 kernel: NUMA: Faking a node at [mem 0x0000000040000000-0x0000000139ffffff] Dec 13 13:19:22.887503 kernel: NUMA: NODE_DATA [mem 0x139821800-0x139826fff] Dec 13 13:19:22.887509 kernel: Zone ranges: Dec 13 13:19:22.887515 kernel: DMA [mem 0x0000000040000000-0x00000000ffffffff] Dec 13 13:19:22.887522 kernel: DMA32 empty Dec 13 13:19:22.887529 kernel: Normal [mem 0x0000000100000000-0x0000000139ffffff] Dec 13 13:19:22.887535 kernel: Movable zone start for each node Dec 13 13:19:22.887541 kernel: Early memory node ranges Dec 13 13:19:22.887548 kernel: node 0: [mem 0x0000000040000000-0x000000013243ffff] Dec 13 13:19:22.887554 kernel: node 0: [mem 0x0000000132440000-0x000000013272ffff] Dec 13 13:19:22.887560 kernel: node 0: [mem 0x0000000132730000-0x0000000135bfffff] Dec 13 13:19:22.887566 kernel: node 0: [mem 0x0000000135c00000-0x0000000135fdffff] Dec 13 13:19:22.887572 kernel: node 0: [mem 0x0000000135fe0000-0x0000000139ffffff] Dec 13 13:19:22.887579 kernel: Initmem setup node 0 [mem 0x0000000040000000-0x0000000139ffffff] Dec 13 13:19:22.887585 kernel: On node 0, zone Normal: 24576 pages in unavailable ranges Dec 13 13:19:22.887592 kernel: psci: probing for conduit method from ACPI. Dec 13 13:19:22.887599 kernel: psci: PSCIv1.1 detected in firmware. Dec 13 13:19:22.887605 kernel: psci: Using standard PSCI v0.2 function IDs Dec 13 13:19:22.887614 kernel: psci: Trusted OS migration not required Dec 13 13:19:22.887621 kernel: psci: SMC Calling Convention v1.1 Dec 13 13:19:22.887628 kernel: smccc: KVM: hypervisor services detected (0x00000000 0x00000000 0x00000000 0x00000003) Dec 13 13:19:22.887636 kernel: percpu: Embedded 31 pages/cpu s86696 r8192 d32088 u126976 Dec 13 13:19:22.887642 kernel: pcpu-alloc: s86696 r8192 d32088 u126976 alloc=31*4096 Dec 13 13:19:22.887649 kernel: pcpu-alloc: [0] 0 [0] 1 Dec 13 13:19:22.887656 kernel: Detected PIPT I-cache on CPU0 Dec 13 13:19:22.887662 kernel: CPU features: detected: GIC system register CPU interface Dec 13 13:19:22.887695 kernel: CPU features: detected: Hardware dirty bit management Dec 13 13:19:22.887703 kernel: CPU features: detected: Spectre-v4 Dec 13 13:19:22.887710 kernel: CPU features: detected: Spectre-BHB Dec 13 13:19:22.887716 kernel: CPU features: kernel page table isolation forced ON by KASLR Dec 13 13:19:22.887723 kernel: CPU features: detected: Kernel page table isolation (KPTI) Dec 13 13:19:22.887730 kernel: CPU features: detected: ARM erratum 1418040 Dec 13 13:19:22.887739 kernel: CPU features: detected: SSBS not fully self-synchronizing Dec 13 13:19:22.887745 kernel: alternatives: applying boot alternatives Dec 13 13:19:22.887753 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyAMA0,115200n8 flatcar.first_boot=detected acpi=force flatcar.oem.id=hetzner verity.usrhash=c48af8adabdaf1d8e07ceb011d2665929c607ddf2c4d40203b31334d745cc472 Dec 13 13:19:22.887760 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Dec 13 13:19:22.887767 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Dec 13 13:19:22.887774 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Dec 13 13:19:22.887780 kernel: Fallback order for Node 0: 0 Dec 13 13:19:22.887787 kernel: Built 1 zonelists, mobility grouping on. Total pages: 1008000 Dec 13 13:19:22.887794 kernel: Policy zone: Normal Dec 13 13:19:22.887800 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Dec 13 13:19:22.887807 kernel: software IO TLB: area num 2. Dec 13 13:19:22.887815 kernel: software IO TLB: mapped [mem 0x00000000fbfff000-0x00000000fffff000] (64MB) Dec 13 13:19:22.887822 kernel: Memory: 3881024K/4096000K available (10304K kernel code, 2184K rwdata, 8088K rodata, 39936K init, 897K bss, 214976K reserved, 0K cma-reserved) Dec 13 13:19:22.887839 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Dec 13 13:19:22.887846 kernel: trace event string verifier disabled Dec 13 13:19:22.887853 kernel: rcu: Preemptible hierarchical RCU implementation. Dec 13 13:19:22.887860 kernel: rcu: RCU event tracing is enabled. Dec 13 13:19:22.887867 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Dec 13 13:19:22.887874 kernel: Trampoline variant of Tasks RCU enabled. Dec 13 13:19:22.887881 kernel: Tracing variant of Tasks RCU enabled. Dec 13 13:19:22.887887 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Dec 13 13:19:22.887894 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Dec 13 13:19:22.887903 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Dec 13 13:19:22.887910 kernel: GICv3: 256 SPIs implemented Dec 13 13:19:22.887916 kernel: GICv3: 0 Extended SPIs implemented Dec 13 13:19:22.887923 kernel: Root IRQ handler: gic_handle_irq Dec 13 13:19:22.887930 kernel: GICv3: GICv3 features: 16 PPIs, DirectLPI Dec 13 13:19:22.887936 kernel: GICv3: CPU0: found redistributor 0 region 0:0x00000000080a0000 Dec 13 13:19:22.887943 kernel: ITS [mem 0x08080000-0x0809ffff] Dec 13 13:19:22.887949 kernel: ITS@0x0000000008080000: allocated 8192 Devices @1000c0000 (indirect, esz 8, psz 64K, shr 1) Dec 13 13:19:22.887956 kernel: ITS@0x0000000008080000: allocated 8192 Interrupt Collections @1000d0000 (flat, esz 8, psz 64K, shr 1) Dec 13 13:19:22.887963 kernel: GICv3: using LPI property table @0x00000001000e0000 Dec 13 13:19:22.887970 kernel: GICv3: CPU0: using allocated LPI pending table @0x00000001000f0000 Dec 13 13:19:22.887976 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Dec 13 13:19:22.887987 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Dec 13 13:19:22.887994 kernel: arch_timer: cp15 timer(s) running at 25.00MHz (virt). Dec 13 13:19:22.888001 kernel: clocksource: arch_sys_counter: mask: 0xffffffffffffff max_cycles: 0x5c40939b5, max_idle_ns: 440795202646 ns Dec 13 13:19:22.888008 kernel: sched_clock: 56 bits at 25MHz, resolution 40ns, wraps every 4398046511100ns Dec 13 13:19:22.888015 kernel: Console: colour dummy device 80x25 Dec 13 13:19:22.888022 kernel: ACPI: Core revision 20230628 Dec 13 13:19:22.888029 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 50.00 BogoMIPS (lpj=25000) Dec 13 13:19:22.888036 kernel: pid_max: default: 32768 minimum: 301 Dec 13 13:19:22.888043 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Dec 13 13:19:22.888050 kernel: landlock: Up and running. Dec 13 13:19:22.888058 kernel: SELinux: Initializing. Dec 13 13:19:22.888065 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Dec 13 13:19:22.888073 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Dec 13 13:19:22.888080 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Dec 13 13:19:22.888087 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Dec 13 13:19:22.888094 kernel: rcu: Hierarchical SRCU implementation. Dec 13 13:19:22.888101 kernel: rcu: Max phase no-delay instances is 400. Dec 13 13:19:22.888108 kernel: Platform MSI: ITS@0x8080000 domain created Dec 13 13:19:22.888115 kernel: PCI/MSI: ITS@0x8080000 domain created Dec 13 13:19:22.888123 kernel: Remapping and enabling EFI services. Dec 13 13:19:22.888130 kernel: smp: Bringing up secondary CPUs ... Dec 13 13:19:22.888137 kernel: Detected PIPT I-cache on CPU1 Dec 13 13:19:22.888144 kernel: GICv3: CPU1: found redistributor 1 region 0:0x00000000080c0000 Dec 13 13:19:22.888151 kernel: GICv3: CPU1: using allocated LPI pending table @0x0000000100100000 Dec 13 13:19:22.888158 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Dec 13 13:19:22.888164 kernel: CPU1: Booted secondary processor 0x0000000001 [0x413fd0c1] Dec 13 13:19:22.888171 kernel: smp: Brought up 1 node, 2 CPUs Dec 13 13:19:22.888178 kernel: SMP: Total of 2 processors activated. Dec 13 13:19:22.888187 kernel: CPU features: detected: 32-bit EL0 Support Dec 13 13:19:22.888194 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence Dec 13 13:19:22.888206 kernel: CPU features: detected: Common not Private translations Dec 13 13:19:22.888214 kernel: CPU features: detected: CRC32 instructions Dec 13 13:19:22.888221 kernel: CPU features: detected: Enhanced Virtualization Traps Dec 13 13:19:22.888228 kernel: CPU features: detected: RCpc load-acquire (LDAPR) Dec 13 13:19:22.888236 kernel: CPU features: detected: LSE atomic instructions Dec 13 13:19:22.888243 kernel: CPU features: detected: Privileged Access Never Dec 13 13:19:22.888250 kernel: CPU features: detected: RAS Extension Support Dec 13 13:19:22.888259 kernel: CPU features: detected: Speculative Store Bypassing Safe (SSBS) Dec 13 13:19:22.888266 kernel: CPU: All CPU(s) started at EL1 Dec 13 13:19:22.888273 kernel: alternatives: applying system-wide alternatives Dec 13 13:19:22.888281 kernel: devtmpfs: initialized Dec 13 13:19:22.888288 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Dec 13 13:19:22.888295 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Dec 13 13:19:22.888303 kernel: pinctrl core: initialized pinctrl subsystem Dec 13 13:19:22.888310 kernel: SMBIOS 3.0.0 present. Dec 13 13:19:22.888318 kernel: DMI: Hetzner vServer/KVM Virtual Machine, BIOS 20171111 11/11/2017 Dec 13 13:19:22.888326 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Dec 13 13:19:22.888333 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations Dec 13 13:19:22.888341 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Dec 13 13:19:22.888348 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Dec 13 13:19:22.888356 kernel: audit: initializing netlink subsys (disabled) Dec 13 13:19:22.888363 kernel: audit: type=2000 audit(0.022:1): state=initialized audit_enabled=0 res=1 Dec 13 13:19:22.888371 kernel: thermal_sys: Registered thermal governor 'step_wise' Dec 13 13:19:22.888379 kernel: cpuidle: using governor menu Dec 13 13:19:22.888387 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Dec 13 13:19:22.888396 kernel: ASID allocator initialised with 32768 entries Dec 13 13:19:22.888404 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Dec 13 13:19:22.888413 kernel: Serial: AMBA PL011 UART driver Dec 13 13:19:22.888421 kernel: Modules: 2G module region forced by RANDOMIZE_MODULE_REGION_FULL Dec 13 13:19:22.888429 kernel: Modules: 0 pages in range for non-PLT usage Dec 13 13:19:22.888437 kernel: Modules: 508880 pages in range for PLT usage Dec 13 13:19:22.888444 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Dec 13 13:19:22.888453 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page Dec 13 13:19:22.888462 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages Dec 13 13:19:22.888469 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page Dec 13 13:19:22.888477 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Dec 13 13:19:22.888484 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page Dec 13 13:19:22.888491 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages Dec 13 13:19:22.888498 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page Dec 13 13:19:22.888506 kernel: ACPI: Added _OSI(Module Device) Dec 13 13:19:22.888513 kernel: ACPI: Added _OSI(Processor Device) Dec 13 13:19:22.888520 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Dec 13 13:19:22.888529 kernel: ACPI: Added _OSI(Processor Aggregator Device) Dec 13 13:19:22.888537 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Dec 13 13:19:22.888544 kernel: ACPI: Interpreter enabled Dec 13 13:19:22.888553 kernel: ACPI: Using GIC for interrupt routing Dec 13 13:19:22.888560 kernel: ACPI: MCFG table detected, 1 entries Dec 13 13:19:22.888567 kernel: ARMH0011:00: ttyAMA0 at MMIO 0x9000000 (irq = 12, base_baud = 0) is a SBSA Dec 13 13:19:22.888575 kernel: printk: console [ttyAMA0] enabled Dec 13 13:19:22.888582 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Dec 13 13:19:22.889757 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Dec 13 13:19:22.889883 kernel: acpi PNP0A08:00: _OSC: platform does not support [LTR] Dec 13 13:19:22.889952 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME AER PCIeCapability] Dec 13 13:19:22.890014 kernel: acpi PNP0A08:00: ECAM area [mem 0x4010000000-0x401fffffff] reserved by PNP0C02:00 Dec 13 13:19:22.890076 kernel: acpi PNP0A08:00: ECAM at [mem 0x4010000000-0x401fffffff] for [bus 00-ff] Dec 13 13:19:22.890085 kernel: ACPI: Remapped I/O 0x000000003eff0000 to [io 0x0000-0xffff window] Dec 13 13:19:22.890093 kernel: PCI host bridge to bus 0000:00 Dec 13 13:19:22.890164 kernel: pci_bus 0000:00: root bus resource [mem 0x10000000-0x3efeffff window] Dec 13 13:19:22.890224 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0xffff window] Dec 13 13:19:22.890287 kernel: pci_bus 0000:00: root bus resource [mem 0x8000000000-0xffffffffff window] Dec 13 13:19:22.890343 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Dec 13 13:19:22.890424 kernel: pci 0000:00:00.0: [1b36:0008] type 00 class 0x060000 Dec 13 13:19:22.890499 kernel: pci 0000:00:01.0: [1af4:1050] type 00 class 0x038000 Dec 13 13:19:22.890563 kernel: pci 0000:00:01.0: reg 0x14: [mem 0x11289000-0x11289fff] Dec 13 13:19:22.890629 kernel: pci 0000:00:01.0: reg 0x20: [mem 0x8000600000-0x8000603fff 64bit pref] Dec 13 13:19:22.891774 kernel: pci 0000:00:02.0: [1b36:000c] type 01 class 0x060400 Dec 13 13:19:22.891877 kernel: pci 0000:00:02.0: reg 0x10: [mem 0x11288000-0x11288fff] Dec 13 13:19:22.891953 kernel: pci 0000:00:02.1: [1b36:000c] type 01 class 0x060400 Dec 13 13:19:22.892018 kernel: pci 0000:00:02.1: reg 0x10: [mem 0x11287000-0x11287fff] Dec 13 13:19:22.892088 kernel: pci 0000:00:02.2: [1b36:000c] type 01 class 0x060400 Dec 13 13:19:22.892160 kernel: pci 0000:00:02.2: reg 0x10: [mem 0x11286000-0x11286fff] Dec 13 13:19:22.892229 kernel: pci 0000:00:02.3: [1b36:000c] type 01 class 0x060400 Dec 13 13:19:22.892293 kernel: pci 0000:00:02.3: reg 0x10: [mem 0x11285000-0x11285fff] Dec 13 13:19:22.892363 kernel: pci 0000:00:02.4: [1b36:000c] type 01 class 0x060400 Dec 13 13:19:22.892427 kernel: pci 0000:00:02.4: reg 0x10: [mem 0x11284000-0x11284fff] Dec 13 13:19:22.892503 kernel: pci 0000:00:02.5: [1b36:000c] type 01 class 0x060400 Dec 13 13:19:22.892570 kernel: pci 0000:00:02.5: reg 0x10: [mem 0x11283000-0x11283fff] Dec 13 13:19:22.892640 kernel: pci 0000:00:02.6: [1b36:000c] type 01 class 0x060400 Dec 13 13:19:22.893751 kernel: pci 0000:00:02.6: reg 0x10: [mem 0x11282000-0x11282fff] Dec 13 13:19:22.893871 kernel: pci 0000:00:02.7: [1b36:000c] type 01 class 0x060400 Dec 13 13:19:22.893946 kernel: pci 0000:00:02.7: reg 0x10: [mem 0x11281000-0x11281fff] Dec 13 13:19:22.894025 kernel: pci 0000:00:03.0: [1b36:000c] type 01 class 0x060400 Dec 13 13:19:22.894095 kernel: pci 0000:00:03.0: reg 0x10: [mem 0x11280000-0x11280fff] Dec 13 13:19:22.894166 kernel: pci 0000:00:04.0: [1b36:0002] type 00 class 0x070002 Dec 13 13:19:22.894229 kernel: pci 0000:00:04.0: reg 0x10: [io 0x8200-0x8207] Dec 13 13:19:22.894302 kernel: pci 0000:01:00.0: [1af4:1041] type 00 class 0x020000 Dec 13 13:19:22.894368 kernel: pci 0000:01:00.0: reg 0x14: [mem 0x11000000-0x11000fff] Dec 13 13:19:22.894433 kernel: pci 0000:01:00.0: reg 0x20: [mem 0x8000000000-0x8000003fff 64bit pref] Dec 13 13:19:22.894496 kernel: pci 0000:01:00.0: reg 0x30: [mem 0xfff80000-0xffffffff pref] Dec 13 13:19:22.894571 kernel: pci 0000:02:00.0: [1b36:000d] type 00 class 0x0c0330 Dec 13 13:19:22.894636 kernel: pci 0000:02:00.0: reg 0x10: [mem 0x10e00000-0x10e03fff 64bit] Dec 13 13:19:22.895793 kernel: pci 0000:03:00.0: [1af4:1043] type 00 class 0x078000 Dec 13 13:19:22.895890 kernel: pci 0000:03:00.0: reg 0x14: [mem 0x10c00000-0x10c00fff] Dec 13 13:19:22.895960 kernel: pci 0000:03:00.0: reg 0x20: [mem 0x8000100000-0x8000103fff 64bit pref] Dec 13 13:19:22.896034 kernel: pci 0000:04:00.0: [1af4:1045] type 00 class 0x00ff00 Dec 13 13:19:22.896109 kernel: pci 0000:04:00.0: reg 0x20: [mem 0x8000200000-0x8000203fff 64bit pref] Dec 13 13:19:22.896184 kernel: pci 0000:05:00.0: [1af4:1044] type 00 class 0x00ff00 Dec 13 13:19:22.896250 kernel: pci 0000:05:00.0: reg 0x20: [mem 0x8000300000-0x8000303fff 64bit pref] Dec 13 13:19:22.896325 kernel: pci 0000:06:00.0: [1af4:1048] type 00 class 0x010000 Dec 13 13:19:22.896391 kernel: pci 0000:06:00.0: reg 0x14: [mem 0x10600000-0x10600fff] Dec 13 13:19:22.896458 kernel: pci 0000:06:00.0: reg 0x20: [mem 0x8000400000-0x8000403fff 64bit pref] Dec 13 13:19:22.896537 kernel: pci 0000:07:00.0: [1af4:1041] type 00 class 0x020000 Dec 13 13:19:22.896607 kernel: pci 0000:07:00.0: reg 0x14: [mem 0x10400000-0x10400fff] Dec 13 13:19:22.896687 kernel: pci 0000:07:00.0: reg 0x20: [mem 0x8000500000-0x8000503fff 64bit pref] Dec 13 13:19:22.896756 kernel: pci 0000:07:00.0: reg 0x30: [mem 0xfff80000-0xffffffff pref] Dec 13 13:19:22.898007 kernel: pci 0000:00:02.0: bridge window [io 0x1000-0x0fff] to [bus 01] add_size 1000 Dec 13 13:19:22.898105 kernel: pci 0000:00:02.0: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 01] add_size 100000 add_align 100000 Dec 13 13:19:22.898174 kernel: pci 0000:00:02.0: bridge window [mem 0x00100000-0x001fffff] to [bus 01] add_size 100000 add_align 100000 Dec 13 13:19:22.898252 kernel: pci 0000:00:02.1: bridge window [io 0x1000-0x0fff] to [bus 02] add_size 1000 Dec 13 13:19:22.898317 kernel: pci 0000:00:02.1: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 02] add_size 200000 add_align 100000 Dec 13 13:19:22.898380 kernel: pci 0000:00:02.1: bridge window [mem 0x00100000-0x001fffff] to [bus 02] add_size 100000 add_align 100000 Dec 13 13:19:22.898447 kernel: pci 0000:00:02.2: bridge window [io 0x1000-0x0fff] to [bus 03] add_size 1000 Dec 13 13:19:22.898510 kernel: pci 0000:00:02.2: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 03] add_size 100000 add_align 100000 Dec 13 13:19:22.898573 kernel: pci 0000:00:02.2: bridge window [mem 0x00100000-0x001fffff] to [bus 03] add_size 100000 add_align 100000 Dec 13 13:19:22.898639 kernel: pci 0000:00:02.3: bridge window [io 0x1000-0x0fff] to [bus 04] add_size 1000 Dec 13 13:19:22.898736 kernel: pci 0000:00:02.3: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 04] add_size 100000 add_align 100000 Dec 13 13:19:22.898804 kernel: pci 0000:00:02.3: bridge window [mem 0x00100000-0x000fffff] to [bus 04] add_size 200000 add_align 100000 Dec 13 13:19:22.898891 kernel: pci 0000:00:02.4: bridge window [io 0x1000-0x0fff] to [bus 05] add_size 1000 Dec 13 13:19:22.898958 kernel: pci 0000:00:02.4: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 05] add_size 100000 add_align 100000 Dec 13 13:19:22.899021 kernel: pci 0000:00:02.4: bridge window [mem 0x00100000-0x000fffff] to [bus 05] add_size 200000 add_align 100000 Dec 13 13:19:22.899088 kernel: pci 0000:00:02.5: bridge window [io 0x1000-0x0fff] to [bus 06] add_size 1000 Dec 13 13:19:22.899151 kernel: pci 0000:00:02.5: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 06] add_size 100000 add_align 100000 Dec 13 13:19:22.899216 kernel: pci 0000:00:02.5: bridge window [mem 0x00100000-0x001fffff] to [bus 06] add_size 100000 add_align 100000 Dec 13 13:19:22.899287 kernel: pci 0000:00:02.6: bridge window [io 0x1000-0x0fff] to [bus 07] add_size 1000 Dec 13 13:19:22.899350 kernel: pci 0000:00:02.6: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 07] add_size 100000 add_align 100000 Dec 13 13:19:22.899413 kernel: pci 0000:00:02.6: bridge window [mem 0x00100000-0x001fffff] to [bus 07] add_size 100000 add_align 100000 Dec 13 13:19:22.899479 kernel: pci 0000:00:02.7: bridge window [io 0x1000-0x0fff] to [bus 08] add_size 1000 Dec 13 13:19:22.899542 kernel: pci 0000:00:02.7: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 08] add_size 200000 add_align 100000 Dec 13 13:19:22.899604 kernel: pci 0000:00:02.7: bridge window [mem 0x00100000-0x000fffff] to [bus 08] add_size 200000 add_align 100000 Dec 13 13:19:22.899688 kernel: pci 0000:00:03.0: bridge window [io 0x1000-0x0fff] to [bus 09] add_size 1000 Dec 13 13:19:22.899759 kernel: pci 0000:00:03.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 09] add_size 200000 add_align 100000 Dec 13 13:19:22.899823 kernel: pci 0000:00:03.0: bridge window [mem 0x00100000-0x000fffff] to [bus 09] add_size 200000 add_align 100000 Dec 13 13:19:22.899903 kernel: pci 0000:00:02.0: BAR 14: assigned [mem 0x10000000-0x101fffff] Dec 13 13:19:22.899969 kernel: pci 0000:00:02.0: BAR 15: assigned [mem 0x8000000000-0x80001fffff 64bit pref] Dec 13 13:19:22.900034 kernel: pci 0000:00:02.1: BAR 14: assigned [mem 0x10200000-0x103fffff] Dec 13 13:19:22.900098 kernel: pci 0000:00:02.1: BAR 15: assigned [mem 0x8000200000-0x80003fffff 64bit pref] Dec 13 13:19:22.900163 kernel: pci 0000:00:02.2: BAR 14: assigned [mem 0x10400000-0x105fffff] Dec 13 13:19:22.900230 kernel: pci 0000:00:02.2: BAR 15: assigned [mem 0x8000400000-0x80005fffff 64bit pref] Dec 13 13:19:22.900294 kernel: pci 0000:00:02.3: BAR 14: assigned [mem 0x10600000-0x107fffff] Dec 13 13:19:22.900359 kernel: pci 0000:00:02.3: BAR 15: assigned [mem 0x8000600000-0x80007fffff 64bit pref] Dec 13 13:19:22.900423 kernel: pci 0000:00:02.4: BAR 14: assigned [mem 0x10800000-0x109fffff] Dec 13 13:19:22.900486 kernel: pci 0000:00:02.4: BAR 15: assigned [mem 0x8000800000-0x80009fffff 64bit pref] Dec 13 13:19:22.900549 kernel: pci 0000:00:02.5: BAR 14: assigned [mem 0x10a00000-0x10bfffff] Dec 13 13:19:22.900612 kernel: pci 0000:00:02.5: BAR 15: assigned [mem 0x8000a00000-0x8000bfffff 64bit pref] Dec 13 13:19:22.901319 kernel: pci 0000:00:02.6: BAR 14: assigned [mem 0x10c00000-0x10dfffff] Dec 13 13:19:22.901420 kernel: pci 0000:00:02.6: BAR 15: assigned [mem 0x8000c00000-0x8000dfffff 64bit pref] Dec 13 13:19:22.901487 kernel: pci 0000:00:02.7: BAR 14: assigned [mem 0x10e00000-0x10ffffff] Dec 13 13:19:22.901549 kernel: pci 0000:00:02.7: BAR 15: assigned [mem 0x8000e00000-0x8000ffffff 64bit pref] Dec 13 13:19:22.901613 kernel: pci 0000:00:03.0: BAR 14: assigned [mem 0x11000000-0x111fffff] Dec 13 13:19:22.901693 kernel: pci 0000:00:03.0: BAR 15: assigned [mem 0x8001000000-0x80011fffff 64bit pref] Dec 13 13:19:22.901764 kernel: pci 0000:00:01.0: BAR 4: assigned [mem 0x8001200000-0x8001203fff 64bit pref] Dec 13 13:19:22.901850 kernel: pci 0000:00:01.0: BAR 1: assigned [mem 0x11200000-0x11200fff] Dec 13 13:19:22.901920 kernel: pci 0000:00:02.0: BAR 0: assigned [mem 0x11201000-0x11201fff] Dec 13 13:19:22.901982 kernel: pci 0000:00:02.0: BAR 13: assigned [io 0x1000-0x1fff] Dec 13 13:19:22.902046 kernel: pci 0000:00:02.1: BAR 0: assigned [mem 0x11202000-0x11202fff] Dec 13 13:19:22.902107 kernel: pci 0000:00:02.1: BAR 13: assigned [io 0x2000-0x2fff] Dec 13 13:19:22.902171 kernel: pci 0000:00:02.2: BAR 0: assigned [mem 0x11203000-0x11203fff] Dec 13 13:19:22.902232 kernel: pci 0000:00:02.2: BAR 13: assigned [io 0x3000-0x3fff] Dec 13 13:19:22.902296 kernel: pci 0000:00:02.3: BAR 0: assigned [mem 0x11204000-0x11204fff] Dec 13 13:19:22.902361 kernel: pci 0000:00:02.3: BAR 13: assigned [io 0x4000-0x4fff] Dec 13 13:19:22.902424 kernel: pci 0000:00:02.4: BAR 0: assigned [mem 0x11205000-0x11205fff] Dec 13 13:19:22.902489 kernel: pci 0000:00:02.4: BAR 13: assigned [io 0x5000-0x5fff] Dec 13 13:19:22.902552 kernel: pci 0000:00:02.5: BAR 0: assigned [mem 0x11206000-0x11206fff] Dec 13 13:19:22.902615 kernel: pci 0000:00:02.5: BAR 13: assigned [io 0x6000-0x6fff] Dec 13 13:19:22.904492 kernel: pci 0000:00:02.6: BAR 0: assigned [mem 0x11207000-0x11207fff] Dec 13 13:19:22.904611 kernel: pci 0000:00:02.6: BAR 13: assigned [io 0x7000-0x7fff] Dec 13 13:19:22.904695 kernel: pci 0000:00:02.7: BAR 0: assigned [mem 0x11208000-0x11208fff] Dec 13 13:19:22.904771 kernel: pci 0000:00:02.7: BAR 13: assigned [io 0x8000-0x8fff] Dec 13 13:19:22.904854 kernel: pci 0000:00:03.0: BAR 0: assigned [mem 0x11209000-0x11209fff] Dec 13 13:19:22.904922 kernel: pci 0000:00:03.0: BAR 13: assigned [io 0x9000-0x9fff] Dec 13 13:19:22.905000 kernel: pci 0000:00:04.0: BAR 0: assigned [io 0xa000-0xa007] Dec 13 13:19:22.905075 kernel: pci 0000:01:00.0: BAR 6: assigned [mem 0x10000000-0x1007ffff pref] Dec 13 13:19:22.905144 kernel: pci 0000:01:00.0: BAR 4: assigned [mem 0x8000000000-0x8000003fff 64bit pref] Dec 13 13:19:22.905215 kernel: pci 0000:01:00.0: BAR 1: assigned [mem 0x10080000-0x10080fff] Dec 13 13:19:22.905297 kernel: pci 0000:00:02.0: PCI bridge to [bus 01] Dec 13 13:19:22.905366 kernel: pci 0000:00:02.0: bridge window [io 0x1000-0x1fff] Dec 13 13:19:22.905439 kernel: pci 0000:00:02.0: bridge window [mem 0x10000000-0x101fffff] Dec 13 13:19:22.905504 kernel: pci 0000:00:02.0: bridge window [mem 0x8000000000-0x80001fffff 64bit pref] Dec 13 13:19:22.905573 kernel: pci 0000:02:00.0: BAR 0: assigned [mem 0x10200000-0x10203fff 64bit] Dec 13 13:19:22.905639 kernel: pci 0000:00:02.1: PCI bridge to [bus 02] Dec 13 13:19:22.905717 kernel: pci 0000:00:02.1: bridge window [io 0x2000-0x2fff] Dec 13 13:19:22.905780 kernel: pci 0000:00:02.1: bridge window [mem 0x10200000-0x103fffff] Dec 13 13:19:22.905857 kernel: pci 0000:00:02.1: bridge window [mem 0x8000200000-0x80003fffff 64bit pref] Dec 13 13:19:22.905930 kernel: pci 0000:03:00.0: BAR 4: assigned [mem 0x8000400000-0x8000403fff 64bit pref] Dec 13 13:19:22.905995 kernel: pci 0000:03:00.0: BAR 1: assigned [mem 0x10400000-0x10400fff] Dec 13 13:19:22.906060 kernel: pci 0000:00:02.2: PCI bridge to [bus 03] Dec 13 13:19:22.906122 kernel: pci 0000:00:02.2: bridge window [io 0x3000-0x3fff] Dec 13 13:19:22.906188 kernel: pci 0000:00:02.2: bridge window [mem 0x10400000-0x105fffff] Dec 13 13:19:22.906250 kernel: pci 0000:00:02.2: bridge window [mem 0x8000400000-0x80005fffff 64bit pref] Dec 13 13:19:22.906330 kernel: pci 0000:04:00.0: BAR 4: assigned [mem 0x8000600000-0x8000603fff 64bit pref] Dec 13 13:19:22.906398 kernel: pci 0000:00:02.3: PCI bridge to [bus 04] Dec 13 13:19:22.906461 kernel: pci 0000:00:02.3: bridge window [io 0x4000-0x4fff] Dec 13 13:19:22.906525 kernel: pci 0000:00:02.3: bridge window [mem 0x10600000-0x107fffff] Dec 13 13:19:22.906587 kernel: pci 0000:00:02.3: bridge window [mem 0x8000600000-0x80007fffff 64bit pref] Dec 13 13:19:22.906770 kernel: pci 0000:05:00.0: BAR 4: assigned [mem 0x8000800000-0x8000803fff 64bit pref] Dec 13 13:19:22.906921 kernel: pci 0000:00:02.4: PCI bridge to [bus 05] Dec 13 13:19:22.906994 kernel: pci 0000:00:02.4: bridge window [io 0x5000-0x5fff] Dec 13 13:19:22.907058 kernel: pci 0000:00:02.4: bridge window [mem 0x10800000-0x109fffff] Dec 13 13:19:22.907136 kernel: pci 0000:00:02.4: bridge window [mem 0x8000800000-0x80009fffff 64bit pref] Dec 13 13:19:22.907211 kernel: pci 0000:06:00.0: BAR 4: assigned [mem 0x8000a00000-0x8000a03fff 64bit pref] Dec 13 13:19:22.907278 kernel: pci 0000:06:00.0: BAR 1: assigned [mem 0x10a00000-0x10a00fff] Dec 13 13:19:22.907345 kernel: pci 0000:00:02.5: PCI bridge to [bus 06] Dec 13 13:19:22.907408 kernel: pci 0000:00:02.5: bridge window [io 0x6000-0x6fff] Dec 13 13:19:22.907475 kernel: pci 0000:00:02.5: bridge window [mem 0x10a00000-0x10bfffff] Dec 13 13:19:22.907538 kernel: pci 0000:00:02.5: bridge window [mem 0x8000a00000-0x8000bfffff 64bit pref] Dec 13 13:19:22.907609 kernel: pci 0000:07:00.0: BAR 6: assigned [mem 0x10c00000-0x10c7ffff pref] Dec 13 13:19:22.907707 kernel: pci 0000:07:00.0: BAR 4: assigned [mem 0x8000c00000-0x8000c03fff 64bit pref] Dec 13 13:19:22.907778 kernel: pci 0000:07:00.0: BAR 1: assigned [mem 0x10c80000-0x10c80fff] Dec 13 13:19:22.907880 kernel: pci 0000:00:02.6: PCI bridge to [bus 07] Dec 13 13:19:22.907949 kernel: pci 0000:00:02.6: bridge window [io 0x7000-0x7fff] Dec 13 13:19:22.908017 kernel: pci 0000:00:02.6: bridge window [mem 0x10c00000-0x10dfffff] Dec 13 13:19:22.908080 kernel: pci 0000:00:02.6: bridge window [mem 0x8000c00000-0x8000dfffff 64bit pref] Dec 13 13:19:22.908145 kernel: pci 0000:00:02.7: PCI bridge to [bus 08] Dec 13 13:19:22.908208 kernel: pci 0000:00:02.7: bridge window [io 0x8000-0x8fff] Dec 13 13:19:22.908270 kernel: pci 0000:00:02.7: bridge window [mem 0x10e00000-0x10ffffff] Dec 13 13:19:22.908336 kernel: pci 0000:00:02.7: bridge window [mem 0x8000e00000-0x8000ffffff 64bit pref] Dec 13 13:19:22.908402 kernel: pci 0000:00:03.0: PCI bridge to [bus 09] Dec 13 13:19:22.908467 kernel: pci 0000:00:03.0: bridge window [io 0x9000-0x9fff] Dec 13 13:19:22.908532 kernel: pci 0000:00:03.0: bridge window [mem 0x11000000-0x111fffff] Dec 13 13:19:22.908595 kernel: pci 0000:00:03.0: bridge window [mem 0x8001000000-0x80011fffff 64bit pref] Dec 13 13:19:22.908660 kernel: pci_bus 0000:00: resource 4 [mem 0x10000000-0x3efeffff window] Dec 13 13:19:22.908753 kernel: pci_bus 0000:00: resource 5 [io 0x0000-0xffff window] Dec 13 13:19:22.908816 kernel: pci_bus 0000:00: resource 6 [mem 0x8000000000-0xffffffffff window] Dec 13 13:19:22.908912 kernel: pci_bus 0000:01: resource 0 [io 0x1000-0x1fff] Dec 13 13:19:22.908974 kernel: pci_bus 0000:01: resource 1 [mem 0x10000000-0x101fffff] Dec 13 13:19:22.909038 kernel: pci_bus 0000:01: resource 2 [mem 0x8000000000-0x80001fffff 64bit pref] Dec 13 13:19:22.909104 kernel: pci_bus 0000:02: resource 0 [io 0x2000-0x2fff] Dec 13 13:19:22.909163 kernel: pci_bus 0000:02: resource 1 [mem 0x10200000-0x103fffff] Dec 13 13:19:22.909220 kernel: pci_bus 0000:02: resource 2 [mem 0x8000200000-0x80003fffff 64bit pref] Dec 13 13:19:22.909286 kernel: pci_bus 0000:03: resource 0 [io 0x3000-0x3fff] Dec 13 13:19:22.909344 kernel: pci_bus 0000:03: resource 1 [mem 0x10400000-0x105fffff] Dec 13 13:19:22.909402 kernel: pci_bus 0000:03: resource 2 [mem 0x8000400000-0x80005fffff 64bit pref] Dec 13 13:19:22.909471 kernel: pci_bus 0000:04: resource 0 [io 0x4000-0x4fff] Dec 13 13:19:22.909529 kernel: pci_bus 0000:04: resource 1 [mem 0x10600000-0x107fffff] Dec 13 13:19:22.909600 kernel: pci_bus 0000:04: resource 2 [mem 0x8000600000-0x80007fffff 64bit pref] Dec 13 13:19:22.909773 kernel: pci_bus 0000:05: resource 0 [io 0x5000-0x5fff] Dec 13 13:19:22.909900 kernel: pci_bus 0000:05: resource 1 [mem 0x10800000-0x109fffff] Dec 13 13:19:22.909963 kernel: pci_bus 0000:05: resource 2 [mem 0x8000800000-0x80009fffff 64bit pref] Dec 13 13:19:22.910029 kernel: pci_bus 0000:06: resource 0 [io 0x6000-0x6fff] Dec 13 13:19:22.910093 kernel: pci_bus 0000:06: resource 1 [mem 0x10a00000-0x10bfffff] Dec 13 13:19:22.910150 kernel: pci_bus 0000:06: resource 2 [mem 0x8000a00000-0x8000bfffff 64bit pref] Dec 13 13:19:22.910225 kernel: pci_bus 0000:07: resource 0 [io 0x7000-0x7fff] Dec 13 13:19:22.910308 kernel: pci_bus 0000:07: resource 1 [mem 0x10c00000-0x10dfffff] Dec 13 13:19:22.910388 kernel: pci_bus 0000:07: resource 2 [mem 0x8000c00000-0x8000dfffff 64bit pref] Dec 13 13:19:22.910454 kernel: pci_bus 0000:08: resource 0 [io 0x8000-0x8fff] Dec 13 13:19:22.910512 kernel: pci_bus 0000:08: resource 1 [mem 0x10e00000-0x10ffffff] Dec 13 13:19:22.910568 kernel: pci_bus 0000:08: resource 2 [mem 0x8000e00000-0x8000ffffff 64bit pref] Dec 13 13:19:22.910638 kernel: pci_bus 0000:09: resource 0 [io 0x9000-0x9fff] Dec 13 13:19:22.912342 kernel: pci_bus 0000:09: resource 1 [mem 0x11000000-0x111fffff] Dec 13 13:19:22.912427 kernel: pci_bus 0000:09: resource 2 [mem 0x8001000000-0x80011fffff 64bit pref] Dec 13 13:19:22.912438 kernel: ACPI: PCI: Interrupt link GSI0 configured for IRQ 35 Dec 13 13:19:22.912446 kernel: ACPI: PCI: Interrupt link GSI1 configured for IRQ 36 Dec 13 13:19:22.912454 kernel: ACPI: PCI: Interrupt link GSI2 configured for IRQ 37 Dec 13 13:19:22.912462 kernel: ACPI: PCI: Interrupt link GSI3 configured for IRQ 38 Dec 13 13:19:22.912470 kernel: iommu: Default domain type: Translated Dec 13 13:19:22.912478 kernel: iommu: DMA domain TLB invalidation policy: strict mode Dec 13 13:19:22.912485 kernel: efivars: Registered efivars operations Dec 13 13:19:22.912493 kernel: vgaarb: loaded Dec 13 13:19:22.912503 kernel: clocksource: Switched to clocksource arch_sys_counter Dec 13 13:19:22.912511 kernel: VFS: Disk quotas dquot_6.6.0 Dec 13 13:19:22.912519 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Dec 13 13:19:22.912527 kernel: pnp: PnP ACPI init Dec 13 13:19:22.912601 kernel: system 00:00: [mem 0x4010000000-0x401fffffff window] could not be reserved Dec 13 13:19:22.912613 kernel: pnp: PnP ACPI: found 1 devices Dec 13 13:19:22.912620 kernel: NET: Registered PF_INET protocol family Dec 13 13:19:22.912628 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Dec 13 13:19:22.912636 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Dec 13 13:19:22.912647 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Dec 13 13:19:22.912655 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Dec 13 13:19:22.912662 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Dec 13 13:19:22.912689 kernel: TCP: Hash tables configured (established 32768 bind 32768) Dec 13 13:19:22.912698 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Dec 13 13:19:22.912706 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Dec 13 13:19:22.912713 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Dec 13 13:19:22.912795 kernel: pci 0000:02:00.0: enabling device (0000 -> 0002) Dec 13 13:19:22.912810 kernel: PCI: CLS 0 bytes, default 64 Dec 13 13:19:22.912818 kernel: kvm [1]: HYP mode not available Dec 13 13:19:22.912841 kernel: Initialise system trusted keyrings Dec 13 13:19:22.912851 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Dec 13 13:19:22.912858 kernel: Key type asymmetric registered Dec 13 13:19:22.912866 kernel: Asymmetric key parser 'x509' registered Dec 13 13:19:22.912874 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Dec 13 13:19:22.912881 kernel: io scheduler mq-deadline registered Dec 13 13:19:22.912889 kernel: io scheduler kyber registered Dec 13 13:19:22.912899 kernel: io scheduler bfq registered Dec 13 13:19:22.912908 kernel: ACPI: \_SB_.PCI0.GSI2: Enabled at IRQ 37 Dec 13 13:19:22.912986 kernel: pcieport 0000:00:02.0: PME: Signaling with IRQ 50 Dec 13 13:19:22.913059 kernel: pcieport 0000:00:02.0: AER: enabled with IRQ 50 Dec 13 13:19:22.913123 kernel: pcieport 0000:00:02.0: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Dec 13 13:19:22.913198 kernel: pcieport 0000:00:02.1: PME: Signaling with IRQ 51 Dec 13 13:19:22.913268 kernel: pcieport 0000:00:02.1: AER: enabled with IRQ 51 Dec 13 13:19:22.913336 kernel: pcieport 0000:00:02.1: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Dec 13 13:19:22.913403 kernel: pcieport 0000:00:02.2: PME: Signaling with IRQ 52 Dec 13 13:19:22.913468 kernel: pcieport 0000:00:02.2: AER: enabled with IRQ 52 Dec 13 13:19:22.913531 kernel: pcieport 0000:00:02.2: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Dec 13 13:19:22.913600 kernel: pcieport 0000:00:02.3: PME: Signaling with IRQ 53 Dec 13 13:19:22.913684 kernel: pcieport 0000:00:02.3: AER: enabled with IRQ 53 Dec 13 13:19:22.913758 kernel: pcieport 0000:00:02.3: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Dec 13 13:19:22.913839 kernel: pcieport 0000:00:02.4: PME: Signaling with IRQ 54 Dec 13 13:19:22.913908 kernel: pcieport 0000:00:02.4: AER: enabled with IRQ 54 Dec 13 13:19:22.913972 kernel: pcieport 0000:00:02.4: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Dec 13 13:19:22.914039 kernel: pcieport 0000:00:02.5: PME: Signaling with IRQ 55 Dec 13 13:19:22.914104 kernel: pcieport 0000:00:02.5: AER: enabled with IRQ 55 Dec 13 13:19:22.914170 kernel: pcieport 0000:00:02.5: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Dec 13 13:19:22.914237 kernel: pcieport 0000:00:02.6: PME: Signaling with IRQ 56 Dec 13 13:19:22.914300 kernel: pcieport 0000:00:02.6: AER: enabled with IRQ 56 Dec 13 13:19:22.914364 kernel: pcieport 0000:00:02.6: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Dec 13 13:19:22.914431 kernel: pcieport 0000:00:02.7: PME: Signaling with IRQ 57 Dec 13 13:19:22.914495 kernel: pcieport 0000:00:02.7: AER: enabled with IRQ 57 Dec 13 13:19:22.914562 kernel: pcieport 0000:00:02.7: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Dec 13 13:19:22.914573 kernel: ACPI: \_SB_.PCI0.GSI3: Enabled at IRQ 38 Dec 13 13:19:22.914636 kernel: pcieport 0000:00:03.0: PME: Signaling with IRQ 58 Dec 13 13:19:22.917804 kernel: pcieport 0000:00:03.0: AER: enabled with IRQ 58 Dec 13 13:19:22.917950 kernel: pcieport 0000:00:03.0: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Dec 13 13:19:22.917963 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0C:00/input/input0 Dec 13 13:19:22.917972 kernel: ACPI: button: Power Button [PWRB] Dec 13 13:19:22.917987 kernel: ACPI: \_SB_.PCI0.GSI1: Enabled at IRQ 36 Dec 13 13:19:22.918060 kernel: virtio-pci 0000:03:00.0: enabling device (0000 -> 0002) Dec 13 13:19:22.918134 kernel: virtio-pci 0000:04:00.0: enabling device (0000 -> 0002) Dec 13 13:19:22.918206 kernel: virtio-pci 0000:07:00.0: enabling device (0000 -> 0002) Dec 13 13:19:22.918217 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Dec 13 13:19:22.918225 kernel: ACPI: \_SB_.PCI0.GSI0: Enabled at IRQ 35 Dec 13 13:19:22.918289 kernel: serial 0000:00:04.0: enabling device (0000 -> 0001) Dec 13 13:19:22.918300 kernel: 0000:00:04.0: ttyS0 at I/O 0xa000 (irq = 45, base_baud = 115200) is a 16550A Dec 13 13:19:22.918320 kernel: thunder_xcv, ver 1.0 Dec 13 13:19:22.918328 kernel: thunder_bgx, ver 1.0 Dec 13 13:19:22.918336 kernel: nicpf, ver 1.0 Dec 13 13:19:22.918352 kernel: nicvf, ver 1.0 Dec 13 13:19:22.918436 kernel: rtc-efi rtc-efi.0: registered as rtc0 Dec 13 13:19:22.918500 kernel: rtc-efi rtc-efi.0: setting system clock to 2024-12-13T13:19:22 UTC (1734095962) Dec 13 13:19:22.918511 kernel: hid: raw HID events driver (C) Jiri Kosina Dec 13 13:19:22.918519 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 7 counters available Dec 13 13:19:22.918529 kernel: watchdog: Delayed init of the lockup detector failed: -19 Dec 13 13:19:22.918537 kernel: watchdog: Hard watchdog permanently disabled Dec 13 13:19:22.918545 kernel: NET: Registered PF_INET6 protocol family Dec 13 13:19:22.918553 kernel: Segment Routing with IPv6 Dec 13 13:19:22.918560 kernel: In-situ OAM (IOAM) with IPv6 Dec 13 13:19:22.918568 kernel: NET: Registered PF_PACKET protocol family Dec 13 13:19:22.918576 kernel: Key type dns_resolver registered Dec 13 13:19:22.918584 kernel: registered taskstats version 1 Dec 13 13:19:22.918591 kernel: Loading compiled-in X.509 certificates Dec 13 13:19:22.918601 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.65-flatcar: 752b3e36c6039904ea643ccad2b3f5f3cb4ebf78' Dec 13 13:19:22.918608 kernel: Key type .fscrypt registered Dec 13 13:19:22.918616 kernel: Key type fscrypt-provisioning registered Dec 13 13:19:22.918624 kernel: ima: No TPM chip found, activating TPM-bypass! Dec 13 13:19:22.918632 kernel: ima: Allocated hash algorithm: sha1 Dec 13 13:19:22.918639 kernel: ima: No architecture policies found Dec 13 13:19:22.918647 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Dec 13 13:19:22.918655 kernel: clk: Disabling unused clocks Dec 13 13:19:22.918662 kernel: Freeing unused kernel memory: 39936K Dec 13 13:19:22.918755 kernel: Run /init as init process Dec 13 13:19:22.918763 kernel: with arguments: Dec 13 13:19:22.918771 kernel: /init Dec 13 13:19:22.918779 kernel: with environment: Dec 13 13:19:22.918789 kernel: HOME=/ Dec 13 13:19:22.918796 kernel: TERM=linux Dec 13 13:19:22.918804 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Dec 13 13:19:22.918814 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Dec 13 13:19:22.918836 systemd[1]: Detected virtualization kvm. Dec 13 13:19:22.918846 systemd[1]: Detected architecture arm64. Dec 13 13:19:22.918854 systemd[1]: Running in initrd. Dec 13 13:19:22.918862 systemd[1]: No hostname configured, using default hostname. Dec 13 13:19:22.918870 systemd[1]: Hostname set to . Dec 13 13:19:22.918878 systemd[1]: Initializing machine ID from VM UUID. Dec 13 13:19:22.918887 systemd[1]: Queued start job for default target initrd.target. Dec 13 13:19:22.918895 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Dec 13 13:19:22.918906 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Dec 13 13:19:22.918915 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Dec 13 13:19:22.918923 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Dec 13 13:19:22.918932 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Dec 13 13:19:22.918941 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Dec 13 13:19:22.918951 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Dec 13 13:19:22.918959 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Dec 13 13:19:22.918969 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Dec 13 13:19:22.918977 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Dec 13 13:19:22.918985 systemd[1]: Reached target paths.target - Path Units. Dec 13 13:19:22.918993 systemd[1]: Reached target slices.target - Slice Units. Dec 13 13:19:22.919001 systemd[1]: Reached target swap.target - Swaps. Dec 13 13:19:22.919009 systemd[1]: Reached target timers.target - Timer Units. Dec 13 13:19:22.919018 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Dec 13 13:19:22.919026 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Dec 13 13:19:22.919036 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Dec 13 13:19:22.919044 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Dec 13 13:19:22.919052 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Dec 13 13:19:22.919060 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Dec 13 13:19:22.919069 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Dec 13 13:19:22.919077 systemd[1]: Reached target sockets.target - Socket Units. Dec 13 13:19:22.919085 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Dec 13 13:19:22.919094 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Dec 13 13:19:22.919102 systemd[1]: Finished network-cleanup.service - Network Cleanup. Dec 13 13:19:22.919112 systemd[1]: Starting systemd-fsck-usr.service... Dec 13 13:19:22.919120 systemd[1]: Starting systemd-journald.service - Journal Service... Dec 13 13:19:22.919128 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Dec 13 13:19:22.919136 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Dec 13 13:19:22.919144 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Dec 13 13:19:22.919153 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Dec 13 13:19:22.919161 systemd[1]: Finished systemd-fsck-usr.service. Dec 13 13:19:22.919171 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Dec 13 13:19:22.919179 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Dec 13 13:19:22.919213 systemd-journald[236]: Collecting audit messages is disabled. Dec 13 13:19:22.919238 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Dec 13 13:19:22.919246 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Dec 13 13:19:22.919255 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Dec 13 13:19:22.919263 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Dec 13 13:19:22.919271 kernel: Bridge firewalling registered Dec 13 13:19:22.919279 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Dec 13 13:19:22.919288 systemd-journald[236]: Journal started Dec 13 13:19:22.919314 systemd-journald[236]: Runtime Journal (/run/log/journal/7a101661a12a43608b01093c64d75866) is 8.0M, max 76.5M, 68.5M free. Dec 13 13:19:22.888766 systemd-modules-load[237]: Inserted module 'overlay' Dec 13 13:19:22.911590 systemd-modules-load[237]: Inserted module 'br_netfilter' Dec 13 13:19:22.925101 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Dec 13 13:19:22.928402 systemd[1]: Started systemd-journald.service - Journal Service. Dec 13 13:19:22.928500 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Dec 13 13:19:22.946968 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Dec 13 13:19:22.948231 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Dec 13 13:19:22.949526 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Dec 13 13:19:22.953892 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Dec 13 13:19:22.958304 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Dec 13 13:19:22.961875 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Dec 13 13:19:22.972799 dracut-cmdline[272]: dracut-dracut-053 Dec 13 13:19:22.982967 dracut-cmdline[272]: Using kernel command line parameters: rd.driver.pre=btrfs BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyAMA0,115200n8 flatcar.first_boot=detected acpi=force flatcar.oem.id=hetzner verity.usrhash=c48af8adabdaf1d8e07ceb011d2665929c607ddf2c4d40203b31334d745cc472 Dec 13 13:19:23.001931 systemd-resolved[274]: Positive Trust Anchors: Dec 13 13:19:23.001947 systemd-resolved[274]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Dec 13 13:19:23.001978 systemd-resolved[274]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Dec 13 13:19:23.008538 systemd-resolved[274]: Defaulting to hostname 'linux'. Dec 13 13:19:23.009747 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Dec 13 13:19:23.011458 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Dec 13 13:19:23.078720 kernel: SCSI subsystem initialized Dec 13 13:19:23.083705 kernel: Loading iSCSI transport class v2.0-870. Dec 13 13:19:23.091725 kernel: iscsi: registered transport (tcp) Dec 13 13:19:23.104706 kernel: iscsi: registered transport (qla4xxx) Dec 13 13:19:23.104769 kernel: QLogic iSCSI HBA Driver Dec 13 13:19:23.151059 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Dec 13 13:19:23.158009 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Dec 13 13:19:23.175843 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Dec 13 13:19:23.175909 kernel: device-mapper: uevent: version 1.0.3 Dec 13 13:19:23.176698 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Dec 13 13:19:23.226734 kernel: raid6: neonx8 gen() 15670 MB/s Dec 13 13:19:23.243732 kernel: raid6: neonx4 gen() 15653 MB/s Dec 13 13:19:23.260735 kernel: raid6: neonx2 gen() 13148 MB/s Dec 13 13:19:23.277726 kernel: raid6: neonx1 gen() 10457 MB/s Dec 13 13:19:23.294723 kernel: raid6: int64x8 gen() 6761 MB/s Dec 13 13:19:23.311732 kernel: raid6: int64x4 gen() 7319 MB/s Dec 13 13:19:23.328737 kernel: raid6: int64x2 gen() 5983 MB/s Dec 13 13:19:23.345734 kernel: raid6: int64x1 gen() 5039 MB/s Dec 13 13:19:23.345794 kernel: raid6: using algorithm neonx8 gen() 15670 MB/s Dec 13 13:19:23.362732 kernel: raid6: .... xor() 11881 MB/s, rmw enabled Dec 13 13:19:23.362809 kernel: raid6: using neon recovery algorithm Dec 13 13:19:23.367816 kernel: xor: measuring software checksum speed Dec 13 13:19:23.367918 kernel: 8regs : 21664 MB/sec Dec 13 13:19:23.367939 kernel: 32regs : 21716 MB/sec Dec 13 13:19:23.367964 kernel: arm64_neon : 22917 MB/sec Dec 13 13:19:23.368711 kernel: xor: using function: arm64_neon (22917 MB/sec) Dec 13 13:19:23.417739 kernel: Btrfs loaded, zoned=no, fsverity=no Dec 13 13:19:23.433070 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Dec 13 13:19:23.439962 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Dec 13 13:19:23.453782 systemd-udevd[456]: Using default interface naming scheme 'v255'. Dec 13 13:19:23.457278 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Dec 13 13:19:23.468736 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Dec 13 13:19:23.483923 dracut-pre-trigger[464]: rd.md=0: removing MD RAID activation Dec 13 13:19:23.520380 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Dec 13 13:19:23.525913 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Dec 13 13:19:23.579753 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Dec 13 13:19:23.587401 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Dec 13 13:19:23.607236 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Dec 13 13:19:23.611094 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Dec 13 13:19:23.612863 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Dec 13 13:19:23.613456 systemd[1]: Reached target remote-fs.target - Remote File Systems. Dec 13 13:19:23.622964 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Dec 13 13:19:23.637406 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Dec 13 13:19:23.679011 kernel: scsi host0: Virtio SCSI HBA Dec 13 13:19:23.691748 kernel: scsi 0:0:0:0: CD-ROM QEMU QEMU CD-ROM 2.5+ PQ: 0 ANSI: 5 Dec 13 13:19:23.691841 kernel: scsi 0:0:0:1: Direct-Access QEMU QEMU HARDDISK 2.5+ PQ: 0 ANSI: 5 Dec 13 13:19:23.693184 kernel: ACPI: bus type USB registered Dec 13 13:19:23.695208 kernel: usbcore: registered new interface driver usbfs Dec 13 13:19:23.695698 kernel: usbcore: registered new interface driver hub Dec 13 13:19:23.696705 kernel: usbcore: registered new device driver usb Dec 13 13:19:23.724298 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Dec 13 13:19:23.726015 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Dec 13 13:19:23.730754 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Dec 13 13:19:23.732464 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Dec 13 13:19:23.732721 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Dec 13 13:19:23.734120 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Dec 13 13:19:23.745075 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Dec 13 13:19:23.755456 kernel: sr 0:0:0:0: Power-on or device reset occurred Dec 13 13:19:23.758803 kernel: sr 0:0:0:0: [sr0] scsi3-mmc drive: 16x/50x cd/rw xa/form2 cdda tray Dec 13 13:19:23.758985 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Dec 13 13:19:23.758997 kernel: sr 0:0:0:0: Attached scsi CD-ROM sr0 Dec 13 13:19:23.768024 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Dec 13 13:19:23.779886 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Dec 13 13:19:23.789458 kernel: sd 0:0:0:1: Power-on or device reset occurred Dec 13 13:19:23.809497 kernel: sd 0:0:0:1: [sda] 80003072 512-byte logical blocks: (41.0 GB/38.1 GiB) Dec 13 13:19:23.809614 kernel: sd 0:0:0:1: [sda] Write Protect is off Dec 13 13:19:23.809719 kernel: sd 0:0:0:1: [sda] Mode Sense: 63 00 00 08 Dec 13 13:19:23.809801 kernel: xhci_hcd 0000:02:00.0: xHCI Host Controller Dec 13 13:19:23.812184 kernel: sd 0:0:0:1: [sda] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA Dec 13 13:19:23.812309 kernel: xhci_hcd 0000:02:00.0: new USB bus registered, assigned bus number 1 Dec 13 13:19:23.812397 kernel: xhci_hcd 0000:02:00.0: hcc params 0x00087001 hci version 0x100 quirks 0x0000000000000010 Dec 13 13:19:23.812477 kernel: xhci_hcd 0000:02:00.0: xHCI Host Controller Dec 13 13:19:23.812555 kernel: xhci_hcd 0000:02:00.0: new USB bus registered, assigned bus number 2 Dec 13 13:19:23.812631 kernel: xhci_hcd 0000:02:00.0: Host supports USB 3.0 SuperSpeed Dec 13 13:19:23.812803 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Dec 13 13:19:23.812857 kernel: GPT:17805311 != 80003071 Dec 13 13:19:23.812870 kernel: GPT:Alternate GPT header not at the end of the disk. Dec 13 13:19:23.812881 kernel: GPT:17805311 != 80003071 Dec 13 13:19:23.812889 kernel: GPT: Use GNU Parted to correct GPT errors. Dec 13 13:19:23.812898 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Dec 13 13:19:23.812907 kernel: hub 1-0:1.0: USB hub found Dec 13 13:19:23.813022 kernel: sd 0:0:0:1: [sda] Attached SCSI disk Dec 13 13:19:23.813109 kernel: hub 1-0:1.0: 4 ports detected Dec 13 13:19:23.813187 kernel: usb usb2: We don't know the algorithms for LPM for this host, disabling LPM. Dec 13 13:19:23.813286 kernel: hub 2-0:1.0: USB hub found Dec 13 13:19:23.813368 kernel: hub 2-0:1.0: 4 ports detected Dec 13 13:19:23.813329 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Dec 13 13:19:23.861706 kernel: BTRFS: device fsid 47b12626-f7d3-4179-9720-ca262eb4c614 devid 1 transid 38 /dev/sda3 scanned by (udev-worker) (517) Dec 13 13:19:23.862072 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - QEMU_HARDDISK EFI-SYSTEM. Dec 13 13:19:23.868576 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/sda6 scanned by (udev-worker) (513) Dec 13 13:19:23.880840 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - QEMU_HARDDISK ROOT. Dec 13 13:19:23.890223 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - QEMU_HARDDISK OEM. Dec 13 13:19:23.894756 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - QEMU_HARDDISK USR-A. Dec 13 13:19:23.895524 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - QEMU_HARDDISK USR-A. Dec 13 13:19:23.901941 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Dec 13 13:19:23.912919 disk-uuid[579]: Primary Header is updated. Dec 13 13:19:23.912919 disk-uuid[579]: Secondary Entries is updated. Dec 13 13:19:23.912919 disk-uuid[579]: Secondary Header is updated. Dec 13 13:19:23.918835 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Dec 13 13:19:24.049805 kernel: usb 1-1: new high-speed USB device number 2 using xhci_hcd Dec 13 13:19:24.292757 kernel: usb 1-2: new high-speed USB device number 3 using xhci_hcd Dec 13 13:19:24.427701 kernel: input: QEMU QEMU USB Tablet as /devices/pci0000:00/0000:00:02.1/0000:02:00.0/usb1/1-1/1-1:1.0/0003:0627:0001.0001/input/input1 Dec 13 13:19:24.427776 kernel: hid-generic 0003:0627:0001.0001: input,hidraw0: USB HID v0.01 Mouse [QEMU QEMU USB Tablet] on usb-0000:02:00.0-1/input0 Dec 13 13:19:24.430712 kernel: input: QEMU QEMU USB Keyboard as /devices/pci0000:00/0000:00:02.1/0000:02:00.0/usb1/1-2/1-2:1.0/0003:0627:0001.0002/input/input2 Dec 13 13:19:24.484167 kernel: hid-generic 0003:0627:0001.0002: input,hidraw1: USB HID v1.11 Keyboard [QEMU QEMU USB Keyboard] on usb-0000:02:00.0-2/input0 Dec 13 13:19:24.484444 kernel: usbcore: registered new interface driver usbhid Dec 13 13:19:24.484489 kernel: usbhid: USB HID core driver Dec 13 13:19:24.932772 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Dec 13 13:19:24.935002 disk-uuid[580]: The operation has completed successfully. Dec 13 13:19:24.998000 systemd[1]: disk-uuid.service: Deactivated successfully. Dec 13 13:19:24.998830 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Dec 13 13:19:25.016956 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Dec 13 13:19:25.022825 sh[595]: Success Dec 13 13:19:25.038693 kernel: device-mapper: verity: sha256 using implementation "sha256-ce" Dec 13 13:19:25.102108 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Dec 13 13:19:25.112564 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Dec 13 13:19:25.113488 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Dec 13 13:19:25.134220 kernel: BTRFS info (device dm-0): first mount of filesystem 47b12626-f7d3-4179-9720-ca262eb4c614 Dec 13 13:19:25.134296 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm Dec 13 13:19:25.134317 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Dec 13 13:19:25.134346 kernel: BTRFS info (device dm-0): disabling log replay at mount time Dec 13 13:19:25.134970 kernel: BTRFS info (device dm-0): using free space tree Dec 13 13:19:25.140697 kernel: BTRFS info (device dm-0): enabling ssd optimizations Dec 13 13:19:25.144181 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Dec 13 13:19:25.144945 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Dec 13 13:19:25.151882 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Dec 13 13:19:25.156448 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Dec 13 13:19:25.169351 kernel: BTRFS info (device sda6): first mount of filesystem d0a3d620-8ab2-45d8-a26c-bb488ffd59f2 Dec 13 13:19:25.169407 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Dec 13 13:19:25.169419 kernel: BTRFS info (device sda6): using free space tree Dec 13 13:19:25.176725 kernel: BTRFS info (device sda6): enabling ssd optimizations Dec 13 13:19:25.176796 kernel: BTRFS info (device sda6): auto enabling async discard Dec 13 13:19:25.189557 systemd[1]: mnt-oem.mount: Deactivated successfully. Dec 13 13:19:25.190883 kernel: BTRFS info (device sda6): last unmount of filesystem d0a3d620-8ab2-45d8-a26c-bb488ffd59f2 Dec 13 13:19:25.197933 systemd[1]: Finished ignition-setup.service - Ignition (setup). Dec 13 13:19:25.205682 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Dec 13 13:19:25.308276 ignition[695]: Ignition 2.20.0 Dec 13 13:19:25.308285 ignition[695]: Stage: fetch-offline Dec 13 13:19:25.308321 ignition[695]: no configs at "/usr/lib/ignition/base.d" Dec 13 13:19:25.308330 ignition[695]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Dec 13 13:19:25.312683 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Dec 13 13:19:25.308489 ignition[695]: parsed url from cmdline: "" Dec 13 13:19:25.308492 ignition[695]: no config URL provided Dec 13 13:19:25.308496 ignition[695]: reading system config file "/usr/lib/ignition/user.ign" Dec 13 13:19:25.308503 ignition[695]: no config at "/usr/lib/ignition/user.ign" Dec 13 13:19:25.308507 ignition[695]: failed to fetch config: resource requires networking Dec 13 13:19:25.308889 ignition[695]: Ignition finished successfully Dec 13 13:19:25.326285 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Dec 13 13:19:25.332908 systemd[1]: Starting systemd-networkd.service - Network Configuration... Dec 13 13:19:25.371304 systemd-networkd[783]: lo: Link UP Dec 13 13:19:25.371316 systemd-networkd[783]: lo: Gained carrier Dec 13 13:19:25.373387 systemd-networkd[783]: Enumeration completed Dec 13 13:19:25.374138 systemd[1]: Started systemd-networkd.service - Network Configuration. Dec 13 13:19:25.374848 systemd-networkd[783]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 13 13:19:25.374852 systemd-networkd[783]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Dec 13 13:19:25.375687 systemd-networkd[783]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 13 13:19:25.375690 systemd-networkd[783]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. Dec 13 13:19:25.375719 systemd[1]: Reached target network.target - Network. Dec 13 13:19:25.376236 systemd-networkd[783]: eth0: Link UP Dec 13 13:19:25.376239 systemd-networkd[783]: eth0: Gained carrier Dec 13 13:19:25.376246 systemd-networkd[783]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 13 13:19:25.381950 systemd-networkd[783]: eth1: Link UP Dec 13 13:19:25.381953 systemd-networkd[783]: eth1: Gained carrier Dec 13 13:19:25.381963 systemd-networkd[783]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 13 13:19:25.382907 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Dec 13 13:19:25.397991 ignition[785]: Ignition 2.20.0 Dec 13 13:19:25.398001 ignition[785]: Stage: fetch Dec 13 13:19:25.398190 ignition[785]: no configs at "/usr/lib/ignition/base.d" Dec 13 13:19:25.398202 ignition[785]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Dec 13 13:19:25.398312 ignition[785]: parsed url from cmdline: "" Dec 13 13:19:25.398316 ignition[785]: no config URL provided Dec 13 13:19:25.398321 ignition[785]: reading system config file "/usr/lib/ignition/user.ign" Dec 13 13:19:25.398328 ignition[785]: no config at "/usr/lib/ignition/user.ign" Dec 13 13:19:25.398417 ignition[785]: GET http://169.254.169.254/hetzner/v1/userdata: attempt #1 Dec 13 13:19:25.399345 ignition[785]: GET error: Get "http://169.254.169.254/hetzner/v1/userdata": dial tcp 169.254.169.254:80: connect: network is unreachable Dec 13 13:19:25.410810 systemd-networkd[783]: eth1: DHCPv4 address 10.0.0.3/32, gateway 10.0.0.1 acquired from 10.0.0.1 Dec 13 13:19:25.433840 systemd-networkd[783]: eth0: DHCPv4 address 23.88.121.109/32, gateway 172.31.1.1 acquired from 172.31.1.1 Dec 13 13:19:25.599919 ignition[785]: GET http://169.254.169.254/hetzner/v1/userdata: attempt #2 Dec 13 13:19:25.604547 ignition[785]: GET result: OK Dec 13 13:19:25.604644 ignition[785]: parsing config with SHA512: be6d7eebc2555f1e75a157a67881daa2ad880f84dfb8b83ca4fae8bf86b6c9e5f27d08ecf39a117a299db46e3ee84d0d462fef5d3362e8980fe73a83c4b87789 Dec 13 13:19:25.611273 unknown[785]: fetched base config from "system" Dec 13 13:19:25.611283 unknown[785]: fetched base config from "system" Dec 13 13:19:25.611739 ignition[785]: fetch: fetch complete Dec 13 13:19:25.611289 unknown[785]: fetched user config from "hetzner" Dec 13 13:19:25.611746 ignition[785]: fetch: fetch passed Dec 13 13:19:25.615656 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Dec 13 13:19:25.611793 ignition[785]: Ignition finished successfully Dec 13 13:19:25.623864 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Dec 13 13:19:25.635995 ignition[793]: Ignition 2.20.0 Dec 13 13:19:25.636011 ignition[793]: Stage: kargs Dec 13 13:19:25.636189 ignition[793]: no configs at "/usr/lib/ignition/base.d" Dec 13 13:19:25.636199 ignition[793]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Dec 13 13:19:25.637131 ignition[793]: kargs: kargs passed Dec 13 13:19:25.637182 ignition[793]: Ignition finished successfully Dec 13 13:19:25.639542 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Dec 13 13:19:25.648197 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Dec 13 13:19:25.663622 ignition[799]: Ignition 2.20.0 Dec 13 13:19:25.663640 ignition[799]: Stage: disks Dec 13 13:19:25.663941 ignition[799]: no configs at "/usr/lib/ignition/base.d" Dec 13 13:19:25.663952 ignition[799]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Dec 13 13:19:25.665117 ignition[799]: disks: disks passed Dec 13 13:19:25.665182 ignition[799]: Ignition finished successfully Dec 13 13:19:25.668047 systemd[1]: Finished ignition-disks.service - Ignition (disks). Dec 13 13:19:25.669229 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Dec 13 13:19:25.670341 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Dec 13 13:19:25.672362 systemd[1]: Reached target local-fs.target - Local File Systems. Dec 13 13:19:25.674063 systemd[1]: Reached target sysinit.target - System Initialization. Dec 13 13:19:25.676096 systemd[1]: Reached target basic.target - Basic System. Dec 13 13:19:25.689984 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Dec 13 13:19:25.712129 systemd-fsck[808]: ROOT: clean, 14/1628000 files, 120691/1617920 blocks Dec 13 13:19:25.717504 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Dec 13 13:19:25.723830 systemd[1]: Mounting sysroot.mount - /sysroot... Dec 13 13:19:25.766698 kernel: EXT4-fs (sda9): mounted filesystem 0aa4851d-a2ba-4d04-90b3-5d00bf608ecc r/w with ordered data mode. Quota mode: none. Dec 13 13:19:25.768200 systemd[1]: Mounted sysroot.mount - /sysroot. Dec 13 13:19:25.770167 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Dec 13 13:19:25.775858 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Dec 13 13:19:25.778893 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Dec 13 13:19:25.789060 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Dec 13 13:19:25.790484 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Dec 13 13:19:25.790517 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Dec 13 13:19:25.799474 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Dec 13 13:19:25.804703 kernel: BTRFS: device label OEM devid 1 transid 15 /dev/sda6 scanned by mount (816) Dec 13 13:19:25.808150 kernel: BTRFS info (device sda6): first mount of filesystem d0a3d620-8ab2-45d8-a26c-bb488ffd59f2 Dec 13 13:19:25.808205 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Dec 13 13:19:25.808217 kernel: BTRFS info (device sda6): using free space tree Dec 13 13:19:25.810718 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Dec 13 13:19:25.821858 kernel: BTRFS info (device sda6): enabling ssd optimizations Dec 13 13:19:25.821920 kernel: BTRFS info (device sda6): auto enabling async discard Dec 13 13:19:25.827050 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Dec 13 13:19:25.856264 coreos-metadata[818]: Dec 13 13:19:25.855 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/hostname: Attempt #1 Dec 13 13:19:25.859046 coreos-metadata[818]: Dec 13 13:19:25.859 INFO Fetch successful Dec 13 13:19:25.861413 coreos-metadata[818]: Dec 13 13:19:25.859 INFO wrote hostname ci-4186-0-0-9-3a4b2091e9 to /sysroot/etc/hostname Dec 13 13:19:25.863584 initrd-setup-root[843]: cut: /sysroot/etc/passwd: No such file or directory Dec 13 13:19:25.863882 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Dec 13 13:19:25.870626 initrd-setup-root[851]: cut: /sysroot/etc/group: No such file or directory Dec 13 13:19:25.876649 initrd-setup-root[858]: cut: /sysroot/etc/shadow: No such file or directory Dec 13 13:19:25.882211 initrd-setup-root[865]: cut: /sysroot/etc/gshadow: No such file or directory Dec 13 13:19:25.986739 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Dec 13 13:19:25.992824 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Dec 13 13:19:25.995926 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Dec 13 13:19:26.003709 kernel: BTRFS info (device sda6): last unmount of filesystem d0a3d620-8ab2-45d8-a26c-bb488ffd59f2 Dec 13 13:19:26.028664 ignition[933]: INFO : Ignition 2.20.0 Dec 13 13:19:26.028664 ignition[933]: INFO : Stage: mount Dec 13 13:19:26.030867 ignition[933]: INFO : no configs at "/usr/lib/ignition/base.d" Dec 13 13:19:26.030867 ignition[933]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Dec 13 13:19:26.030867 ignition[933]: INFO : mount: mount passed Dec 13 13:19:26.030867 ignition[933]: INFO : Ignition finished successfully Dec 13 13:19:26.034381 systemd[1]: Finished ignition-mount.service - Ignition (mount). Dec 13 13:19:26.041945 systemd[1]: Starting ignition-files.service - Ignition (files)... Dec 13 13:19:26.044125 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Dec 13 13:19:26.134135 systemd[1]: sysroot-oem.mount: Deactivated successfully. Dec 13 13:19:26.139891 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Dec 13 13:19:26.154722 kernel: BTRFS: device label OEM devid 1 transid 16 /dev/sda6 scanned by mount (944) Dec 13 13:19:26.156048 kernel: BTRFS info (device sda6): first mount of filesystem d0a3d620-8ab2-45d8-a26c-bb488ffd59f2 Dec 13 13:19:26.156085 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Dec 13 13:19:26.156114 kernel: BTRFS info (device sda6): using free space tree Dec 13 13:19:26.161197 kernel: BTRFS info (device sda6): enabling ssd optimizations Dec 13 13:19:26.161244 kernel: BTRFS info (device sda6): auto enabling async discard Dec 13 13:19:26.165139 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Dec 13 13:19:26.189428 ignition[962]: INFO : Ignition 2.20.0 Dec 13 13:19:26.189428 ignition[962]: INFO : Stage: files Dec 13 13:19:26.190934 ignition[962]: INFO : no configs at "/usr/lib/ignition/base.d" Dec 13 13:19:26.190934 ignition[962]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Dec 13 13:19:26.190934 ignition[962]: DEBUG : files: compiled without relabeling support, skipping Dec 13 13:19:26.193521 ignition[962]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Dec 13 13:19:26.193521 ignition[962]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Dec 13 13:19:26.195554 ignition[962]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Dec 13 13:19:26.196541 ignition[962]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Dec 13 13:19:26.196541 ignition[962]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Dec 13 13:19:26.196116 unknown[962]: wrote ssh authorized keys file for user: core Dec 13 13:19:26.199485 ignition[962]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Dec 13 13:19:26.199485 ignition[962]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-arm64.tar.gz: attempt #1 Dec 13 13:19:26.346720 ignition[962]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Dec 13 13:19:26.625142 ignition[962]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Dec 13 13:19:26.625142 ignition[962]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Dec 13 13:19:26.627823 ignition[962]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-arm64.tar.gz: attempt #1 Dec 13 13:19:26.822036 systemd-networkd[783]: eth0: Gained IPv6LL Dec 13 13:19:27.206076 systemd-networkd[783]: eth1: Gained IPv6LL Dec 13 13:19:27.304254 ignition[962]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Dec 13 13:19:27.571379 ignition[962]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Dec 13 13:19:27.573317 ignition[962]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Dec 13 13:19:27.573317 ignition[962]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Dec 13 13:19:27.573317 ignition[962]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Dec 13 13:19:27.573317 ignition[962]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Dec 13 13:19:27.573317 ignition[962]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Dec 13 13:19:27.573317 ignition[962]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Dec 13 13:19:27.573317 ignition[962]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Dec 13 13:19:27.573317 ignition[962]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Dec 13 13:19:27.573317 ignition[962]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Dec 13 13:19:27.582576 ignition[962]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Dec 13 13:19:27.582576 ignition[962]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" Dec 13 13:19:27.582576 ignition[962]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" Dec 13 13:19:27.582576 ignition[962]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" Dec 13 13:19:27.582576 ignition[962]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.30.1-arm64.raw: attempt #1 Dec 13 13:19:28.093365 ignition[962]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Dec 13 13:19:28.405832 ignition[962]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" Dec 13 13:19:28.407348 ignition[962]: INFO : files: op(c): [started] processing unit "prepare-helm.service" Dec 13 13:19:28.408610 ignition[962]: INFO : files: op(c): op(d): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Dec 13 13:19:28.408610 ignition[962]: INFO : files: op(c): op(d): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Dec 13 13:19:28.408610 ignition[962]: INFO : files: op(c): [finished] processing unit "prepare-helm.service" Dec 13 13:19:28.408610 ignition[962]: INFO : files: op(e): [started] processing unit "coreos-metadata.service" Dec 13 13:19:28.408610 ignition[962]: INFO : files: op(e): op(f): [started] writing systemd drop-in "00-custom-metadata.conf" at "/sysroot/etc/systemd/system/coreos-metadata.service.d/00-custom-metadata.conf" Dec 13 13:19:28.408610 ignition[962]: INFO : files: op(e): op(f): [finished] writing systemd drop-in "00-custom-metadata.conf" at "/sysroot/etc/systemd/system/coreos-metadata.service.d/00-custom-metadata.conf" Dec 13 13:19:28.408610 ignition[962]: INFO : files: op(e): [finished] processing unit "coreos-metadata.service" Dec 13 13:19:28.408610 ignition[962]: INFO : files: op(10): [started] setting preset to enabled for "prepare-helm.service" Dec 13 13:19:28.408610 ignition[962]: INFO : files: op(10): [finished] setting preset to enabled for "prepare-helm.service" Dec 13 13:19:28.418947 ignition[962]: INFO : files: createResultFile: createFiles: op(11): [started] writing file "/sysroot/etc/.ignition-result.json" Dec 13 13:19:28.418947 ignition[962]: INFO : files: createResultFile: createFiles: op(11): [finished] writing file "/sysroot/etc/.ignition-result.json" Dec 13 13:19:28.418947 ignition[962]: INFO : files: files passed Dec 13 13:19:28.418947 ignition[962]: INFO : Ignition finished successfully Dec 13 13:19:28.411822 systemd[1]: Finished ignition-files.service - Ignition (files). Dec 13 13:19:28.419032 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Dec 13 13:19:28.426115 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Dec 13 13:19:28.434504 systemd[1]: ignition-quench.service: Deactivated successfully. Dec 13 13:19:28.434625 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Dec 13 13:19:28.442415 initrd-setup-root-after-ignition[989]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Dec 13 13:19:28.442415 initrd-setup-root-after-ignition[989]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Dec 13 13:19:28.445321 initrd-setup-root-after-ignition[993]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Dec 13 13:19:28.447899 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Dec 13 13:19:28.448850 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Dec 13 13:19:28.456985 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Dec 13 13:19:28.489131 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Dec 13 13:19:28.490696 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Dec 13 13:19:28.493291 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Dec 13 13:19:28.494089 systemd[1]: Reached target initrd.target - Initrd Default Target. Dec 13 13:19:28.495361 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Dec 13 13:19:28.496924 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Dec 13 13:19:28.524967 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Dec 13 13:19:28.535032 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Dec 13 13:19:28.545573 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Dec 13 13:19:28.546872 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Dec 13 13:19:28.548229 systemd[1]: Stopped target timers.target - Timer Units. Dec 13 13:19:28.549605 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Dec 13 13:19:28.549757 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Dec 13 13:19:28.552372 systemd[1]: Stopped target initrd.target - Initrd Default Target. Dec 13 13:19:28.553849 systemd[1]: Stopped target basic.target - Basic System. Dec 13 13:19:28.555156 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Dec 13 13:19:28.556701 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Dec 13 13:19:28.557891 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Dec 13 13:19:28.559330 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Dec 13 13:19:28.560643 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Dec 13 13:19:28.562004 systemd[1]: Stopped target sysinit.target - System Initialization. Dec 13 13:19:28.563392 systemd[1]: Stopped target local-fs.target - Local File Systems. Dec 13 13:19:28.564597 systemd[1]: Stopped target swap.target - Swaps. Dec 13 13:19:28.565638 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Dec 13 13:19:28.565823 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Dec 13 13:19:28.567309 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Dec 13 13:19:28.568140 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Dec 13 13:19:28.569417 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Dec 13 13:19:28.569493 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Dec 13 13:19:28.570681 systemd[1]: dracut-initqueue.service: Deactivated successfully. Dec 13 13:19:28.570823 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Dec 13 13:19:28.572478 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Dec 13 13:19:28.572597 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Dec 13 13:19:28.574064 systemd[1]: ignition-files.service: Deactivated successfully. Dec 13 13:19:28.574158 systemd[1]: Stopped ignition-files.service - Ignition (files). Dec 13 13:19:28.575125 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Dec 13 13:19:28.575219 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Dec 13 13:19:28.585071 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Dec 13 13:19:28.590741 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Dec 13 13:19:28.591399 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Dec 13 13:19:28.591559 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Dec 13 13:19:28.596855 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Dec 13 13:19:28.597035 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Dec 13 13:19:28.604598 systemd[1]: initrd-cleanup.service: Deactivated successfully. Dec 13 13:19:28.606823 ignition[1013]: INFO : Ignition 2.20.0 Dec 13 13:19:28.606823 ignition[1013]: INFO : Stage: umount Dec 13 13:19:28.606823 ignition[1013]: INFO : no configs at "/usr/lib/ignition/base.d" Dec 13 13:19:28.606823 ignition[1013]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Dec 13 13:19:28.611278 ignition[1013]: INFO : umount: umount passed Dec 13 13:19:28.611278 ignition[1013]: INFO : Ignition finished successfully Dec 13 13:19:28.607516 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Dec 13 13:19:28.611150 systemd[1]: ignition-mount.service: Deactivated successfully. Dec 13 13:19:28.612092 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Dec 13 13:19:28.616611 systemd[1]: sysroot-boot.mount: Deactivated successfully. Dec 13 13:19:28.617586 systemd[1]: ignition-disks.service: Deactivated successfully. Dec 13 13:19:28.618900 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Dec 13 13:19:28.620318 systemd[1]: ignition-kargs.service: Deactivated successfully. Dec 13 13:19:28.620382 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Dec 13 13:19:28.621638 systemd[1]: ignition-fetch.service: Deactivated successfully. Dec 13 13:19:28.621708 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Dec 13 13:19:28.622869 systemd[1]: Stopped target network.target - Network. Dec 13 13:19:28.623905 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Dec 13 13:19:28.623971 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Dec 13 13:19:28.625320 systemd[1]: Stopped target paths.target - Path Units. Dec 13 13:19:28.626352 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Dec 13 13:19:28.630813 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Dec 13 13:19:28.633333 systemd[1]: Stopped target slices.target - Slice Units. Dec 13 13:19:28.633970 systemd[1]: Stopped target sockets.target - Socket Units. Dec 13 13:19:28.635243 systemd[1]: iscsid.socket: Deactivated successfully. Dec 13 13:19:28.635292 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Dec 13 13:19:28.637079 systemd[1]: iscsiuio.socket: Deactivated successfully. Dec 13 13:19:28.637115 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Dec 13 13:19:28.639242 systemd[1]: ignition-setup.service: Deactivated successfully. Dec 13 13:19:28.639298 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Dec 13 13:19:28.640684 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Dec 13 13:19:28.640727 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Dec 13 13:19:28.642488 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Dec 13 13:19:28.643313 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Dec 13 13:19:28.644882 systemd[1]: sysroot-boot.service: Deactivated successfully. Dec 13 13:19:28.645048 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Dec 13 13:19:28.646270 systemd[1]: initrd-setup-root.service: Deactivated successfully. Dec 13 13:19:28.646362 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Dec 13 13:19:28.649736 systemd-networkd[783]: eth0: DHCPv6 lease lost Dec 13 13:19:28.652747 systemd-networkd[783]: eth1: DHCPv6 lease lost Dec 13 13:19:28.655721 systemd[1]: systemd-resolved.service: Deactivated successfully. Dec 13 13:19:28.655969 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Dec 13 13:19:28.660262 systemd[1]: systemd-networkd.service: Deactivated successfully. Dec 13 13:19:28.660711 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Dec 13 13:19:28.662058 systemd[1]: systemd-networkd.socket: Deactivated successfully. Dec 13 13:19:28.662093 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Dec 13 13:19:28.677961 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Dec 13 13:19:28.678584 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Dec 13 13:19:28.678664 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Dec 13 13:19:28.679509 systemd[1]: systemd-sysctl.service: Deactivated successfully. Dec 13 13:19:28.679566 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Dec 13 13:19:28.681938 systemd[1]: systemd-modules-load.service: Deactivated successfully. Dec 13 13:19:28.682003 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Dec 13 13:19:28.683264 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Dec 13 13:19:28.683314 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Dec 13 13:19:28.684636 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Dec 13 13:19:28.700447 systemd[1]: network-cleanup.service: Deactivated successfully. Dec 13 13:19:28.700647 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Dec 13 13:19:28.703726 systemd[1]: systemd-udevd.service: Deactivated successfully. Dec 13 13:19:28.703954 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Dec 13 13:19:28.705978 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Dec 13 13:19:28.706024 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Dec 13 13:19:28.707255 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Dec 13 13:19:28.707286 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Dec 13 13:19:28.708298 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Dec 13 13:19:28.708347 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Dec 13 13:19:28.710100 systemd[1]: dracut-cmdline.service: Deactivated successfully. Dec 13 13:19:28.710150 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Dec 13 13:19:28.711965 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Dec 13 13:19:28.712018 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Dec 13 13:19:28.724454 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Dec 13 13:19:28.725640 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Dec 13 13:19:28.725737 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Dec 13 13:19:28.726651 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Dec 13 13:19:28.726839 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Dec 13 13:19:28.728517 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Dec 13 13:19:28.728563 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Dec 13 13:19:28.729590 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Dec 13 13:19:28.729627 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Dec 13 13:19:28.736279 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Dec 13 13:19:28.736370 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Dec 13 13:19:28.738924 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Dec 13 13:19:28.747962 systemd[1]: Starting initrd-switch-root.service - Switch Root... Dec 13 13:19:28.758539 systemd[1]: Switching root. Dec 13 13:19:28.794851 systemd-journald[236]: Journal stopped Dec 13 13:19:29.771136 systemd-journald[236]: Received SIGTERM from PID 1 (systemd). Dec 13 13:19:29.771217 kernel: SELinux: policy capability network_peer_controls=1 Dec 13 13:19:29.771234 kernel: SELinux: policy capability open_perms=1 Dec 13 13:19:29.771243 kernel: SELinux: policy capability extended_socket_class=1 Dec 13 13:19:29.771254 kernel: SELinux: policy capability always_check_network=0 Dec 13 13:19:29.771263 kernel: SELinux: policy capability cgroup_seclabel=1 Dec 13 13:19:29.771272 kernel: SELinux: policy capability nnp_nosuid_transition=1 Dec 13 13:19:29.771281 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Dec 13 13:19:29.771290 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Dec 13 13:19:29.771300 kernel: audit: type=1403 audit(1734095968.992:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Dec 13 13:19:29.771310 systemd[1]: Successfully loaded SELinux policy in 35.175ms. Dec 13 13:19:29.771332 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 11.214ms. Dec 13 13:19:29.771343 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Dec 13 13:19:29.771353 systemd[1]: Detected virtualization kvm. Dec 13 13:19:29.771367 systemd[1]: Detected architecture arm64. Dec 13 13:19:29.771377 systemd[1]: Detected first boot. Dec 13 13:19:29.771386 systemd[1]: Hostname set to . Dec 13 13:19:29.771397 systemd[1]: Initializing machine ID from VM UUID. Dec 13 13:19:29.771412 zram_generator::config[1056]: No configuration found. Dec 13 13:19:29.771429 systemd[1]: Populated /etc with preset unit settings. Dec 13 13:19:29.771440 systemd[1]: initrd-switch-root.service: Deactivated successfully. Dec 13 13:19:29.771450 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Dec 13 13:19:29.771460 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Dec 13 13:19:29.771472 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Dec 13 13:19:29.771482 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Dec 13 13:19:29.771492 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Dec 13 13:19:29.771503 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Dec 13 13:19:29.771513 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Dec 13 13:19:29.771523 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Dec 13 13:19:29.771533 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Dec 13 13:19:29.771543 systemd[1]: Created slice user.slice - User and Session Slice. Dec 13 13:19:29.771552 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Dec 13 13:19:29.771562 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Dec 13 13:19:29.771573 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Dec 13 13:19:29.771585 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Dec 13 13:19:29.771595 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Dec 13 13:19:29.771605 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Dec 13 13:19:29.771616 systemd[1]: Expecting device dev-ttyAMA0.device - /dev/ttyAMA0... Dec 13 13:19:29.771626 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Dec 13 13:19:29.771636 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Dec 13 13:19:29.771646 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Dec 13 13:19:29.771657 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Dec 13 13:19:29.771685 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Dec 13 13:19:29.771698 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Dec 13 13:19:29.771709 systemd[1]: Reached target remote-fs.target - Remote File Systems. Dec 13 13:19:29.771718 systemd[1]: Reached target slices.target - Slice Units. Dec 13 13:19:29.771729 systemd[1]: Reached target swap.target - Swaps. Dec 13 13:19:29.771738 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Dec 13 13:19:29.771748 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Dec 13 13:19:29.771771 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Dec 13 13:19:29.771786 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Dec 13 13:19:29.771796 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Dec 13 13:19:29.771806 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Dec 13 13:19:29.771816 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Dec 13 13:19:29.771826 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Dec 13 13:19:29.771835 systemd[1]: Mounting media.mount - External Media Directory... Dec 13 13:19:29.771846 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Dec 13 13:19:29.771856 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Dec 13 13:19:29.771866 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Dec 13 13:19:29.771880 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Dec 13 13:19:29.771891 systemd[1]: Reached target machines.target - Containers. Dec 13 13:19:29.771901 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Dec 13 13:19:29.771912 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Dec 13 13:19:29.771922 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Dec 13 13:19:29.771933 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Dec 13 13:19:29.771943 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Dec 13 13:19:29.771953 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Dec 13 13:19:29.771963 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Dec 13 13:19:29.771973 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Dec 13 13:19:29.771983 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Dec 13 13:19:29.771994 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Dec 13 13:19:29.772004 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Dec 13 13:19:29.772015 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Dec 13 13:19:29.772025 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Dec 13 13:19:29.772035 systemd[1]: Stopped systemd-fsck-usr.service. Dec 13 13:19:29.772046 systemd[1]: Starting systemd-journald.service - Journal Service... Dec 13 13:19:29.772056 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Dec 13 13:19:29.772066 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Dec 13 13:19:29.772076 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Dec 13 13:19:29.772086 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Dec 13 13:19:29.772096 kernel: loop: module loaded Dec 13 13:19:29.772107 systemd[1]: verity-setup.service: Deactivated successfully. Dec 13 13:19:29.772117 systemd[1]: Stopped verity-setup.service. Dec 13 13:19:29.772126 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Dec 13 13:19:29.772137 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Dec 13 13:19:29.772146 systemd[1]: Mounted media.mount - External Media Directory. Dec 13 13:19:29.772158 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Dec 13 13:19:29.772168 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Dec 13 13:19:29.772177 kernel: fuse: init (API version 7.39) Dec 13 13:19:29.772190 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Dec 13 13:19:29.772204 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Dec 13 13:19:29.772214 systemd[1]: modprobe@configfs.service: Deactivated successfully. Dec 13 13:19:29.772224 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Dec 13 13:19:29.772233 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Dec 13 13:19:29.772243 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Dec 13 13:19:29.772293 systemd-journald[1137]: Collecting audit messages is disabled. Dec 13 13:19:29.772316 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Dec 13 13:19:29.772326 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Dec 13 13:19:29.772341 systemd[1]: modprobe@fuse.service: Deactivated successfully. Dec 13 13:19:29.772354 systemd-journald[1137]: Journal started Dec 13 13:19:29.772379 systemd-journald[1137]: Runtime Journal (/run/log/journal/7a101661a12a43608b01093c64d75866) is 8.0M, max 76.5M, 68.5M free. Dec 13 13:19:29.776717 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Dec 13 13:19:29.490448 systemd[1]: Queued start job for default target multi-user.target. Dec 13 13:19:29.509636 systemd[1]: Unnecessary job was removed for dev-sda6.device - /dev/sda6. Dec 13 13:19:29.510340 systemd[1]: systemd-journald.service: Deactivated successfully. Dec 13 13:19:29.780260 kernel: ACPI: bus type drm_connector registered Dec 13 13:19:29.780319 systemd[1]: Started systemd-journald.service - Journal Service. Dec 13 13:19:29.781356 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Dec 13 13:19:29.784828 systemd[1]: modprobe@drm.service: Deactivated successfully. Dec 13 13:19:29.784979 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Dec 13 13:19:29.786617 systemd[1]: modprobe@loop.service: Deactivated successfully. Dec 13 13:19:29.787168 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Dec 13 13:19:29.788219 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Dec 13 13:19:29.789372 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Dec 13 13:19:29.790548 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Dec 13 13:19:29.803995 systemd[1]: Reached target network-pre.target - Preparation for Network. Dec 13 13:19:29.809894 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Dec 13 13:19:29.813830 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Dec 13 13:19:29.816510 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Dec 13 13:19:29.816559 systemd[1]: Reached target local-fs.target - Local File Systems. Dec 13 13:19:29.819609 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Dec 13 13:19:29.828941 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Dec 13 13:19:29.834914 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Dec 13 13:19:29.837908 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Dec 13 13:19:29.844940 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Dec 13 13:19:29.851922 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Dec 13 13:19:29.852787 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Dec 13 13:19:29.857923 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Dec 13 13:19:29.858594 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Dec 13 13:19:29.863120 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Dec 13 13:19:29.873638 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Dec 13 13:19:29.883241 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Dec 13 13:19:29.886570 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Dec 13 13:19:29.889063 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Dec 13 13:19:29.892118 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Dec 13 13:19:29.917149 systemd-journald[1137]: Time spent on flushing to /var/log/journal/7a101661a12a43608b01093c64d75866 is 28.628ms for 1130 entries. Dec 13 13:19:29.917149 systemd-journald[1137]: System Journal (/var/log/journal/7a101661a12a43608b01093c64d75866) is 8.0M, max 584.8M, 576.8M free. Dec 13 13:19:29.957359 systemd-journald[1137]: Received client request to flush runtime journal. Dec 13 13:19:29.957403 kernel: loop0: detected capacity change from 0 to 116784 Dec 13 13:19:29.918125 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Dec 13 13:19:29.921340 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Dec 13 13:19:29.933923 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Dec 13 13:19:29.961256 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Dec 13 13:19:29.977638 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Dec 13 13:19:29.985990 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Dec 13 13:19:29.993837 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Dec 13 13:19:29.996944 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Dec 13 13:19:30.002045 systemd-tmpfiles[1171]: ACLs are not supported, ignoring. Dec 13 13:19:30.002428 systemd-tmpfiles[1171]: ACLs are not supported, ignoring. Dec 13 13:19:30.014410 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Dec 13 13:19:30.022988 kernel: loop1: detected capacity change from 0 to 194096 Dec 13 13:19:30.028964 systemd[1]: Starting systemd-sysusers.service - Create System Users... Dec 13 13:19:30.030293 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Dec 13 13:19:30.031593 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Dec 13 13:19:30.037189 udevadm[1186]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Dec 13 13:19:30.065739 kernel: loop2: detected capacity change from 0 to 8 Dec 13 13:19:30.088500 kernel: loop3: detected capacity change from 0 to 113552 Dec 13 13:19:30.099848 systemd[1]: Finished systemd-sysusers.service - Create System Users. Dec 13 13:19:30.109882 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Dec 13 13:19:30.130694 kernel: loop4: detected capacity change from 0 to 116784 Dec 13 13:19:30.143941 systemd-tmpfiles[1195]: ACLs are not supported, ignoring. Dec 13 13:19:30.143956 systemd-tmpfiles[1195]: ACLs are not supported, ignoring. Dec 13 13:19:30.150738 kernel: loop5: detected capacity change from 0 to 194096 Dec 13 13:19:30.152738 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Dec 13 13:19:30.191241 kernel: loop6: detected capacity change from 0 to 8 Dec 13 13:19:30.196703 kernel: loop7: detected capacity change from 0 to 113552 Dec 13 13:19:30.216405 (sd-merge)[1197]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-hetzner'. Dec 13 13:19:30.216925 (sd-merge)[1197]: Merged extensions into '/usr'. Dec 13 13:19:30.225924 systemd[1]: Reloading requested from client PID 1170 ('systemd-sysext') (unit systemd-sysext.service)... Dec 13 13:19:30.225945 systemd[1]: Reloading... Dec 13 13:19:30.342699 zram_generator::config[1222]: No configuration found. Dec 13 13:19:30.458590 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Dec 13 13:19:30.510459 ldconfig[1165]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Dec 13 13:19:30.514835 systemd[1]: Reloading finished in 287 ms. Dec 13 13:19:30.556250 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Dec 13 13:19:30.560406 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Dec 13 13:19:30.568074 systemd[1]: Starting ensure-sysext.service... Dec 13 13:19:30.573984 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Dec 13 13:19:30.585365 systemd[1]: Reloading requested from client PID 1261 ('systemctl') (unit ensure-sysext.service)... Dec 13 13:19:30.585383 systemd[1]: Reloading... Dec 13 13:19:30.599561 systemd-tmpfiles[1262]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Dec 13 13:19:30.600173 systemd-tmpfiles[1262]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Dec 13 13:19:30.601431 systemd-tmpfiles[1262]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Dec 13 13:19:30.601822 systemd-tmpfiles[1262]: ACLs are not supported, ignoring. Dec 13 13:19:30.602012 systemd-tmpfiles[1262]: ACLs are not supported, ignoring. Dec 13 13:19:30.605165 systemd-tmpfiles[1262]: Detected autofs mount point /boot during canonicalization of boot. Dec 13 13:19:30.605306 systemd-tmpfiles[1262]: Skipping /boot Dec 13 13:19:30.615621 systemd-tmpfiles[1262]: Detected autofs mount point /boot during canonicalization of boot. Dec 13 13:19:30.616299 systemd-tmpfiles[1262]: Skipping /boot Dec 13 13:19:30.664696 zram_generator::config[1289]: No configuration found. Dec 13 13:19:30.787662 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Dec 13 13:19:30.833598 systemd[1]: Reloading finished in 247 ms. Dec 13 13:19:30.849457 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Dec 13 13:19:30.855429 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Dec 13 13:19:30.872227 systemd[1]: Starting audit-rules.service - Load Audit Rules... Dec 13 13:19:30.878335 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Dec 13 13:19:30.882002 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Dec 13 13:19:30.892725 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Dec 13 13:19:30.897011 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Dec 13 13:19:30.903011 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Dec 13 13:19:30.905936 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Dec 13 13:19:30.911071 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Dec 13 13:19:30.917051 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Dec 13 13:19:30.922152 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Dec 13 13:19:30.923907 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Dec 13 13:19:30.925614 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Dec 13 13:19:30.925802 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Dec 13 13:19:30.930200 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Dec 13 13:19:30.934315 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Dec 13 13:19:30.935164 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Dec 13 13:19:30.942987 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Dec 13 13:19:30.946250 systemd[1]: Finished ensure-sysext.service. Dec 13 13:19:30.948685 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Dec 13 13:19:30.954350 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Dec 13 13:19:30.955448 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Dec 13 13:19:30.957714 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Dec 13 13:19:30.970077 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Dec 13 13:19:30.974949 systemd[1]: Starting systemd-update-done.service - Update is Completed... Dec 13 13:19:30.977166 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Dec 13 13:19:30.977360 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Dec 13 13:19:30.979128 systemd[1]: modprobe@drm.service: Deactivated successfully. Dec 13 13:19:30.982986 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Dec 13 13:19:30.997085 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Dec 13 13:19:31.002638 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Dec 13 13:19:31.006266 systemd[1]: modprobe@loop.service: Deactivated successfully. Dec 13 13:19:31.006409 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Dec 13 13:19:31.013576 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Dec 13 13:19:31.015500 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Dec 13 13:19:31.023555 systemd-udevd[1338]: Using default interface naming scheme 'v255'. Dec 13 13:19:31.035601 systemd[1]: Finished systemd-update-done.service - Update is Completed. Dec 13 13:19:31.037304 augenrules[1369]: No rules Dec 13 13:19:31.038824 systemd[1]: audit-rules.service: Deactivated successfully. Dec 13 13:19:31.039044 systemd[1]: Finished audit-rules.service - Load Audit Rules. Dec 13 13:19:31.045205 systemd[1]: Started systemd-userdbd.service - User Database Manager. Dec 13 13:19:31.066777 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Dec 13 13:19:31.078891 systemd[1]: Starting systemd-networkd.service - Network Configuration... Dec 13 13:19:31.143183 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Dec 13 13:19:31.144228 systemd[1]: Reached target time-set.target - System Time Set. Dec 13 13:19:31.147854 systemd-resolved[1337]: Positive Trust Anchors: Dec 13 13:19:31.147874 systemd-resolved[1337]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Dec 13 13:19:31.147906 systemd-resolved[1337]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Dec 13 13:19:31.154171 systemd-resolved[1337]: Using system hostname 'ci-4186-0-0-9-3a4b2091e9'. Dec 13 13:19:31.155899 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Dec 13 13:19:31.156850 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Dec 13 13:19:31.190366 systemd-networkd[1383]: lo: Link UP Dec 13 13:19:31.190376 systemd-networkd[1383]: lo: Gained carrier Dec 13 13:19:31.191806 systemd-networkd[1383]: Enumeration completed Dec 13 13:19:31.192052 systemd[1]: Started systemd-networkd.service - Network Configuration. Dec 13 13:19:31.192801 systemd[1]: Reached target network.target - Network. Dec 13 13:19:31.202073 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Dec 13 13:19:31.207135 systemd[1]: Condition check resulted in dev-ttyAMA0.device - /dev/ttyAMA0 being skipped. Dec 13 13:19:31.227747 kernel: BTRFS info: devid 1 device path /dev/mapper/usr changed to /dev/dm-0 scanned by (udev-worker) (1388) Dec 13 13:19:31.229690 kernel: BTRFS info: devid 1 device path /dev/dm-0 changed to /dev/mapper/usr scanned by (udev-worker) (1388) Dec 13 13:19:31.247144 systemd-networkd[1383]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 13 13:19:31.247771 systemd-networkd[1383]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Dec 13 13:19:31.250013 systemd-networkd[1383]: eth0: Link UP Dec 13 13:19:31.250020 systemd-networkd[1383]: eth0: Gained carrier Dec 13 13:19:31.250040 systemd-networkd[1383]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 13 13:19:31.294696 kernel: mousedev: PS/2 mouse device common for all mice Dec 13 13:19:31.305824 systemd-networkd[1383]: eth0: DHCPv4 address 23.88.121.109/32, gateway 172.31.1.1 acquired from 172.31.1.1 Dec 13 13:19:31.307790 systemd-timesyncd[1352]: Network configuration changed, trying to establish connection. Dec 13 13:19:31.322532 systemd-networkd[1383]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 13 13:19:31.322543 systemd-networkd[1383]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. Dec 13 13:19:31.324575 systemd-timesyncd[1352]: Network configuration changed, trying to establish connection. Dec 13 13:19:31.325071 systemd-networkd[1383]: eth1: Link UP Dec 13 13:19:31.325075 systemd-networkd[1383]: eth1: Gained carrier Dec 13 13:19:31.325095 systemd-networkd[1383]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 13 13:19:31.330053 systemd-timesyncd[1352]: Network configuration changed, trying to establish connection. Dec 13 13:19:31.331127 systemd[1]: Condition check resulted in dev-virtio\x2dports-org.qemu.guest_agent.0.device - /dev/virtio-ports/org.qemu.guest_agent.0 being skipped. Dec 13 13:19:31.331295 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Dec 13 13:19:31.341138 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Dec 13 13:19:31.344029 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Dec 13 13:19:31.348978 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Dec 13 13:19:31.349766 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Dec 13 13:19:31.349810 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Dec 13 13:19:31.350164 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Dec 13 13:19:31.350634 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Dec 13 13:19:31.358001 systemd-networkd[1383]: eth1: DHCPv4 address 10.0.0.3/32, gateway 10.0.0.1 acquired from 10.0.0.1 Dec 13 13:19:31.358944 systemd-timesyncd[1352]: Network configuration changed, trying to establish connection. Dec 13 13:19:31.372971 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Dec 13 13:19:31.373183 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Dec 13 13:19:31.374353 systemd[1]: modprobe@loop.service: Deactivated successfully. Dec 13 13:19:31.375310 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Dec 13 13:19:31.376446 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Dec 13 13:19:31.376513 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Dec 13 13:19:31.417702 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 38 scanned by (udev-worker) (1397) Dec 13 13:19:31.424038 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Dec 13 13:19:31.442840 kernel: [drm] pci: virtio-gpu-pci detected at 0000:00:01.0 Dec 13 13:19:31.442929 kernel: [drm] features: -virgl +edid -resource_blob -host_visible Dec 13 13:19:31.442946 kernel: [drm] features: -context_init Dec 13 13:19:31.445778 kernel: [drm] number of scanouts: 1 Dec 13 13:19:31.445869 kernel: [drm] number of cap sets: 0 Dec 13 13:19:31.452719 kernel: [drm] Initialized virtio_gpu 0.1.0 0 for 0000:00:01.0 on minor 0 Dec 13 13:19:31.461865 kernel: Console: switching to colour frame buffer device 160x50 Dec 13 13:19:31.473743 kernel: virtio-pci 0000:00:01.0: [drm] fb0: virtio_gpudrmfb frame buffer device Dec 13 13:19:31.477887 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - QEMU_HARDDISK OEM. Dec 13 13:19:31.486973 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Dec 13 13:19:31.489514 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Dec 13 13:19:31.490817 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Dec 13 13:19:31.502934 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Dec 13 13:19:31.506546 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Dec 13 13:19:31.555425 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Dec 13 13:19:31.604370 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Dec 13 13:19:31.620071 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Dec 13 13:19:31.638798 lvm[1447]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Dec 13 13:19:31.665344 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Dec 13 13:19:31.667886 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Dec 13 13:19:31.668842 systemd[1]: Reached target sysinit.target - System Initialization. Dec 13 13:19:31.669565 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Dec 13 13:19:31.670403 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Dec 13 13:19:31.671365 systemd[1]: Started logrotate.timer - Daily rotation of log files. Dec 13 13:19:31.672293 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Dec 13 13:19:31.673119 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Dec 13 13:19:31.673855 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Dec 13 13:19:31.673901 systemd[1]: Reached target paths.target - Path Units. Dec 13 13:19:31.674393 systemd[1]: Reached target timers.target - Timer Units. Dec 13 13:19:31.676256 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Dec 13 13:19:31.678518 systemd[1]: Starting docker.socket - Docker Socket for the API... Dec 13 13:19:31.684781 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Dec 13 13:19:31.689187 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Dec 13 13:19:31.692713 systemd[1]: Listening on docker.socket - Docker Socket for the API. Dec 13 13:19:31.693592 systemd[1]: Reached target sockets.target - Socket Units. Dec 13 13:19:31.694212 systemd[1]: Reached target basic.target - Basic System. Dec 13 13:19:31.694792 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Dec 13 13:19:31.694822 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Dec 13 13:19:31.700960 systemd[1]: Starting containerd.service - containerd container runtime... Dec 13 13:19:31.707148 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Dec 13 13:19:31.710922 lvm[1451]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Dec 13 13:19:31.715984 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Dec 13 13:19:31.722392 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Dec 13 13:19:31.730952 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Dec 13 13:19:31.731931 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Dec 13 13:19:31.743065 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Dec 13 13:19:31.747761 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Dec 13 13:19:31.757944 systemd[1]: Started qemu-guest-agent.service - QEMU Guest Agent. Dec 13 13:19:31.765962 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Dec 13 13:19:31.770834 coreos-metadata[1453]: Dec 13 13:19:31.769 INFO Fetching http://169.254.169.254/hetzner/v1/metadata: Attempt #1 Dec 13 13:19:31.771945 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Dec 13 13:19:31.774548 jq[1455]: false Dec 13 13:19:31.777161 systemd[1]: Starting systemd-logind.service - User Login Management... Dec 13 13:19:31.780548 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Dec 13 13:19:31.781584 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Dec 13 13:19:31.785140 systemd[1]: Starting update-engine.service - Update Engine... Dec 13 13:19:31.788947 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Dec 13 13:19:31.792994 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Dec 13 13:19:31.794030 coreos-metadata[1453]: Dec 13 13:19:31.793 INFO Fetch successful Dec 13 13:19:31.794030 coreos-metadata[1453]: Dec 13 13:19:31.793 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/private-networks: Attempt #1 Dec 13 13:19:31.799373 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Dec 13 13:19:31.799546 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Dec 13 13:19:31.799890 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Dec 13 13:19:31.800040 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Dec 13 13:19:31.803149 dbus-daemon[1454]: [system] SELinux support is enabled Dec 13 13:19:31.806312 systemd[1]: Started dbus.service - D-Bus System Message Bus. Dec 13 13:19:31.810236 coreos-metadata[1453]: Dec 13 13:19:31.810 INFO Fetch successful Dec 13 13:19:31.823092 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Dec 13 13:19:31.823167 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Dec 13 13:19:31.824003 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Dec 13 13:19:31.824021 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Dec 13 13:19:31.836111 extend-filesystems[1457]: Found loop4 Dec 13 13:19:31.836111 extend-filesystems[1457]: Found loop5 Dec 13 13:19:31.836111 extend-filesystems[1457]: Found loop6 Dec 13 13:19:31.836111 extend-filesystems[1457]: Found loop7 Dec 13 13:19:31.836111 extend-filesystems[1457]: Found sda Dec 13 13:19:31.836111 extend-filesystems[1457]: Found sda1 Dec 13 13:19:31.836111 extend-filesystems[1457]: Found sda2 Dec 13 13:19:31.836111 extend-filesystems[1457]: Found sda3 Dec 13 13:19:31.836111 extend-filesystems[1457]: Found usr Dec 13 13:19:31.836111 extend-filesystems[1457]: Found sda4 Dec 13 13:19:31.836111 extend-filesystems[1457]: Found sda6 Dec 13 13:19:31.836111 extend-filesystems[1457]: Found sda7 Dec 13 13:19:31.836111 extend-filesystems[1457]: Found sda9 Dec 13 13:19:31.836111 extend-filesystems[1457]: Checking size of /dev/sda9 Dec 13 13:19:31.855151 (ntainerd)[1477]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Dec 13 13:19:31.934954 update_engine[1468]: I20241213 13:19:31.875285 1468 main.cc:92] Flatcar Update Engine starting Dec 13 13:19:31.934954 update_engine[1468]: I20241213 13:19:31.882072 1468 update_check_scheduler.cc:74] Next update check in 7m16s Dec 13 13:19:31.935175 extend-filesystems[1457]: Resized partition /dev/sda9 Dec 13 13:19:31.941724 jq[1470]: true Dec 13 13:19:31.870370 systemd[1]: motdgen.service: Deactivated successfully. Dec 13 13:19:31.942041 extend-filesystems[1500]: resize2fs 1.47.1 (20-May-2024) Dec 13 13:19:31.870582 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Dec 13 13:19:31.893077 systemd[1]: Started update-engine.service - Update Engine. Dec 13 13:19:31.951830 tar[1475]: linux-arm64/helm Dec 13 13:19:31.981200 kernel: EXT4-fs (sda9): resizing filesystem from 1617920 to 9393147 blocks Dec 13 13:19:31.921989 systemd[1]: Started locksmithd.service - Cluster reboot manager. Dec 13 13:19:31.981407 jq[1489]: true Dec 13 13:19:31.942826 systemd-logind[1465]: New seat seat0. Dec 13 13:19:31.945925 systemd-logind[1465]: Watching system buttons on /dev/input/event0 (Power Button) Dec 13 13:19:31.945942 systemd-logind[1465]: Watching system buttons on /dev/input/event2 (QEMU QEMU USB Keyboard) Dec 13 13:19:31.946206 systemd[1]: Started systemd-logind.service - User Login Management. Dec 13 13:19:32.026641 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 38 scanned by (udev-worker) (1394) Dec 13 13:19:32.045890 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Dec 13 13:19:32.047632 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Dec 13 13:19:32.102888 locksmithd[1498]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Dec 13 13:19:32.118524 bash[1528]: Updated "/home/core/.ssh/authorized_keys" Dec 13 13:19:32.122103 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Dec 13 13:19:32.126854 kernel: EXT4-fs (sda9): resized filesystem to 9393147 Dec 13 13:19:32.135031 systemd[1]: Starting sshkeys.service... Dec 13 13:19:32.148925 extend-filesystems[1500]: Filesystem at /dev/sda9 is mounted on /; on-line resizing required Dec 13 13:19:32.148925 extend-filesystems[1500]: old_desc_blocks = 1, new_desc_blocks = 5 Dec 13 13:19:32.148925 extend-filesystems[1500]: The filesystem on /dev/sda9 is now 9393147 (4k) blocks long. Dec 13 13:19:32.159229 extend-filesystems[1457]: Resized filesystem in /dev/sda9 Dec 13 13:19:32.159229 extend-filesystems[1457]: Found sr0 Dec 13 13:19:32.149960 systemd[1]: extend-filesystems.service: Deactivated successfully. Dec 13 13:19:32.151721 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Dec 13 13:19:32.165642 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Dec 13 13:19:32.176360 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Dec 13 13:19:32.212773 coreos-metadata[1538]: Dec 13 13:19:32.212 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/public-keys: Attempt #1 Dec 13 13:19:32.217493 coreos-metadata[1538]: Dec 13 13:19:32.216 INFO Fetch successful Dec 13 13:19:32.222117 unknown[1538]: wrote ssh authorized keys file for user: core Dec 13 13:19:32.253724 update-ssh-keys[1542]: Updated "/home/core/.ssh/authorized_keys" Dec 13 13:19:32.254807 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Dec 13 13:19:32.261064 systemd[1]: Finished sshkeys.service. Dec 13 13:19:32.310947 containerd[1477]: time="2024-12-13T13:19:32.310826040Z" level=info msg="starting containerd" revision=9b2ad7760328148397346d10c7b2004271249db4 version=v1.7.23 Dec 13 13:19:32.372787 containerd[1477]: time="2024-12-13T13:19:32.372634680Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Dec 13 13:19:32.375675 containerd[1477]: time="2024-12-13T13:19:32.374888080Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.65-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Dec 13 13:19:32.375675 containerd[1477]: time="2024-12-13T13:19:32.374936640Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Dec 13 13:19:32.375675 containerd[1477]: time="2024-12-13T13:19:32.374957440Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Dec 13 13:19:32.375675 containerd[1477]: time="2024-12-13T13:19:32.375124800Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Dec 13 13:19:32.375675 containerd[1477]: time="2024-12-13T13:19:32.375142400Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Dec 13 13:19:32.375675 containerd[1477]: time="2024-12-13T13:19:32.375203760Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Dec 13 13:19:32.375675 containerd[1477]: time="2024-12-13T13:19:32.375215080Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Dec 13 13:19:32.375675 containerd[1477]: time="2024-12-13T13:19:32.375381320Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Dec 13 13:19:32.375675 containerd[1477]: time="2024-12-13T13:19:32.375395440Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Dec 13 13:19:32.375675 containerd[1477]: time="2024-12-13T13:19:32.375407240Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Dec 13 13:19:32.375675 containerd[1477]: time="2024-12-13T13:19:32.375416480Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Dec 13 13:19:32.375952 containerd[1477]: time="2024-12-13T13:19:32.375485080Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Dec 13 13:19:32.375952 containerd[1477]: time="2024-12-13T13:19:32.375755560Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Dec 13 13:19:32.375952 containerd[1477]: time="2024-12-13T13:19:32.375863560Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Dec 13 13:19:32.375952 containerd[1477]: time="2024-12-13T13:19:32.375875800Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Dec 13 13:19:32.376022 containerd[1477]: time="2024-12-13T13:19:32.375954360Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Dec 13 13:19:32.376022 containerd[1477]: time="2024-12-13T13:19:32.375995160Z" level=info msg="metadata content store policy set" policy=shared Dec 13 13:19:32.382682 containerd[1477]: time="2024-12-13T13:19:32.380974760Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Dec 13 13:19:32.382682 containerd[1477]: time="2024-12-13T13:19:32.381057160Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Dec 13 13:19:32.382682 containerd[1477]: time="2024-12-13T13:19:32.381074960Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Dec 13 13:19:32.382682 containerd[1477]: time="2024-12-13T13:19:32.381093480Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Dec 13 13:19:32.382682 containerd[1477]: time="2024-12-13T13:19:32.381111880Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Dec 13 13:19:32.382682 containerd[1477]: time="2024-12-13T13:19:32.381306800Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Dec 13 13:19:32.382682 containerd[1477]: time="2024-12-13T13:19:32.381575000Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Dec 13 13:19:32.382682 containerd[1477]: time="2024-12-13T13:19:32.381700560Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Dec 13 13:19:32.382682 containerd[1477]: time="2024-12-13T13:19:32.381723320Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Dec 13 13:19:32.382682 containerd[1477]: time="2024-12-13T13:19:32.381765880Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Dec 13 13:19:32.382682 containerd[1477]: time="2024-12-13T13:19:32.381781720Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Dec 13 13:19:32.382682 containerd[1477]: time="2024-12-13T13:19:32.381797080Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Dec 13 13:19:32.382682 containerd[1477]: time="2024-12-13T13:19:32.381812240Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Dec 13 13:19:32.382682 containerd[1477]: time="2024-12-13T13:19:32.381827040Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Dec 13 13:19:32.383044 containerd[1477]: time="2024-12-13T13:19:32.381842760Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Dec 13 13:19:32.383044 containerd[1477]: time="2024-12-13T13:19:32.381859680Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Dec 13 13:19:32.383044 containerd[1477]: time="2024-12-13T13:19:32.381873520Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Dec 13 13:19:32.383044 containerd[1477]: time="2024-12-13T13:19:32.381886640Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Dec 13 13:19:32.383044 containerd[1477]: time="2024-12-13T13:19:32.381910000Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Dec 13 13:19:32.383044 containerd[1477]: time="2024-12-13T13:19:32.381927800Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Dec 13 13:19:32.383044 containerd[1477]: time="2024-12-13T13:19:32.381944200Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Dec 13 13:19:32.383044 containerd[1477]: time="2024-12-13T13:19:32.381958600Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Dec 13 13:19:32.383044 containerd[1477]: time="2024-12-13T13:19:32.381972840Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Dec 13 13:19:32.383044 containerd[1477]: time="2024-12-13T13:19:32.381993360Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Dec 13 13:19:32.383044 containerd[1477]: time="2024-12-13T13:19:32.382015480Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Dec 13 13:19:32.383044 containerd[1477]: time="2024-12-13T13:19:32.382031480Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Dec 13 13:19:32.383044 containerd[1477]: time="2024-12-13T13:19:32.382052680Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Dec 13 13:19:32.383044 containerd[1477]: time="2024-12-13T13:19:32.382068480Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Dec 13 13:19:32.383287 containerd[1477]: time="2024-12-13T13:19:32.382082040Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Dec 13 13:19:32.383287 containerd[1477]: time="2024-12-13T13:19:32.382094520Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Dec 13 13:19:32.383287 containerd[1477]: time="2024-12-13T13:19:32.382107800Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Dec 13 13:19:32.383287 containerd[1477]: time="2024-12-13T13:19:32.382123880Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Dec 13 13:19:32.383287 containerd[1477]: time="2024-12-13T13:19:32.382151280Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Dec 13 13:19:32.383287 containerd[1477]: time="2024-12-13T13:19:32.382165520Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Dec 13 13:19:32.383287 containerd[1477]: time="2024-12-13T13:19:32.382177320Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Dec 13 13:19:32.383287 containerd[1477]: time="2024-12-13T13:19:32.382349480Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Dec 13 13:19:32.383287 containerd[1477]: time="2024-12-13T13:19:32.382369600Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Dec 13 13:19:32.383287 containerd[1477]: time="2024-12-13T13:19:32.382380840Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Dec 13 13:19:32.383287 containerd[1477]: time="2024-12-13T13:19:32.382393800Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Dec 13 13:19:32.383287 containerd[1477]: time="2024-12-13T13:19:32.382403400Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Dec 13 13:19:32.383287 containerd[1477]: time="2024-12-13T13:19:32.382415680Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Dec 13 13:19:32.383287 containerd[1477]: time="2024-12-13T13:19:32.382426000Z" level=info msg="NRI interface is disabled by configuration." Dec 13 13:19:32.383507 containerd[1477]: time="2024-12-13T13:19:32.382436400Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Dec 13 13:19:32.383528 containerd[1477]: time="2024-12-13T13:19:32.382981080Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Dec 13 13:19:32.383528 containerd[1477]: time="2024-12-13T13:19:32.383049960Z" level=info msg="Connect containerd service" Dec 13 13:19:32.383528 containerd[1477]: time="2024-12-13T13:19:32.383099840Z" level=info msg="using legacy CRI server" Dec 13 13:19:32.383528 containerd[1477]: time="2024-12-13T13:19:32.383109120Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Dec 13 13:19:32.383528 containerd[1477]: time="2024-12-13T13:19:32.383392480Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Dec 13 13:19:32.389130 containerd[1477]: time="2024-12-13T13:19:32.387613200Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Dec 13 13:19:32.389130 containerd[1477]: time="2024-12-13T13:19:32.388357560Z" level=info msg="Start subscribing containerd event" Dec 13 13:19:32.389130 containerd[1477]: time="2024-12-13T13:19:32.388603200Z" level=info msg="Start recovering state" Dec 13 13:19:32.389130 containerd[1477]: time="2024-12-13T13:19:32.388709280Z" level=info msg="Start event monitor" Dec 13 13:19:32.389130 containerd[1477]: time="2024-12-13T13:19:32.388759240Z" level=info msg="Start snapshots syncer" Dec 13 13:19:32.389130 containerd[1477]: time="2024-12-13T13:19:32.388773800Z" level=info msg="Start cni network conf syncer for default" Dec 13 13:19:32.389130 containerd[1477]: time="2024-12-13T13:19:32.388781240Z" level=info msg="Start streaming server" Dec 13 13:19:32.391678 containerd[1477]: time="2024-12-13T13:19:32.389504880Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Dec 13 13:19:32.391678 containerd[1477]: time="2024-12-13T13:19:32.389571000Z" level=info msg=serving... address=/run/containerd/containerd.sock Dec 13 13:19:32.391102 systemd[1]: Started containerd.service - containerd container runtime. Dec 13 13:19:32.392554 containerd[1477]: time="2024-12-13T13:19:32.392525600Z" level=info msg="containerd successfully booted in 0.088292s" Dec 13 13:19:32.603744 tar[1475]: linux-arm64/LICENSE Dec 13 13:19:32.603988 tar[1475]: linux-arm64/README.md Dec 13 13:19:32.617277 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Dec 13 13:19:32.645863 systemd-networkd[1383]: eth1: Gained IPv6LL Dec 13 13:19:32.647692 systemd-timesyncd[1352]: Network configuration changed, trying to establish connection. Dec 13 13:19:32.651978 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Dec 13 13:19:32.653308 systemd[1]: Reached target network-online.target - Network is Online. Dec 13 13:19:32.664048 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 13 13:19:32.668703 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Dec 13 13:19:32.721130 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Dec 13 13:19:32.749634 sshd_keygen[1490]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Dec 13 13:19:32.778095 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Dec 13 13:19:32.792567 systemd[1]: Starting issuegen.service - Generate /run/issue... Dec 13 13:19:32.798134 systemd[1]: issuegen.service: Deactivated successfully. Dec 13 13:19:32.799034 systemd[1]: Finished issuegen.service - Generate /run/issue. Dec 13 13:19:32.807419 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Dec 13 13:19:32.819326 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Dec 13 13:19:32.828424 systemd[1]: Started getty@tty1.service - Getty on tty1. Dec 13 13:19:32.831557 systemd[1]: Started serial-getty@ttyAMA0.service - Serial Getty on ttyAMA0. Dec 13 13:19:32.834072 systemd[1]: Reached target getty.target - Login Prompts. Dec 13 13:19:33.030571 systemd-networkd[1383]: eth0: Gained IPv6LL Dec 13 13:19:33.031872 systemd-timesyncd[1352]: Network configuration changed, trying to establish connection. Dec 13 13:19:33.379244 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 13 13:19:33.384057 systemd[1]: Reached target multi-user.target - Multi-User System. Dec 13 13:19:33.384177 (kubelet)[1584]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 13 13:19:33.389769 systemd[1]: Startup finished in 784ms (kernel) + 6.310s (initrd) + 4.432s (userspace) = 11.527s. Dec 13 13:19:33.402298 agetty[1578]: failed to open credentials directory Dec 13 13:19:33.402355 agetty[1579]: failed to open credentials directory Dec 13 13:19:33.932655 kubelet[1584]: E1213 13:19:33.932599 1584 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 13 13:19:33.935599 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 13 13:19:33.936128 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 13 13:19:44.186694 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Dec 13 13:19:44.194413 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 13 13:19:44.301416 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 13 13:19:44.307077 (kubelet)[1605]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 13 13:19:44.363902 kubelet[1605]: E1213 13:19:44.362754 1605 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 13 13:19:44.366716 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 13 13:19:44.366847 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 13 13:19:54.429351 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Dec 13 13:19:54.435144 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 13 13:19:54.557297 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 13 13:19:54.572036 (kubelet)[1621]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 13 13:19:54.623650 kubelet[1621]: E1213 13:19:54.623604 1621 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 13 13:19:54.626271 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 13 13:19:54.626418 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 13 13:20:03.811204 systemd-resolved[1337]: Clock change detected. Flushing caches. Dec 13 13:20:03.812085 systemd-timesyncd[1352]: Contacted time server 141.82.25.202:123 (2.flatcar.pool.ntp.org). Dec 13 13:20:03.812298 systemd-timesyncd[1352]: Initial clock synchronization to Fri 2024-12-13 13:20:03.811130 UTC. Dec 13 13:20:05.102304 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Dec 13 13:20:05.109120 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 13 13:20:05.216666 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 13 13:20:05.225140 (kubelet)[1638]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 13 13:20:05.273289 kubelet[1638]: E1213 13:20:05.273228 1638 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 13 13:20:05.275391 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 13 13:20:05.275533 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 13 13:20:15.352428 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 4. Dec 13 13:20:15.370608 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 13 13:20:15.488496 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 13 13:20:15.493647 (kubelet)[1654]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 13 13:20:15.549586 kubelet[1654]: E1213 13:20:15.549530 1654 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 13 13:20:15.553843 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 13 13:20:15.554187 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 13 13:20:17.303930 update_engine[1468]: I20241213 13:20:17.303411 1468 update_attempter.cc:509] Updating boot flags... Dec 13 13:20:17.350808 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 38 scanned by (udev-worker) (1670) Dec 13 13:20:17.417830 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 38 scanned by (udev-worker) (1673) Dec 13 13:20:17.468766 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 38 scanned by (udev-worker) (1673) Dec 13 13:20:25.602181 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 5. Dec 13 13:20:25.609152 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 13 13:20:25.715170 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 13 13:20:25.726302 (kubelet)[1690]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 13 13:20:25.769767 kubelet[1690]: E1213 13:20:25.769697 1690 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 13 13:20:25.772565 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 13 13:20:25.772909 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 13 13:20:35.852248 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 6. Dec 13 13:20:35.860056 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 13 13:20:35.979066 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 13 13:20:35.979127 (kubelet)[1707]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 13 13:20:36.027850 kubelet[1707]: E1213 13:20:36.027794 1707 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 13 13:20:36.030304 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 13 13:20:36.030534 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 13 13:20:46.102268 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 7. Dec 13 13:20:46.116060 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 13 13:20:46.241488 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 13 13:20:46.258308 (kubelet)[1723]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 13 13:20:46.307353 kubelet[1723]: E1213 13:20:46.307296 1723 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 13 13:20:46.309645 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 13 13:20:46.309801 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 13 13:20:56.352439 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 8. Dec 13 13:20:56.370045 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 13 13:20:56.488665 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 13 13:20:56.501322 (kubelet)[1739]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 13 13:20:56.550580 kubelet[1739]: E1213 13:20:56.550520 1739 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 13 13:20:56.552524 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 13 13:20:56.552647 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 13 13:21:06.602255 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 9. Dec 13 13:21:06.617162 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 13 13:21:06.727264 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 13 13:21:06.733373 (kubelet)[1755]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 13 13:21:06.780821 kubelet[1755]: E1213 13:21:06.780764 1755 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 13 13:21:06.783618 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 13 13:21:06.783789 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 13 13:21:16.852497 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 10. Dec 13 13:21:16.869671 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 13 13:21:17.002060 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 13 13:21:17.002193 (kubelet)[1771]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 13 13:21:17.056722 kubelet[1771]: E1213 13:21:17.056654 1771 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 13 13:21:17.059857 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 13 13:21:17.060178 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 13 13:21:19.809821 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Dec 13 13:21:19.811309 systemd[1]: Started sshd@0-23.88.121.109:22-147.75.109.163:45410.service - OpenSSH per-connection server daemon (147.75.109.163:45410). Dec 13 13:21:20.814149 sshd[1780]: Accepted publickey for core from 147.75.109.163 port 45410 ssh2: RSA SHA256:mujnIWCHI6g7RhOIFUnU4L23td/nvZma7inzBKSUIRw Dec 13 13:21:20.817262 sshd-session[1780]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 13:21:20.829871 systemd-logind[1465]: New session 1 of user core. Dec 13 13:21:20.831819 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Dec 13 13:21:20.843194 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Dec 13 13:21:20.856407 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Dec 13 13:21:20.865245 systemd[1]: Starting user@500.service - User Manager for UID 500... Dec 13 13:21:20.868322 (systemd)[1784]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Dec 13 13:21:20.983795 systemd[1784]: Queued start job for default target default.target. Dec 13 13:21:20.993871 systemd[1784]: Created slice app.slice - User Application Slice. Dec 13 13:21:20.994054 systemd[1784]: Reached target paths.target - Paths. Dec 13 13:21:20.994138 systemd[1784]: Reached target timers.target - Timers. Dec 13 13:21:20.995949 systemd[1784]: Starting dbus.socket - D-Bus User Message Bus Socket... Dec 13 13:21:21.011072 systemd[1784]: Listening on dbus.socket - D-Bus User Message Bus Socket. Dec 13 13:21:21.011422 systemd[1784]: Reached target sockets.target - Sockets. Dec 13 13:21:21.011442 systemd[1784]: Reached target basic.target - Basic System. Dec 13 13:21:21.011497 systemd[1784]: Reached target default.target - Main User Target. Dec 13 13:21:21.011528 systemd[1784]: Startup finished in 136ms. Dec 13 13:21:21.011880 systemd[1]: Started user@500.service - User Manager for UID 500. Dec 13 13:21:21.022142 systemd[1]: Started session-1.scope - Session 1 of User core. Dec 13 13:21:21.724325 systemd[1]: Started sshd@1-23.88.121.109:22-147.75.109.163:45420.service - OpenSSH per-connection server daemon (147.75.109.163:45420). Dec 13 13:21:22.722368 sshd[1795]: Accepted publickey for core from 147.75.109.163 port 45420 ssh2: RSA SHA256:mujnIWCHI6g7RhOIFUnU4L23td/nvZma7inzBKSUIRw Dec 13 13:21:22.724461 sshd-session[1795]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 13:21:22.730270 systemd-logind[1465]: New session 2 of user core. Dec 13 13:21:22.740095 systemd[1]: Started session-2.scope - Session 2 of User core. Dec 13 13:21:23.405643 sshd[1797]: Connection closed by 147.75.109.163 port 45420 Dec 13 13:21:23.406428 sshd-session[1795]: pam_unix(sshd:session): session closed for user core Dec 13 13:21:23.410592 systemd-logind[1465]: Session 2 logged out. Waiting for processes to exit. Dec 13 13:21:23.412222 systemd[1]: sshd@1-23.88.121.109:22-147.75.109.163:45420.service: Deactivated successfully. Dec 13 13:21:23.414521 systemd[1]: session-2.scope: Deactivated successfully. Dec 13 13:21:23.416041 systemd-logind[1465]: Removed session 2. Dec 13 13:21:23.584306 systemd[1]: Started sshd@2-23.88.121.109:22-147.75.109.163:45422.service - OpenSSH per-connection server daemon (147.75.109.163:45422). Dec 13 13:21:24.557225 sshd[1802]: Accepted publickey for core from 147.75.109.163 port 45422 ssh2: RSA SHA256:mujnIWCHI6g7RhOIFUnU4L23td/nvZma7inzBKSUIRw Dec 13 13:21:24.559513 sshd-session[1802]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 13:21:24.563721 systemd-logind[1465]: New session 3 of user core. Dec 13 13:21:24.576054 systemd[1]: Started session-3.scope - Session 3 of User core. Dec 13 13:21:25.227818 sshd[1804]: Connection closed by 147.75.109.163 port 45422 Dec 13 13:21:25.228570 sshd-session[1802]: pam_unix(sshd:session): session closed for user core Dec 13 13:21:25.233406 systemd[1]: sshd@2-23.88.121.109:22-147.75.109.163:45422.service: Deactivated successfully. Dec 13 13:21:25.234877 systemd[1]: session-3.scope: Deactivated successfully. Dec 13 13:21:25.236023 systemd-logind[1465]: Session 3 logged out. Waiting for processes to exit. Dec 13 13:21:25.237920 systemd-logind[1465]: Removed session 3. Dec 13 13:21:25.404048 systemd[1]: Started sshd@3-23.88.121.109:22-147.75.109.163:45432.service - OpenSSH per-connection server daemon (147.75.109.163:45432). Dec 13 13:21:26.398897 sshd[1809]: Accepted publickey for core from 147.75.109.163 port 45432 ssh2: RSA SHA256:mujnIWCHI6g7RhOIFUnU4L23td/nvZma7inzBKSUIRw Dec 13 13:21:26.400705 sshd-session[1809]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 13:21:26.406674 systemd-logind[1465]: New session 4 of user core. Dec 13 13:21:26.412337 systemd[1]: Started session-4.scope - Session 4 of User core. Dec 13 13:21:27.088651 sshd[1811]: Connection closed by 147.75.109.163 port 45432 Dec 13 13:21:27.089767 sshd-session[1809]: pam_unix(sshd:session): session closed for user core Dec 13 13:21:27.094039 systemd[1]: sshd@3-23.88.121.109:22-147.75.109.163:45432.service: Deactivated successfully. Dec 13 13:21:27.096127 systemd[1]: session-4.scope: Deactivated successfully. Dec 13 13:21:27.097257 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 11. Dec 13 13:21:27.099029 systemd-logind[1465]: Session 4 logged out. Waiting for processes to exit. Dec 13 13:21:27.105108 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 13 13:21:27.107347 systemd-logind[1465]: Removed session 4. Dec 13 13:21:27.254047 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 13 13:21:27.257870 (kubelet)[1823]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 13 13:21:27.258401 systemd[1]: Started sshd@4-23.88.121.109:22-147.75.109.163:52612.service - OpenSSH per-connection server daemon (147.75.109.163:52612). Dec 13 13:21:27.301622 kubelet[1823]: E1213 13:21:27.301560 1823 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 13 13:21:27.304519 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 13 13:21:27.304665 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 13 13:21:28.250953 sshd[1825]: Accepted publickey for core from 147.75.109.163 port 52612 ssh2: RSA SHA256:mujnIWCHI6g7RhOIFUnU4L23td/nvZma7inzBKSUIRw Dec 13 13:21:28.253645 sshd-session[1825]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 13:21:28.258939 systemd-logind[1465]: New session 5 of user core. Dec 13 13:21:28.270062 systemd[1]: Started session-5.scope - Session 5 of User core. Dec 13 13:21:28.779105 sudo[1836]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Dec 13 13:21:28.779392 sudo[1836]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Dec 13 13:21:28.796526 sudo[1836]: pam_unix(sudo:session): session closed for user root Dec 13 13:21:28.957773 sshd[1835]: Connection closed by 147.75.109.163 port 52612 Dec 13 13:21:28.959958 sshd-session[1825]: pam_unix(sshd:session): session closed for user core Dec 13 13:21:28.969309 systemd[1]: sshd@4-23.88.121.109:22-147.75.109.163:52612.service: Deactivated successfully. Dec 13 13:21:28.972655 systemd[1]: session-5.scope: Deactivated successfully. Dec 13 13:21:28.976924 systemd-logind[1465]: Session 5 logged out. Waiting for processes to exit. Dec 13 13:21:28.978902 systemd-logind[1465]: Removed session 5. Dec 13 13:21:29.136170 systemd[1]: Started sshd@5-23.88.121.109:22-147.75.109.163:52614.service - OpenSSH per-connection server daemon (147.75.109.163:52614). Dec 13 13:21:30.116598 sshd[1841]: Accepted publickey for core from 147.75.109.163 port 52614 ssh2: RSA SHA256:mujnIWCHI6g7RhOIFUnU4L23td/nvZma7inzBKSUIRw Dec 13 13:21:30.118907 sshd-session[1841]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 13:21:30.124496 systemd-logind[1465]: New session 6 of user core. Dec 13 13:21:30.136058 systemd[1]: Started session-6.scope - Session 6 of User core. Dec 13 13:21:30.636804 sudo[1845]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Dec 13 13:21:30.637127 sudo[1845]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Dec 13 13:21:30.641048 sudo[1845]: pam_unix(sudo:session): session closed for user root Dec 13 13:21:30.648054 sudo[1844]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Dec 13 13:21:30.648342 sudo[1844]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Dec 13 13:21:30.664258 systemd[1]: Starting audit-rules.service - Load Audit Rules... Dec 13 13:21:30.696756 augenrules[1867]: No rules Dec 13 13:21:30.698170 systemd[1]: audit-rules.service: Deactivated successfully. Dec 13 13:21:30.698372 systemd[1]: Finished audit-rules.service - Load Audit Rules. Dec 13 13:21:30.700010 sudo[1844]: pam_unix(sudo:session): session closed for user root Dec 13 13:21:30.858895 sshd[1843]: Connection closed by 147.75.109.163 port 52614 Dec 13 13:21:30.859457 sshd-session[1841]: pam_unix(sshd:session): session closed for user core Dec 13 13:21:30.864283 systemd[1]: sshd@5-23.88.121.109:22-147.75.109.163:52614.service: Deactivated successfully. Dec 13 13:21:30.866410 systemd[1]: session-6.scope: Deactivated successfully. Dec 13 13:21:30.867724 systemd-logind[1465]: Session 6 logged out. Waiting for processes to exit. Dec 13 13:21:30.868993 systemd-logind[1465]: Removed session 6. Dec 13 13:21:31.036493 systemd[1]: Started sshd@6-23.88.121.109:22-147.75.109.163:52630.service - OpenSSH per-connection server daemon (147.75.109.163:52630). Dec 13 13:21:32.019254 sshd[1875]: Accepted publickey for core from 147.75.109.163 port 52630 ssh2: RSA SHA256:mujnIWCHI6g7RhOIFUnU4L23td/nvZma7inzBKSUIRw Dec 13 13:21:32.021501 sshd-session[1875]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 13:21:32.027909 systemd-logind[1465]: New session 7 of user core. Dec 13 13:21:32.033057 systemd[1]: Started session-7.scope - Session 7 of User core. Dec 13 13:21:32.537786 sudo[1878]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Dec 13 13:21:32.538074 sudo[1878]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Dec 13 13:21:32.872331 systemd[1]: Starting docker.service - Docker Application Container Engine... Dec 13 13:21:32.872426 (dockerd)[1897]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Dec 13 13:21:33.096564 dockerd[1897]: time="2024-12-13T13:21:33.096482597Z" level=info msg="Starting up" Dec 13 13:21:33.205486 dockerd[1897]: time="2024-12-13T13:21:33.204594856Z" level=info msg="Loading containers: start." Dec 13 13:21:33.393815 kernel: Initializing XFRM netlink socket Dec 13 13:21:33.480603 systemd-networkd[1383]: docker0: Link UP Dec 13 13:21:33.510248 dockerd[1897]: time="2024-12-13T13:21:33.510147870Z" level=info msg="Loading containers: done." Dec 13 13:21:33.523233 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck1611083791-merged.mount: Deactivated successfully. Dec 13 13:21:33.528785 dockerd[1897]: time="2024-12-13T13:21:33.528668553Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Dec 13 13:21:33.529012 dockerd[1897]: time="2024-12-13T13:21:33.528816353Z" level=info msg="Docker daemon" commit=41ca978a0a5400cc24b274137efa9f25517fcc0b containerd-snapshotter=false storage-driver=overlay2 version=27.3.1 Dec 13 13:21:33.529062 dockerd[1897]: time="2024-12-13T13:21:33.529013913Z" level=info msg="Daemon has completed initialization" Dec 13 13:21:33.565940 dockerd[1897]: time="2024-12-13T13:21:33.565172759Z" level=info msg="API listen on /run/docker.sock" Dec 13 13:21:33.565397 systemd[1]: Started docker.service - Docker Application Container Engine. Dec 13 13:21:34.682851 containerd[1477]: time="2024-12-13T13:21:34.682809707Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.30.8\"" Dec 13 13:21:35.379141 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2634004546.mount: Deactivated successfully. Dec 13 13:21:36.266461 containerd[1477]: time="2024-12-13T13:21:36.266397352Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.30.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 13:21:36.268110 containerd[1477]: time="2024-12-13T13:21:36.267840512Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.30.8: active requests=0, bytes read=29864102" Dec 13 13:21:36.269413 containerd[1477]: time="2024-12-13T13:21:36.269375392Z" level=info msg="ImageCreate event name:\"sha256:8202e87ffef091fe4f11dd113ff6f2ab16c70279775d224ddd8aa95e2dd0b966\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 13:21:36.273534 containerd[1477]: time="2024-12-13T13:21:36.273439193Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:f0e1b3de0c2e98e6c6abd73edf9d3b8e4d44460656cde0ebb92e2d9206961fcb\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 13:21:36.275485 containerd[1477]: time="2024-12-13T13:21:36.275243513Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.30.8\" with image id \"sha256:8202e87ffef091fe4f11dd113ff6f2ab16c70279775d224ddd8aa95e2dd0b966\", repo tag \"registry.k8s.io/kube-apiserver:v1.30.8\", repo digest \"registry.k8s.io/kube-apiserver@sha256:f0e1b3de0c2e98e6c6abd73edf9d3b8e4d44460656cde0ebb92e2d9206961fcb\", size \"29860810\" in 1.592389285s" Dec 13 13:21:36.275485 containerd[1477]: time="2024-12-13T13:21:36.275307953Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.30.8\" returns image reference \"sha256:8202e87ffef091fe4f11dd113ff6f2ab16c70279775d224ddd8aa95e2dd0b966\"" Dec 13 13:21:36.300859 containerd[1477]: time="2024-12-13T13:21:36.300795037Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.30.8\"" Dec 13 13:21:37.352192 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 12. Dec 13 13:21:37.359220 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 13 13:21:37.476443 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 13 13:21:37.487194 (kubelet)[2156]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 13 13:21:37.538353 kubelet[2156]: E1213 13:21:37.538302 2156 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 13 13:21:37.541675 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 13 13:21:37.542042 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 13 13:21:38.381417 containerd[1477]: time="2024-12-13T13:21:38.381275281Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.30.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 13:21:38.383408 containerd[1477]: time="2024-12-13T13:21:38.383349362Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.30.8: active requests=0, bytes read=26900714" Dec 13 13:21:38.384445 containerd[1477]: time="2024-12-13T13:21:38.384395162Z" level=info msg="ImageCreate event name:\"sha256:4b2191aa4d4d6ca9fbd7704b35401bfa6b0b90de75db22c425053e97fd5c8338\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 13:21:38.389846 containerd[1477]: time="2024-12-13T13:21:38.389685362Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:124f66b7e877eb5a80a40503057299bb60e6a5f2130905f4e3293dabf194c397\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 13:21:38.391763 containerd[1477]: time="2024-12-13T13:21:38.391523523Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.30.8\" with image id \"sha256:4b2191aa4d4d6ca9fbd7704b35401bfa6b0b90de75db22c425053e97fd5c8338\", repo tag \"registry.k8s.io/kube-controller-manager:v1.30.8\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:124f66b7e877eb5a80a40503057299bb60e6a5f2130905f4e3293dabf194c397\", size \"28303015\" in 2.090661286s" Dec 13 13:21:38.391763 containerd[1477]: time="2024-12-13T13:21:38.391561483Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.30.8\" returns image reference \"sha256:4b2191aa4d4d6ca9fbd7704b35401bfa6b0b90de75db22c425053e97fd5c8338\"" Dec 13 13:21:38.413408 containerd[1477]: time="2024-12-13T13:21:38.413369645Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.30.8\"" Dec 13 13:21:39.982760 containerd[1477]: time="2024-12-13T13:21:39.982680797Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.30.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 13:21:39.984498 containerd[1477]: time="2024-12-13T13:21:39.984416157Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.30.8: active requests=0, bytes read=16164352" Dec 13 13:21:39.986892 containerd[1477]: time="2024-12-13T13:21:39.986801677Z" level=info msg="ImageCreate event name:\"sha256:d43326c1723208785a33cdc1507082792eb041ca0d789c103c90180e31f65ca8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 13:21:39.990994 containerd[1477]: time="2024-12-13T13:21:39.990906397Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:c8bdeac2590c99c1a77e33995423ddb6633ff90a82a2aa455442e0a8079ef8c7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 13:21:39.993282 containerd[1477]: time="2024-12-13T13:21:39.993229118Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.30.8\" with image id \"sha256:d43326c1723208785a33cdc1507082792eb041ca0d789c103c90180e31f65ca8\", repo tag \"registry.k8s.io/kube-scheduler:v1.30.8\", repo digest \"registry.k8s.io/kube-scheduler@sha256:c8bdeac2590c99c1a77e33995423ddb6633ff90a82a2aa455442e0a8079ef8c7\", size \"17566671\" in 1.579812313s" Dec 13 13:21:39.993282 containerd[1477]: time="2024-12-13T13:21:39.993275438Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.30.8\" returns image reference \"sha256:d43326c1723208785a33cdc1507082792eb041ca0d789c103c90180e31f65ca8\"" Dec 13 13:21:40.024927 containerd[1477]: time="2024-12-13T13:21:40.024882241Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.30.8\"" Dec 13 13:21:41.008660 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2922977796.mount: Deactivated successfully. Dec 13 13:21:41.573445 containerd[1477]: time="2024-12-13T13:21:41.573376838Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.30.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 13:21:41.574934 containerd[1477]: time="2024-12-13T13:21:41.574858096Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.30.8: active requests=0, bytes read=25662037" Dec 13 13:21:41.576410 containerd[1477]: time="2024-12-13T13:21:41.576344074Z" level=info msg="ImageCreate event name:\"sha256:4612aebc0675831aedbbde7cd56b85db91f1fdcf05ef923072961538ec497adb\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 13:21:41.580328 containerd[1477]: time="2024-12-13T13:21:41.580234507Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:f6d6be9417e22af78905000ac4fd134896bacd2188ea63c7cac8edd7a5d7e9b5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 13:21:41.581199 containerd[1477]: time="2024-12-13T13:21:41.581060699Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.30.8\" with image id \"sha256:4612aebc0675831aedbbde7cd56b85db91f1fdcf05ef923072961538ec497adb\", repo tag \"registry.k8s.io/kube-proxy:v1.30.8\", repo digest \"registry.k8s.io/kube-proxy@sha256:f6d6be9417e22af78905000ac4fd134896bacd2188ea63c7cac8edd7a5d7e9b5\", size \"25661030\" in 1.556135898s" Dec 13 13:21:41.581199 containerd[1477]: time="2024-12-13T13:21:41.581095901Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.30.8\" returns image reference \"sha256:4612aebc0675831aedbbde7cd56b85db91f1fdcf05ef923072961538ec497adb\"" Dec 13 13:21:41.607545 containerd[1477]: time="2024-12-13T13:21:41.607307970Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\"" Dec 13 13:21:42.179718 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount676427852.mount: Deactivated successfully. Dec 13 13:21:42.747508 containerd[1477]: time="2024-12-13T13:21:42.745495792Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 13:21:42.747508 containerd[1477]: time="2024-12-13T13:21:42.747050012Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.1: active requests=0, bytes read=16485461" Dec 13 13:21:42.747508 containerd[1477]: time="2024-12-13T13:21:42.747428026Z" level=info msg="ImageCreate event name:\"sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 13:21:42.751320 containerd[1477]: time="2024-12-13T13:21:42.751263813Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 13:21:42.752669 containerd[1477]: time="2024-12-13T13:21:42.752621624Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.1\" with image id \"sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\", size \"16482581\" in 1.145267053s" Dec 13 13:21:42.752669 containerd[1477]: time="2024-12-13T13:21:42.752663226Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\" returns image reference \"sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93\"" Dec 13 13:21:42.776251 containerd[1477]: time="2024-12-13T13:21:42.776216565Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\"" Dec 13 13:21:43.280279 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount295310180.mount: Deactivated successfully. Dec 13 13:21:43.287403 containerd[1477]: time="2024-12-13T13:21:43.287333487Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 13:21:43.289320 containerd[1477]: time="2024-12-13T13:21:43.289065391Z" level=info msg="stop pulling image registry.k8s.io/pause:3.9: active requests=0, bytes read=268841" Dec 13 13:21:43.290186 containerd[1477]: time="2024-12-13T13:21:43.290151471Z" level=info msg="ImageCreate event name:\"sha256:829e9de338bd5fdd3f16f68f83a9fb288fbc8453e881e5d5cfd0f6f2ff72b43e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 13:21:43.294726 containerd[1477]: time="2024-12-13T13:21:43.293851929Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 13:21:43.294726 containerd[1477]: time="2024-12-13T13:21:43.294589756Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.9\" with image id \"sha256:829e9de338bd5fdd3f16f68f83a9fb288fbc8453e881e5d5cfd0f6f2ff72b43e\", repo tag \"registry.k8s.io/pause:3.9\", repo digest \"registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097\", size \"268051\" in 518.153743ms" Dec 13 13:21:43.294726 containerd[1477]: time="2024-12-13T13:21:43.294634318Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\" returns image reference \"sha256:829e9de338bd5fdd3f16f68f83a9fb288fbc8453e881e5d5cfd0f6f2ff72b43e\"" Dec 13 13:21:43.318340 containerd[1477]: time="2024-12-13T13:21:43.318287875Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.12-0\"" Dec 13 13:21:43.866155 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount679775247.mount: Deactivated successfully. Dec 13 13:21:47.602139 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 13. Dec 13 13:21:47.609040 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 13 13:21:47.636651 containerd[1477]: time="2024-12-13T13:21:47.635317116Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.12-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 13:21:47.642175 containerd[1477]: time="2024-12-13T13:21:47.642080780Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.12-0: active requests=0, bytes read=66191552" Dec 13 13:21:47.650291 containerd[1477]: time="2024-12-13T13:21:47.650237651Z" level=info msg="ImageCreate event name:\"sha256:014faa467e29798aeef733fe6d1a3b5e382688217b053ad23410e6cccd5d22fd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 13:21:47.670271 containerd[1477]: time="2024-12-13T13:21:47.668844709Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:44a8e24dcbba3470ee1fee21d5e88d128c936e9b55d4bc51fbef8086f8ed123b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 13:21:47.671644 containerd[1477]: time="2024-12-13T13:21:47.671504357Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.12-0\" with image id \"sha256:014faa467e29798aeef733fe6d1a3b5e382688217b053ad23410e6cccd5d22fd\", repo tag \"registry.k8s.io/etcd:3.5.12-0\", repo digest \"registry.k8s.io/etcd@sha256:44a8e24dcbba3470ee1fee21d5e88d128c936e9b55d4bc51fbef8086f8ed123b\", size \"66189079\" in 4.352922071s" Dec 13 13:21:47.671810 containerd[1477]: time="2024-12-13T13:21:47.671789207Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.12-0\" returns image reference \"sha256:014faa467e29798aeef733fe6d1a3b5e382688217b053ad23410e6cccd5d22fd\"" Dec 13 13:21:47.737585 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 13 13:21:47.743874 (kubelet)[2312]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 13 13:21:47.808414 kubelet[2312]: E1213 13:21:47.808322 2312 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 13 13:21:47.812202 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 13 13:21:47.812520 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 13 13:21:55.097785 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Dec 13 13:21:55.112612 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 13 13:21:55.143826 systemd[1]: Reloading requested from client PID 2374 ('systemctl') (unit session-7.scope)... Dec 13 13:21:55.143843 systemd[1]: Reloading... Dec 13 13:21:55.256775 zram_generator::config[2415]: No configuration found. Dec 13 13:21:55.362733 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Dec 13 13:21:55.438496 systemd[1]: Reloading finished in 294 ms. Dec 13 13:21:55.509819 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 13 13:21:55.525948 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Dec 13 13:21:55.527392 systemd[1]: kubelet.service: Deactivated successfully. Dec 13 13:21:55.527612 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Dec 13 13:21:55.535211 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 13 13:21:55.642249 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 13 13:21:55.656341 (kubelet)[2465]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Dec 13 13:21:55.704665 kubelet[2465]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 13 13:21:55.704665 kubelet[2465]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 13 13:21:55.704665 kubelet[2465]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 13 13:21:55.705172 kubelet[2465]: I1213 13:21:55.704770 2465 server.go:205] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 13 13:21:57.070827 kubelet[2465]: I1213 13:21:57.070008 2465 server.go:484] "Kubelet version" kubeletVersion="v1.30.1" Dec 13 13:21:57.070827 kubelet[2465]: I1213 13:21:57.070042 2465 server.go:486] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 13 13:21:57.070827 kubelet[2465]: I1213 13:21:57.070253 2465 server.go:927] "Client rotation is on, will bootstrap in background" Dec 13 13:21:57.093033 kubelet[2465]: I1213 13:21:57.092664 2465 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Dec 13 13:21:57.093033 kubelet[2465]: E1213 13:21:57.093001 2465 certificate_manager.go:562] kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post "https://23.88.121.109:6443/apis/certificates.k8s.io/v1/certificatesigningrequests": dial tcp 23.88.121.109:6443: connect: connection refused Dec 13 13:21:57.101727 kubelet[2465]: I1213 13:21:57.101684 2465 server.go:742] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Dec 13 13:21:57.103801 kubelet[2465]: I1213 13:21:57.103673 2465 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 13 13:21:57.103995 kubelet[2465]: I1213 13:21:57.103756 2465 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4186-0-0-9-3a4b2091e9","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} Dec 13 13:21:57.104097 kubelet[2465]: I1213 13:21:57.104047 2465 topology_manager.go:138] "Creating topology manager with none policy" Dec 13 13:21:57.104097 kubelet[2465]: I1213 13:21:57.104059 2465 container_manager_linux.go:301] "Creating device plugin manager" Dec 13 13:21:57.104529 kubelet[2465]: I1213 13:21:57.104370 2465 state_mem.go:36] "Initialized new in-memory state store" Dec 13 13:21:57.105720 kubelet[2465]: I1213 13:21:57.105591 2465 kubelet.go:400] "Attempting to sync node with API server" Dec 13 13:21:57.105720 kubelet[2465]: I1213 13:21:57.105617 2465 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 13 13:21:57.105908 kubelet[2465]: I1213 13:21:57.105821 2465 kubelet.go:312] "Adding apiserver pod source" Dec 13 13:21:57.105908 kubelet[2465]: I1213 13:21:57.105837 2465 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 13 13:21:57.109680 kubelet[2465]: W1213 13:21:57.109567 2465 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://23.88.121.109:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 23.88.121.109:6443: connect: connection refused Dec 13 13:21:57.110783 kubelet[2465]: E1213 13:21:57.109881 2465 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://23.88.121.109:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 23.88.121.109:6443: connect: connection refused Dec 13 13:21:57.110783 kubelet[2465]: I1213 13:21:57.110097 2465 kuberuntime_manager.go:261] "Container runtime initialized" containerRuntime="containerd" version="v1.7.23" apiVersion="v1" Dec 13 13:21:57.110783 kubelet[2465]: I1213 13:21:57.110609 2465 kubelet.go:815] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 13 13:21:57.111007 kubelet[2465]: W1213 13:21:57.110733 2465 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Dec 13 13:21:57.112127 kubelet[2465]: I1213 13:21:57.112101 2465 server.go:1264] "Started kubelet" Dec 13 13:21:57.114615 kubelet[2465]: W1213 13:21:57.114566 2465 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://23.88.121.109:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4186-0-0-9-3a4b2091e9&limit=500&resourceVersion=0": dial tcp 23.88.121.109:6443: connect: connection refused Dec 13 13:21:57.114615 kubelet[2465]: E1213 13:21:57.114619 2465 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://23.88.121.109:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4186-0-0-9-3a4b2091e9&limit=500&resourceVersion=0": dial tcp 23.88.121.109:6443: connect: connection refused Dec 13 13:21:57.114771 kubelet[2465]: I1213 13:21:57.114721 2465 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 13 13:21:57.115822 kubelet[2465]: I1213 13:21:57.115801 2465 server.go:455] "Adding debug handlers to kubelet server" Dec 13 13:21:57.115978 kubelet[2465]: I1213 13:21:57.115806 2465 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 13 13:21:57.116384 kubelet[2465]: I1213 13:21:57.116365 2465 server.go:227] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 13 13:21:57.116802 kubelet[2465]: E1213 13:21:57.116610 2465 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://23.88.121.109:6443/api/v1/namespaces/default/events\": dial tcp 23.88.121.109:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4186-0-0-9-3a4b2091e9.1810bf3b1ad96965 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4186-0-0-9-3a4b2091e9,UID:ci-4186-0-0-9-3a4b2091e9,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4186-0-0-9-3a4b2091e9,},FirstTimestamp:2024-12-13 13:21:57.112072549 +0000 UTC m=+1.450344840,LastTimestamp:2024-12-13 13:21:57.112072549 +0000 UTC m=+1.450344840,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4186-0-0-9-3a4b2091e9,}" Dec 13 13:21:57.118304 kubelet[2465]: I1213 13:21:57.118277 2465 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 13 13:21:57.124929 kubelet[2465]: E1213 13:21:57.123363 2465 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"ci-4186-0-0-9-3a4b2091e9\" not found" Dec 13 13:21:57.124929 kubelet[2465]: I1213 13:21:57.123482 2465 volume_manager.go:291] "Starting Kubelet Volume Manager" Dec 13 13:21:57.124929 kubelet[2465]: I1213 13:21:57.123590 2465 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Dec 13 13:21:57.124929 kubelet[2465]: I1213 13:21:57.123668 2465 reconciler.go:26] "Reconciler: start to sync state" Dec 13 13:21:57.124929 kubelet[2465]: W1213 13:21:57.124089 2465 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://23.88.121.109:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 23.88.121.109:6443: connect: connection refused Dec 13 13:21:57.124929 kubelet[2465]: E1213 13:21:57.124138 2465 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://23.88.121.109:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 23.88.121.109:6443: connect: connection refused Dec 13 13:21:57.124929 kubelet[2465]: E1213 13:21:57.124305 2465 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://23.88.121.109:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4186-0-0-9-3a4b2091e9?timeout=10s\": dial tcp 23.88.121.109:6443: connect: connection refused" interval="200ms" Dec 13 13:21:57.125460 kubelet[2465]: E1213 13:21:57.125435 2465 kubelet.go:1467] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Dec 13 13:21:57.125704 kubelet[2465]: I1213 13:21:57.125681 2465 factory.go:221] Registration of the systemd container factory successfully Dec 13 13:21:57.125958 kubelet[2465]: I1213 13:21:57.125933 2465 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Dec 13 13:21:57.129453 kubelet[2465]: I1213 13:21:57.129393 2465 factory.go:221] Registration of the containerd container factory successfully Dec 13 13:21:57.141785 kubelet[2465]: I1213 13:21:57.141618 2465 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 13 13:21:57.143466 kubelet[2465]: I1213 13:21:57.143036 2465 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 13 13:21:57.143466 kubelet[2465]: I1213 13:21:57.143077 2465 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 13 13:21:57.143466 kubelet[2465]: I1213 13:21:57.143101 2465 kubelet.go:2337] "Starting kubelet main sync loop" Dec 13 13:21:57.143466 kubelet[2465]: E1213 13:21:57.143167 2465 kubelet.go:2361] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 13 13:21:57.152076 kubelet[2465]: W1213 13:21:57.152009 2465 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://23.88.121.109:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 23.88.121.109:6443: connect: connection refused Dec 13 13:21:57.152275 kubelet[2465]: E1213 13:21:57.152258 2465 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://23.88.121.109:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 23.88.121.109:6443: connect: connection refused Dec 13 13:21:57.162547 kubelet[2465]: I1213 13:21:57.162517 2465 cpu_manager.go:214] "Starting CPU manager" policy="none" Dec 13 13:21:57.162547 kubelet[2465]: I1213 13:21:57.162536 2465 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Dec 13 13:21:57.162730 kubelet[2465]: I1213 13:21:57.162561 2465 state_mem.go:36] "Initialized new in-memory state store" Dec 13 13:21:57.166520 kubelet[2465]: I1213 13:21:57.166487 2465 policy_none.go:49] "None policy: Start" Dec 13 13:21:57.167732 kubelet[2465]: I1213 13:21:57.167648 2465 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 13 13:21:57.167835 kubelet[2465]: I1213 13:21:57.167764 2465 state_mem.go:35] "Initializing new in-memory state store" Dec 13 13:21:57.175313 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Dec 13 13:21:57.196237 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Dec 13 13:21:57.201461 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Dec 13 13:21:57.211760 kubelet[2465]: I1213 13:21:57.211625 2465 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 13 13:21:57.212543 kubelet[2465]: I1213 13:21:57.212347 2465 container_log_manager.go:186] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 13 13:21:57.213875 kubelet[2465]: I1213 13:21:57.213434 2465 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 13 13:21:57.217230 kubelet[2465]: E1213 13:21:57.217203 2465 eviction_manager.go:282] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4186-0-0-9-3a4b2091e9\" not found" Dec 13 13:21:57.226769 kubelet[2465]: I1213 13:21:57.226499 2465 kubelet_node_status.go:73] "Attempting to register node" node="ci-4186-0-0-9-3a4b2091e9" Dec 13 13:21:57.227319 kubelet[2465]: E1213 13:21:57.227278 2465 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://23.88.121.109:6443/api/v1/nodes\": dial tcp 23.88.121.109:6443: connect: connection refused" node="ci-4186-0-0-9-3a4b2091e9" Dec 13 13:21:57.244371 kubelet[2465]: I1213 13:21:57.244290 2465 topology_manager.go:215] "Topology Admit Handler" podUID="fefcd6b7a40abae3f206c49cecce6d10" podNamespace="kube-system" podName="kube-apiserver-ci-4186-0-0-9-3a4b2091e9" Dec 13 13:21:57.246782 kubelet[2465]: I1213 13:21:57.246597 2465 topology_manager.go:215] "Topology Admit Handler" podUID="aad4b9b343b480895f715fb9ea17da19" podNamespace="kube-system" podName="kube-controller-manager-ci-4186-0-0-9-3a4b2091e9" Dec 13 13:21:57.250245 kubelet[2465]: I1213 13:21:57.249907 2465 topology_manager.go:215] "Topology Admit Handler" podUID="5daf0f4e60133131538ff6808845b16e" podNamespace="kube-system" podName="kube-scheduler-ci-4186-0-0-9-3a4b2091e9" Dec 13 13:21:57.260451 systemd[1]: Created slice kubepods-burstable-podfefcd6b7a40abae3f206c49cecce6d10.slice - libcontainer container kubepods-burstable-podfefcd6b7a40abae3f206c49cecce6d10.slice. Dec 13 13:21:57.284137 systemd[1]: Created slice kubepods-burstable-podaad4b9b343b480895f715fb9ea17da19.slice - libcontainer container kubepods-burstable-podaad4b9b343b480895f715fb9ea17da19.slice. Dec 13 13:21:57.297615 systemd[1]: Created slice kubepods-burstable-pod5daf0f4e60133131538ff6808845b16e.slice - libcontainer container kubepods-burstable-pod5daf0f4e60133131538ff6808845b16e.slice. Dec 13 13:21:57.324952 kubelet[2465]: E1213 13:21:57.324814 2465 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://23.88.121.109:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4186-0-0-9-3a4b2091e9?timeout=10s\": dial tcp 23.88.121.109:6443: connect: connection refused" interval="400ms" Dec 13 13:21:57.424776 kubelet[2465]: I1213 13:21:57.424312 2465 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/aad4b9b343b480895f715fb9ea17da19-flexvolume-dir\") pod \"kube-controller-manager-ci-4186-0-0-9-3a4b2091e9\" (UID: \"aad4b9b343b480895f715fb9ea17da19\") " pod="kube-system/kube-controller-manager-ci-4186-0-0-9-3a4b2091e9" Dec 13 13:21:57.424776 kubelet[2465]: I1213 13:21:57.424530 2465 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/aad4b9b343b480895f715fb9ea17da19-kubeconfig\") pod \"kube-controller-manager-ci-4186-0-0-9-3a4b2091e9\" (UID: \"aad4b9b343b480895f715fb9ea17da19\") " pod="kube-system/kube-controller-manager-ci-4186-0-0-9-3a4b2091e9" Dec 13 13:21:57.425243 kubelet[2465]: I1213 13:21:57.425217 2465 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/aad4b9b343b480895f715fb9ea17da19-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4186-0-0-9-3a4b2091e9\" (UID: \"aad4b9b343b480895f715fb9ea17da19\") " pod="kube-system/kube-controller-manager-ci-4186-0-0-9-3a4b2091e9" Dec 13 13:21:57.425390 kubelet[2465]: I1213 13:21:57.425252 2465 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/fefcd6b7a40abae3f206c49cecce6d10-k8s-certs\") pod \"kube-apiserver-ci-4186-0-0-9-3a4b2091e9\" (UID: \"fefcd6b7a40abae3f206c49cecce6d10\") " pod="kube-system/kube-apiserver-ci-4186-0-0-9-3a4b2091e9" Dec 13 13:21:57.425390 kubelet[2465]: I1213 13:21:57.425276 2465 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/fefcd6b7a40abae3f206c49cecce6d10-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4186-0-0-9-3a4b2091e9\" (UID: \"fefcd6b7a40abae3f206c49cecce6d10\") " pod="kube-system/kube-apiserver-ci-4186-0-0-9-3a4b2091e9" Dec 13 13:21:57.425390 kubelet[2465]: I1213 13:21:57.425295 2465 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/aad4b9b343b480895f715fb9ea17da19-k8s-certs\") pod \"kube-controller-manager-ci-4186-0-0-9-3a4b2091e9\" (UID: \"aad4b9b343b480895f715fb9ea17da19\") " pod="kube-system/kube-controller-manager-ci-4186-0-0-9-3a4b2091e9" Dec 13 13:21:57.425390 kubelet[2465]: I1213 13:21:57.425314 2465 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/5daf0f4e60133131538ff6808845b16e-kubeconfig\") pod \"kube-scheduler-ci-4186-0-0-9-3a4b2091e9\" (UID: \"5daf0f4e60133131538ff6808845b16e\") " pod="kube-system/kube-scheduler-ci-4186-0-0-9-3a4b2091e9" Dec 13 13:21:57.425390 kubelet[2465]: I1213 13:21:57.425346 2465 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/fefcd6b7a40abae3f206c49cecce6d10-ca-certs\") pod \"kube-apiserver-ci-4186-0-0-9-3a4b2091e9\" (UID: \"fefcd6b7a40abae3f206c49cecce6d10\") " pod="kube-system/kube-apiserver-ci-4186-0-0-9-3a4b2091e9" Dec 13 13:21:57.425609 kubelet[2465]: I1213 13:21:57.425368 2465 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/aad4b9b343b480895f715fb9ea17da19-ca-certs\") pod \"kube-controller-manager-ci-4186-0-0-9-3a4b2091e9\" (UID: \"aad4b9b343b480895f715fb9ea17da19\") " pod="kube-system/kube-controller-manager-ci-4186-0-0-9-3a4b2091e9" Dec 13 13:21:57.430605 kubelet[2465]: I1213 13:21:57.430578 2465 kubelet_node_status.go:73] "Attempting to register node" node="ci-4186-0-0-9-3a4b2091e9" Dec 13 13:21:57.431040 kubelet[2465]: E1213 13:21:57.431012 2465 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://23.88.121.109:6443/api/v1/nodes\": dial tcp 23.88.121.109:6443: connect: connection refused" node="ci-4186-0-0-9-3a4b2091e9" Dec 13 13:21:57.582996 containerd[1477]: time="2024-12-13T13:21:57.582822205Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4186-0-0-9-3a4b2091e9,Uid:fefcd6b7a40abae3f206c49cecce6d10,Namespace:kube-system,Attempt:0,}" Dec 13 13:21:57.595295 containerd[1477]: time="2024-12-13T13:21:57.595210639Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4186-0-0-9-3a4b2091e9,Uid:aad4b9b343b480895f715fb9ea17da19,Namespace:kube-system,Attempt:0,}" Dec 13 13:21:57.601690 containerd[1477]: time="2024-12-13T13:21:57.601268433Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4186-0-0-9-3a4b2091e9,Uid:5daf0f4e60133131538ff6808845b16e,Namespace:kube-system,Attempt:0,}" Dec 13 13:21:57.725774 kubelet[2465]: E1213 13:21:57.725687 2465 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://23.88.121.109:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4186-0-0-9-3a4b2091e9?timeout=10s\": dial tcp 23.88.121.109:6443: connect: connection refused" interval="800ms" Dec 13 13:21:57.834988 kubelet[2465]: I1213 13:21:57.834715 2465 kubelet_node_status.go:73] "Attempting to register node" node="ci-4186-0-0-9-3a4b2091e9" Dec 13 13:21:57.835420 kubelet[2465]: E1213 13:21:57.835375 2465 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://23.88.121.109:6443/api/v1/nodes\": dial tcp 23.88.121.109:6443: connect: connection refused" node="ci-4186-0-0-9-3a4b2091e9" Dec 13 13:21:58.004127 kubelet[2465]: W1213 13:21:58.003992 2465 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://23.88.121.109:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 23.88.121.109:6443: connect: connection refused Dec 13 13:21:58.004127 kubelet[2465]: E1213 13:21:58.004091 2465 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://23.88.121.109:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 23.88.121.109:6443: connect: connection refused Dec 13 13:21:58.006826 kubelet[2465]: W1213 13:21:58.006664 2465 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://23.88.121.109:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 23.88.121.109:6443: connect: connection refused Dec 13 13:21:58.006826 kubelet[2465]: E1213 13:21:58.006711 2465 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://23.88.121.109:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 23.88.121.109:6443: connect: connection refused Dec 13 13:21:58.100721 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount144438253.mount: Deactivated successfully. Dec 13 13:21:58.108881 containerd[1477]: time="2024-12-13T13:21:58.108815752Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Dec 13 13:21:58.113455 containerd[1477]: time="2024-12-13T13:21:58.113022495Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=269193" Dec 13 13:21:58.115976 containerd[1477]: time="2024-12-13T13:21:58.115816244Z" level=info msg="ImageCreate event name:\"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Dec 13 13:21:58.118275 containerd[1477]: time="2024-12-13T13:21:58.117984618Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Dec 13 13:21:58.118984 containerd[1477]: time="2024-12-13T13:21:58.118922721Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Dec 13 13:21:58.120159 containerd[1477]: time="2024-12-13T13:21:58.120120151Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Dec 13 13:21:58.121101 containerd[1477]: time="2024-12-13T13:21:58.121073174Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Dec 13 13:21:58.122939 containerd[1477]: time="2024-12-13T13:21:58.122149081Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Dec 13 13:21:58.124185 containerd[1477]: time="2024-12-13T13:21:58.124152850Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 541.182322ms" Dec 13 13:21:58.126781 containerd[1477]: time="2024-12-13T13:21:58.126747074Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 531.383032ms" Dec 13 13:21:58.130522 containerd[1477]: time="2024-12-13T13:21:58.130372364Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 528.989529ms" Dec 13 13:21:58.165572 kubelet[2465]: W1213 13:21:58.164973 2465 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://23.88.121.109:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 23.88.121.109:6443: connect: connection refused Dec 13 13:21:58.165572 kubelet[2465]: E1213 13:21:58.165040 2465 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://23.88.121.109:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 23.88.121.109:6443: connect: connection refused Dec 13 13:21:58.252167 containerd[1477]: time="2024-12-13T13:21:58.251889485Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 13:21:58.252167 containerd[1477]: time="2024-12-13T13:21:58.252027049Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 13:21:58.252167 containerd[1477]: time="2024-12-13T13:21:58.252044329Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 13:21:58.252362 containerd[1477]: time="2024-12-13T13:21:58.252181452Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 13:21:58.256371 containerd[1477]: time="2024-12-13T13:21:58.256210712Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 13:21:58.256586 containerd[1477]: time="2024-12-13T13:21:58.256542760Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 13:21:58.258858 containerd[1477]: time="2024-12-13T13:21:58.257540705Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 13:21:58.258858 containerd[1477]: time="2024-12-13T13:21:58.258170440Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 13:21:58.258858 containerd[1477]: time="2024-12-13T13:21:58.257665228Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 13:21:58.258858 containerd[1477]: time="2024-12-13T13:21:58.257722029Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 13:21:58.258858 containerd[1477]: time="2024-12-13T13:21:58.257748910Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 13:21:58.258858 containerd[1477]: time="2024-12-13T13:21:58.257844992Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 13:21:58.284993 systemd[1]: Started cri-containerd-1bc350dd4920e74bcd0be073eb6ed7a7680cd1170253bccc828a177b4fcaf343.scope - libcontainer container 1bc350dd4920e74bcd0be073eb6ed7a7680cd1170253bccc828a177b4fcaf343. Dec 13 13:21:58.290463 systemd[1]: Started cri-containerd-96ebce14f7d74ca74001f6124edd8d49b93f4b4732c13760badec1bbbcd39f9a.scope - libcontainer container 96ebce14f7d74ca74001f6124edd8d49b93f4b4732c13760badec1bbbcd39f9a. Dec 13 13:21:58.292205 systemd[1]: Started cri-containerd-ab08713986e5c1552494e0cabb6701001c06bd11230be29a1b60962de02fa05f.scope - libcontainer container ab08713986e5c1552494e0cabb6701001c06bd11230be29a1b60962de02fa05f. Dec 13 13:21:58.342755 containerd[1477]: time="2024-12-13T13:21:58.342627766Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4186-0-0-9-3a4b2091e9,Uid:aad4b9b343b480895f715fb9ea17da19,Namespace:kube-system,Attempt:0,} returns sandbox id \"96ebce14f7d74ca74001f6124edd8d49b93f4b4732c13760badec1bbbcd39f9a\"" Dec 13 13:21:58.350544 containerd[1477]: time="2024-12-13T13:21:58.349577418Z" level=info msg="CreateContainer within sandbox \"96ebce14f7d74ca74001f6124edd8d49b93f4b4732c13760badec1bbbcd39f9a\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Dec 13 13:21:58.354825 containerd[1477]: time="2024-12-13T13:21:58.354783547Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4186-0-0-9-3a4b2091e9,Uid:fefcd6b7a40abae3f206c49cecce6d10,Namespace:kube-system,Attempt:0,} returns sandbox id \"1bc350dd4920e74bcd0be073eb6ed7a7680cd1170253bccc828a177b4fcaf343\"" Dec 13 13:21:58.358119 containerd[1477]: time="2024-12-13T13:21:58.358069548Z" level=info msg="CreateContainer within sandbox \"1bc350dd4920e74bcd0be073eb6ed7a7680cd1170253bccc828a177b4fcaf343\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Dec 13 13:21:58.365362 containerd[1477]: time="2024-12-13T13:21:58.365312567Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4186-0-0-9-3a4b2091e9,Uid:5daf0f4e60133131538ff6808845b16e,Namespace:kube-system,Attempt:0,} returns sandbox id \"ab08713986e5c1552494e0cabb6701001c06bd11230be29a1b60962de02fa05f\"" Dec 13 13:21:58.369549 containerd[1477]: time="2024-12-13T13:21:58.369482430Z" level=info msg="CreateContainer within sandbox \"ab08713986e5c1552494e0cabb6701001c06bd11230be29a1b60962de02fa05f\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Dec 13 13:21:58.381481 containerd[1477]: time="2024-12-13T13:21:58.381421685Z" level=info msg="CreateContainer within sandbox \"96ebce14f7d74ca74001f6124edd8d49b93f4b4732c13760badec1bbbcd39f9a\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"8b6a6b55b44aa0a5f3080d85d9c17b2c54270376d226859be2a7c72c6659adcb\"" Dec 13 13:21:58.382311 containerd[1477]: time="2024-12-13T13:21:58.382209944Z" level=info msg="CreateContainer within sandbox \"1bc350dd4920e74bcd0be073eb6ed7a7680cd1170253bccc828a177b4fcaf343\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"22615ee618af8c529a88f514173c2f238a62674449091c777c30586696e95203\"" Dec 13 13:21:58.382515 containerd[1477]: time="2024-12-13T13:21:58.382493671Z" level=info msg="StartContainer for \"8b6a6b55b44aa0a5f3080d85d9c17b2c54270376d226859be2a7c72c6659adcb\"" Dec 13 13:21:58.385176 containerd[1477]: time="2024-12-13T13:21:58.385150537Z" level=info msg="StartContainer for \"22615ee618af8c529a88f514173c2f238a62674449091c777c30586696e95203\"" Dec 13 13:21:58.394088 containerd[1477]: time="2024-12-13T13:21:58.394041076Z" level=info msg="CreateContainer within sandbox \"ab08713986e5c1552494e0cabb6701001c06bd11230be29a1b60962de02fa05f\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"afeae58b5d0f4f03848f198a7146faa4a0346103434f00c53c3f77deb38561b1\"" Dec 13 13:21:58.394605 containerd[1477]: time="2024-12-13T13:21:58.394576770Z" level=info msg="StartContainer for \"afeae58b5d0f4f03848f198a7146faa4a0346103434f00c53c3f77deb38561b1\"" Dec 13 13:21:58.421407 systemd[1]: Started cri-containerd-8b6a6b55b44aa0a5f3080d85d9c17b2c54270376d226859be2a7c72c6659adcb.scope - libcontainer container 8b6a6b55b44aa0a5f3080d85d9c17b2c54270376d226859be2a7c72c6659adcb. Dec 13 13:21:58.429220 systemd[1]: Started cri-containerd-afeae58b5d0f4f03848f198a7146faa4a0346103434f00c53c3f77deb38561b1.scope - libcontainer container afeae58b5d0f4f03848f198a7146faa4a0346103434f00c53c3f77deb38561b1. Dec 13 13:21:58.441155 systemd[1]: Started cri-containerd-22615ee618af8c529a88f514173c2f238a62674449091c777c30586696e95203.scope - libcontainer container 22615ee618af8c529a88f514173c2f238a62674449091c777c30586696e95203. Dec 13 13:21:58.504647 containerd[1477]: time="2024-12-13T13:21:58.504454683Z" level=info msg="StartContainer for \"22615ee618af8c529a88f514173c2f238a62674449091c777c30586696e95203\" returns successfully" Dec 13 13:21:58.504647 containerd[1477]: time="2024-12-13T13:21:58.504588847Z" level=info msg="StartContainer for \"8b6a6b55b44aa0a5f3080d85d9c17b2c54270376d226859be2a7c72c6659adcb\" returns successfully" Dec 13 13:21:58.509724 containerd[1477]: time="2024-12-13T13:21:58.509646692Z" level=info msg="StartContainer for \"afeae58b5d0f4f03848f198a7146faa4a0346103434f00c53c3f77deb38561b1\" returns successfully" Dec 13 13:21:58.526761 kubelet[2465]: E1213 13:21:58.526684 2465 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://23.88.121.109:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4186-0-0-9-3a4b2091e9?timeout=10s\": dial tcp 23.88.121.109:6443: connect: connection refused" interval="1.6s" Dec 13 13:21:58.599629 kubelet[2465]: W1213 13:21:58.599539 2465 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://23.88.121.109:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4186-0-0-9-3a4b2091e9&limit=500&resourceVersion=0": dial tcp 23.88.121.109:6443: connect: connection refused Dec 13 13:21:58.599629 kubelet[2465]: E1213 13:21:58.599606 2465 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://23.88.121.109:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4186-0-0-9-3a4b2091e9&limit=500&resourceVersion=0": dial tcp 23.88.121.109:6443: connect: connection refused Dec 13 13:21:58.639156 kubelet[2465]: I1213 13:21:58.639059 2465 kubelet_node_status.go:73] "Attempting to register node" node="ci-4186-0-0-9-3a4b2091e9" Dec 13 13:22:00.610768 kubelet[2465]: E1213 13:22:00.610694 2465 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-4186-0-0-9-3a4b2091e9\" not found" node="ci-4186-0-0-9-3a4b2091e9" Dec 13 13:22:00.694874 kubelet[2465]: I1213 13:22:00.693826 2465 kubelet_node_status.go:76] "Successfully registered node" node="ci-4186-0-0-9-3a4b2091e9" Dec 13 13:22:01.110134 kubelet[2465]: I1213 13:22:01.110093 2465 apiserver.go:52] "Watching apiserver" Dec 13 13:22:01.124691 kubelet[2465]: I1213 13:22:01.124636 2465 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" Dec 13 13:22:03.095230 systemd[1]: Reloading requested from client PID 2738 ('systemctl') (unit session-7.scope)... Dec 13 13:22:03.095255 systemd[1]: Reloading... Dec 13 13:22:03.215772 zram_generator::config[2778]: No configuration found. Dec 13 13:22:03.317701 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Dec 13 13:22:03.398353 systemd[1]: Reloading finished in 302 ms. Dec 13 13:22:03.440932 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Dec 13 13:22:03.457407 systemd[1]: kubelet.service: Deactivated successfully. Dec 13 13:22:03.457873 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Dec 13 13:22:03.457991 systemd[1]: kubelet.service: Consumed 1.889s CPU time, 113.4M memory peak, 0B memory swap peak. Dec 13 13:22:03.465219 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 13 13:22:03.574247 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 13 13:22:03.588289 (kubelet)[2824]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Dec 13 13:22:03.646394 kubelet[2824]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 13 13:22:03.646394 kubelet[2824]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 13 13:22:03.646394 kubelet[2824]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 13 13:22:03.646777 kubelet[2824]: I1213 13:22:03.646470 2824 server.go:205] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 13 13:22:03.651795 kubelet[2824]: I1213 13:22:03.651654 2824 server.go:484] "Kubelet version" kubeletVersion="v1.30.1" Dec 13 13:22:03.651795 kubelet[2824]: I1213 13:22:03.651681 2824 server.go:486] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 13 13:22:03.652340 kubelet[2824]: I1213 13:22:03.652317 2824 server.go:927] "Client rotation is on, will bootstrap in background" Dec 13 13:22:03.653688 kubelet[2824]: I1213 13:22:03.653656 2824 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 13 13:22:03.657326 kubelet[2824]: I1213 13:22:03.656191 2824 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Dec 13 13:22:03.664164 kubelet[2824]: I1213 13:22:03.663702 2824 server.go:742] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Dec 13 13:22:03.664164 kubelet[2824]: I1213 13:22:03.663924 2824 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 13 13:22:03.664388 kubelet[2824]: I1213 13:22:03.663950 2824 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4186-0-0-9-3a4b2091e9","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} Dec 13 13:22:03.664540 kubelet[2824]: I1213 13:22:03.664526 2824 topology_manager.go:138] "Creating topology manager with none policy" Dec 13 13:22:03.664596 kubelet[2824]: I1213 13:22:03.664588 2824 container_manager_linux.go:301] "Creating device plugin manager" Dec 13 13:22:03.664680 kubelet[2824]: I1213 13:22:03.664671 2824 state_mem.go:36] "Initialized new in-memory state store" Dec 13 13:22:03.664911 kubelet[2824]: I1213 13:22:03.664895 2824 kubelet.go:400] "Attempting to sync node with API server" Dec 13 13:22:03.665012 kubelet[2824]: I1213 13:22:03.664999 2824 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 13 13:22:03.665094 kubelet[2824]: I1213 13:22:03.665085 2824 kubelet.go:312] "Adding apiserver pod source" Dec 13 13:22:03.665153 kubelet[2824]: I1213 13:22:03.665145 2824 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 13 13:22:03.668054 kubelet[2824]: I1213 13:22:03.667837 2824 kuberuntime_manager.go:261] "Container runtime initialized" containerRuntime="containerd" version="v1.7.23" apiVersion="v1" Dec 13 13:22:03.668054 kubelet[2824]: I1213 13:22:03.668024 2824 kubelet.go:815] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 13 13:22:03.668659 kubelet[2824]: I1213 13:22:03.668484 2824 server.go:1264] "Started kubelet" Dec 13 13:22:03.671902 kubelet[2824]: I1213 13:22:03.670444 2824 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 13 13:22:03.675704 kubelet[2824]: I1213 13:22:03.675265 2824 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 13 13:22:03.677169 kubelet[2824]: I1213 13:22:03.676426 2824 server.go:455] "Adding debug handlers to kubelet server" Dec 13 13:22:03.677492 kubelet[2824]: I1213 13:22:03.677427 2824 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 13 13:22:03.679746 kubelet[2824]: I1213 13:22:03.677643 2824 server.go:227] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 13 13:22:03.679746 kubelet[2824]: I1213 13:22:03.679388 2824 volume_manager.go:291] "Starting Kubelet Volume Manager" Dec 13 13:22:03.682723 kubelet[2824]: I1213 13:22:03.681312 2824 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Dec 13 13:22:03.682723 kubelet[2824]: I1213 13:22:03.681479 2824 reconciler.go:26] "Reconciler: start to sync state" Dec 13 13:22:03.683264 kubelet[2824]: I1213 13:22:03.683214 2824 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 13 13:22:03.685406 kubelet[2824]: I1213 13:22:03.684395 2824 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 13 13:22:03.685406 kubelet[2824]: I1213 13:22:03.684436 2824 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 13 13:22:03.685406 kubelet[2824]: I1213 13:22:03.684457 2824 kubelet.go:2337] "Starting kubelet main sync loop" Dec 13 13:22:03.685406 kubelet[2824]: E1213 13:22:03.684498 2824 kubelet.go:2361] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 13 13:22:03.699756 kubelet[2824]: I1213 13:22:03.699034 2824 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Dec 13 13:22:03.710749 kubelet[2824]: I1213 13:22:03.709359 2824 factory.go:221] Registration of the containerd container factory successfully Dec 13 13:22:03.710749 kubelet[2824]: I1213 13:22:03.709393 2824 factory.go:221] Registration of the systemd container factory successfully Dec 13 13:22:03.782659 kubelet[2824]: I1213 13:22:03.782622 2824 kubelet_node_status.go:73] "Attempting to register node" node="ci-4186-0-0-9-3a4b2091e9" Dec 13 13:22:03.785784 kubelet[2824]: E1213 13:22:03.785703 2824 kubelet.go:2361] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Dec 13 13:22:03.786786 kubelet[2824]: I1213 13:22:03.786754 2824 cpu_manager.go:214] "Starting CPU manager" policy="none" Dec 13 13:22:03.786901 kubelet[2824]: I1213 13:22:03.786887 2824 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Dec 13 13:22:03.786962 kubelet[2824]: I1213 13:22:03.786954 2824 state_mem.go:36] "Initialized new in-memory state store" Dec 13 13:22:03.787308 kubelet[2824]: I1213 13:22:03.787213 2824 state_mem.go:88] "Updated default CPUSet" cpuSet="" Dec 13 13:22:03.787308 kubelet[2824]: I1213 13:22:03.787231 2824 state_mem.go:96] "Updated CPUSet assignments" assignments={} Dec 13 13:22:03.787308 kubelet[2824]: I1213 13:22:03.787250 2824 policy_none.go:49] "None policy: Start" Dec 13 13:22:03.789894 kubelet[2824]: I1213 13:22:03.788908 2824 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 13 13:22:03.789894 kubelet[2824]: I1213 13:22:03.788938 2824 state_mem.go:35] "Initializing new in-memory state store" Dec 13 13:22:03.789894 kubelet[2824]: I1213 13:22:03.789107 2824 state_mem.go:75] "Updated machine memory state" Dec 13 13:22:03.794715 kubelet[2824]: I1213 13:22:03.794687 2824 kubelet_node_status.go:112] "Node was previously registered" node="ci-4186-0-0-9-3a4b2091e9" Dec 13 13:22:03.795164 kubelet[2824]: I1213 13:22:03.795148 2824 kubelet_node_status.go:76] "Successfully registered node" node="ci-4186-0-0-9-3a4b2091e9" Dec 13 13:22:03.802417 kubelet[2824]: I1213 13:22:03.802387 2824 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 13 13:22:03.805814 kubelet[2824]: I1213 13:22:03.805489 2824 container_log_manager.go:186] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 13 13:22:03.805814 kubelet[2824]: I1213 13:22:03.805595 2824 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 13 13:22:03.987251 kubelet[2824]: I1213 13:22:03.987019 2824 topology_manager.go:215] "Topology Admit Handler" podUID="aad4b9b343b480895f715fb9ea17da19" podNamespace="kube-system" podName="kube-controller-manager-ci-4186-0-0-9-3a4b2091e9" Dec 13 13:22:03.987251 kubelet[2824]: I1213 13:22:03.987192 2824 topology_manager.go:215] "Topology Admit Handler" podUID="5daf0f4e60133131538ff6808845b16e" podNamespace="kube-system" podName="kube-scheduler-ci-4186-0-0-9-3a4b2091e9" Dec 13 13:22:03.987468 kubelet[2824]: I1213 13:22:03.987276 2824 topology_manager.go:215] "Topology Admit Handler" podUID="fefcd6b7a40abae3f206c49cecce6d10" podNamespace="kube-system" podName="kube-apiserver-ci-4186-0-0-9-3a4b2091e9" Dec 13 13:22:03.999553 kubelet[2824]: E1213 13:22:03.999221 2824 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-apiserver-ci-4186-0-0-9-3a4b2091e9\" already exists" pod="kube-system/kube-apiserver-ci-4186-0-0-9-3a4b2091e9" Dec 13 13:22:04.000281 kubelet[2824]: E1213 13:22:04.000245 2824 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-ci-4186-0-0-9-3a4b2091e9\" already exists" pod="kube-system/kube-controller-manager-ci-4186-0-0-9-3a4b2091e9" Dec 13 13:22:04.083610 kubelet[2824]: I1213 13:22:04.083569 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/5daf0f4e60133131538ff6808845b16e-kubeconfig\") pod \"kube-scheduler-ci-4186-0-0-9-3a4b2091e9\" (UID: \"5daf0f4e60133131538ff6808845b16e\") " pod="kube-system/kube-scheduler-ci-4186-0-0-9-3a4b2091e9" Dec 13 13:22:04.083856 kubelet[2824]: I1213 13:22:04.083658 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/fefcd6b7a40abae3f206c49cecce6d10-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4186-0-0-9-3a4b2091e9\" (UID: \"fefcd6b7a40abae3f206c49cecce6d10\") " pod="kube-system/kube-apiserver-ci-4186-0-0-9-3a4b2091e9" Dec 13 13:22:04.083856 kubelet[2824]: I1213 13:22:04.083696 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/aad4b9b343b480895f715fb9ea17da19-ca-certs\") pod \"kube-controller-manager-ci-4186-0-0-9-3a4b2091e9\" (UID: \"aad4b9b343b480895f715fb9ea17da19\") " pod="kube-system/kube-controller-manager-ci-4186-0-0-9-3a4b2091e9" Dec 13 13:22:04.083856 kubelet[2824]: I1213 13:22:04.083716 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/aad4b9b343b480895f715fb9ea17da19-k8s-certs\") pod \"kube-controller-manager-ci-4186-0-0-9-3a4b2091e9\" (UID: \"aad4b9b343b480895f715fb9ea17da19\") " pod="kube-system/kube-controller-manager-ci-4186-0-0-9-3a4b2091e9" Dec 13 13:22:04.083856 kubelet[2824]: I1213 13:22:04.083733 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/aad4b9b343b480895f715fb9ea17da19-kubeconfig\") pod \"kube-controller-manager-ci-4186-0-0-9-3a4b2091e9\" (UID: \"aad4b9b343b480895f715fb9ea17da19\") " pod="kube-system/kube-controller-manager-ci-4186-0-0-9-3a4b2091e9" Dec 13 13:22:04.083856 kubelet[2824]: I1213 13:22:04.083772 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/aad4b9b343b480895f715fb9ea17da19-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4186-0-0-9-3a4b2091e9\" (UID: \"aad4b9b343b480895f715fb9ea17da19\") " pod="kube-system/kube-controller-manager-ci-4186-0-0-9-3a4b2091e9" Dec 13 13:22:04.084376 kubelet[2824]: I1213 13:22:04.083797 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/fefcd6b7a40abae3f206c49cecce6d10-ca-certs\") pod \"kube-apiserver-ci-4186-0-0-9-3a4b2091e9\" (UID: \"fefcd6b7a40abae3f206c49cecce6d10\") " pod="kube-system/kube-apiserver-ci-4186-0-0-9-3a4b2091e9" Dec 13 13:22:04.084376 kubelet[2824]: I1213 13:22:04.083831 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/fefcd6b7a40abae3f206c49cecce6d10-k8s-certs\") pod \"kube-apiserver-ci-4186-0-0-9-3a4b2091e9\" (UID: \"fefcd6b7a40abae3f206c49cecce6d10\") " pod="kube-system/kube-apiserver-ci-4186-0-0-9-3a4b2091e9" Dec 13 13:22:04.084376 kubelet[2824]: I1213 13:22:04.083853 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/aad4b9b343b480895f715fb9ea17da19-flexvolume-dir\") pod \"kube-controller-manager-ci-4186-0-0-9-3a4b2091e9\" (UID: \"aad4b9b343b480895f715fb9ea17da19\") " pod="kube-system/kube-controller-manager-ci-4186-0-0-9-3a4b2091e9" Dec 13 13:22:04.092094 sudo[2857]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Dec 13 13:22:04.092425 sudo[2857]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=0) Dec 13 13:22:04.601722 sudo[2857]: pam_unix(sudo:session): session closed for user root Dec 13 13:22:04.666769 kubelet[2824]: I1213 13:22:04.666479 2824 apiserver.go:52] "Watching apiserver" Dec 13 13:22:04.681614 kubelet[2824]: I1213 13:22:04.681562 2824 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" Dec 13 13:22:04.775066 kubelet[2824]: E1213 13:22:04.775025 2824 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-apiserver-ci-4186-0-0-9-3a4b2091e9\" already exists" pod="kube-system/kube-apiserver-ci-4186-0-0-9-3a4b2091e9" Dec 13 13:22:04.804107 kubelet[2824]: I1213 13:22:04.804018 2824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4186-0-0-9-3a4b2091e9" podStartSLOduration=2.80397814 podStartE2EDuration="2.80397814s" podCreationTimestamp="2024-12-13 13:22:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-12-13 13:22:04.791647719 +0000 UTC m=+1.198921245" watchObservedRunningTime="2024-12-13 13:22:04.80397814 +0000 UTC m=+1.211251666" Dec 13 13:22:04.820533 kubelet[2824]: I1213 13:22:04.820460 2824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4186-0-0-9-3a4b2091e9" podStartSLOduration=1.820437569 podStartE2EDuration="1.820437569s" podCreationTimestamp="2024-12-13 13:22:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-12-13 13:22:04.805445851 +0000 UTC m=+1.212719377" watchObservedRunningTime="2024-12-13 13:22:04.820437569 +0000 UTC m=+1.227711095" Dec 13 13:22:04.836853 kubelet[2824]: I1213 13:22:04.836775 2824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4186-0-0-9-3a4b2091e9" podStartSLOduration=1.836726514 podStartE2EDuration="1.836726514s" podCreationTimestamp="2024-12-13 13:22:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-12-13 13:22:04.821628474 +0000 UTC m=+1.228902000" watchObservedRunningTime="2024-12-13 13:22:04.836726514 +0000 UTC m=+1.244000080" Dec 13 13:22:06.221345 sudo[1878]: pam_unix(sudo:session): session closed for user root Dec 13 13:22:06.378903 sshd[1877]: Connection closed by 147.75.109.163 port 52630 Dec 13 13:22:06.379705 sshd-session[1875]: pam_unix(sshd:session): session closed for user core Dec 13 13:22:06.384492 systemd-logind[1465]: Session 7 logged out. Waiting for processes to exit. Dec 13 13:22:06.385448 systemd[1]: sshd@6-23.88.121.109:22-147.75.109.163:52630.service: Deactivated successfully. Dec 13 13:22:06.390663 systemd[1]: session-7.scope: Deactivated successfully. Dec 13 13:22:06.391010 systemd[1]: session-7.scope: Consumed 9.211s CPU time, 192.6M memory peak, 0B memory swap peak. Dec 13 13:22:06.392213 systemd-logind[1465]: Removed session 7. Dec 13 13:22:17.564494 kubelet[2824]: I1213 13:22:17.564311 2824 kuberuntime_manager.go:1523] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Dec 13 13:22:17.565770 containerd[1477]: time="2024-12-13T13:22:17.565429366Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Dec 13 13:22:17.566165 kubelet[2824]: I1213 13:22:17.565697 2824 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Dec 13 13:22:17.629193 kubelet[2824]: I1213 13:22:17.629117 2824 topology_manager.go:215] "Topology Admit Handler" podUID="87af4ba6-4abb-49f4-8748-18d3ba478286" podNamespace="kube-system" podName="cilium-mzcnw" Dec 13 13:22:17.631326 kubelet[2824]: I1213 13:22:17.631261 2824 topology_manager.go:215] "Topology Admit Handler" podUID="353c6283-ae18-42a6-8707-59d30708af57" podNamespace="kube-system" podName="kube-proxy-8f7v2" Dec 13 13:22:17.637779 kubelet[2824]: W1213 13:22:17.636556 2824 reflector.go:547] object-"kube-system"/"kube-proxy": failed to list *v1.ConfigMap: configmaps "kube-proxy" is forbidden: User "system:node:ci-4186-0-0-9-3a4b2091e9" cannot list resource "configmaps" in API group "" in the namespace "kube-system": no relationship found between node 'ci-4186-0-0-9-3a4b2091e9' and this object Dec 13 13:22:17.637779 kubelet[2824]: E1213 13:22:17.636625 2824 reflector.go:150] object-"kube-system"/"kube-proxy": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps "kube-proxy" is forbidden: User "system:node:ci-4186-0-0-9-3a4b2091e9" cannot list resource "configmaps" in API group "" in the namespace "kube-system": no relationship found between node 'ci-4186-0-0-9-3a4b2091e9' and this object Dec 13 13:22:17.645551 systemd[1]: Created slice kubepods-burstable-pod87af4ba6_4abb_49f4_8748_18d3ba478286.slice - libcontainer container kubepods-burstable-pod87af4ba6_4abb_49f4_8748_18d3ba478286.slice. Dec 13 13:22:17.654384 systemd[1]: Created slice kubepods-besteffort-pod353c6283_ae18_42a6_8707_59d30708af57.slice - libcontainer container kubepods-besteffort-pod353c6283_ae18_42a6_8707_59d30708af57.slice. Dec 13 13:22:17.672256 kubelet[2824]: I1213 13:22:17.671964 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-cilium-run\") pod \"cilium-mzcnw\" (UID: \"87af4ba6-4abb-49f4-8748-18d3ba478286\") " pod="kube-system/cilium-mzcnw" Dec 13 13:22:17.672256 kubelet[2824]: I1213 13:22:17.672024 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-etc-cni-netd\") pod \"cilium-mzcnw\" (UID: \"87af4ba6-4abb-49f4-8748-18d3ba478286\") " pod="kube-system/cilium-mzcnw" Dec 13 13:22:17.672256 kubelet[2824]: I1213 13:22:17.672049 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-host-proc-sys-kernel\") pod \"cilium-mzcnw\" (UID: \"87af4ba6-4abb-49f4-8748-18d3ba478286\") " pod="kube-system/cilium-mzcnw" Dec 13 13:22:17.672256 kubelet[2824]: I1213 13:22:17.672077 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/353c6283-ae18-42a6-8707-59d30708af57-xtables-lock\") pod \"kube-proxy-8f7v2\" (UID: \"353c6283-ae18-42a6-8707-59d30708af57\") " pod="kube-system/kube-proxy-8f7v2" Dec 13 13:22:17.672256 kubelet[2824]: I1213 13:22:17.672141 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-lib-modules\") pod \"cilium-mzcnw\" (UID: \"87af4ba6-4abb-49f4-8748-18d3ba478286\") " pod="kube-system/cilium-mzcnw" Dec 13 13:22:17.672256 kubelet[2824]: I1213 13:22:17.672160 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/87af4ba6-4abb-49f4-8748-18d3ba478286-clustermesh-secrets\") pod \"cilium-mzcnw\" (UID: \"87af4ba6-4abb-49f4-8748-18d3ba478286\") " pod="kube-system/cilium-mzcnw" Dec 13 13:22:17.672526 kubelet[2824]: I1213 13:22:17.672176 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmf64\" (UniqueName: \"kubernetes.io/projected/87af4ba6-4abb-49f4-8748-18d3ba478286-kube-api-access-tmf64\") pod \"cilium-mzcnw\" (UID: \"87af4ba6-4abb-49f4-8748-18d3ba478286\") " pod="kube-system/cilium-mzcnw" Dec 13 13:22:17.672526 kubelet[2824]: I1213 13:22:17.672194 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-bpf-maps\") pod \"cilium-mzcnw\" (UID: \"87af4ba6-4abb-49f4-8748-18d3ba478286\") " pod="kube-system/cilium-mzcnw" Dec 13 13:22:17.672526 kubelet[2824]: I1213 13:22:17.672210 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/87af4ba6-4abb-49f4-8748-18d3ba478286-cilium-config-path\") pod \"cilium-mzcnw\" (UID: \"87af4ba6-4abb-49f4-8748-18d3ba478286\") " pod="kube-system/cilium-mzcnw" Dec 13 13:22:17.672526 kubelet[2824]: I1213 13:22:17.672226 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-host-proc-sys-net\") pod \"cilium-mzcnw\" (UID: \"87af4ba6-4abb-49f4-8748-18d3ba478286\") " pod="kube-system/cilium-mzcnw" Dec 13 13:22:17.672526 kubelet[2824]: I1213 13:22:17.672244 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/353c6283-ae18-42a6-8707-59d30708af57-kube-proxy\") pod \"kube-proxy-8f7v2\" (UID: \"353c6283-ae18-42a6-8707-59d30708af57\") " pod="kube-system/kube-proxy-8f7v2" Dec 13 13:22:17.672712 kubelet[2824]: I1213 13:22:17.672266 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m624w\" (UniqueName: \"kubernetes.io/projected/353c6283-ae18-42a6-8707-59d30708af57-kube-api-access-m624w\") pod \"kube-proxy-8f7v2\" (UID: \"353c6283-ae18-42a6-8707-59d30708af57\") " pod="kube-system/kube-proxy-8f7v2" Dec 13 13:22:17.672712 kubelet[2824]: I1213 13:22:17.672283 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-xtables-lock\") pod \"cilium-mzcnw\" (UID: \"87af4ba6-4abb-49f4-8748-18d3ba478286\") " pod="kube-system/cilium-mzcnw" Dec 13 13:22:17.672712 kubelet[2824]: I1213 13:22:17.672297 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-cni-path\") pod \"cilium-mzcnw\" (UID: \"87af4ba6-4abb-49f4-8748-18d3ba478286\") " pod="kube-system/cilium-mzcnw" Dec 13 13:22:17.672712 kubelet[2824]: I1213 13:22:17.672315 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-hostproc\") pod \"cilium-mzcnw\" (UID: \"87af4ba6-4abb-49f4-8748-18d3ba478286\") " pod="kube-system/cilium-mzcnw" Dec 13 13:22:17.672712 kubelet[2824]: I1213 13:22:17.672334 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-cilium-cgroup\") pod \"cilium-mzcnw\" (UID: \"87af4ba6-4abb-49f4-8748-18d3ba478286\") " pod="kube-system/cilium-mzcnw" Dec 13 13:22:17.672712 kubelet[2824]: I1213 13:22:17.672350 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/353c6283-ae18-42a6-8707-59d30708af57-lib-modules\") pod \"kube-proxy-8f7v2\" (UID: \"353c6283-ae18-42a6-8707-59d30708af57\") " pod="kube-system/kube-proxy-8f7v2" Dec 13 13:22:17.674829 kubelet[2824]: I1213 13:22:17.672367 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/87af4ba6-4abb-49f4-8748-18d3ba478286-hubble-tls\") pod \"cilium-mzcnw\" (UID: \"87af4ba6-4abb-49f4-8748-18d3ba478286\") " pod="kube-system/cilium-mzcnw" Dec 13 13:22:17.801600 kubelet[2824]: E1213 13:22:17.800956 2824 projected.go:294] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found Dec 13 13:22:17.801600 kubelet[2824]: E1213 13:22:17.800992 2824 projected.go:200] Error preparing data for projected volume kube-api-access-tmf64 for pod kube-system/cilium-mzcnw: configmap "kube-root-ca.crt" not found Dec 13 13:22:17.801779 kubelet[2824]: E1213 13:22:17.801617 2824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/87af4ba6-4abb-49f4-8748-18d3ba478286-kube-api-access-tmf64 podName:87af4ba6-4abb-49f4-8748-18d3ba478286 nodeName:}" failed. No retries permitted until 2024-12-13 13:22:18.301590143 +0000 UTC m=+14.708863629 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-tmf64" (UniqueName: "kubernetes.io/projected/87af4ba6-4abb-49f4-8748-18d3ba478286-kube-api-access-tmf64") pod "cilium-mzcnw" (UID: "87af4ba6-4abb-49f4-8748-18d3ba478286") : configmap "kube-root-ca.crt" not found Dec 13 13:22:17.804063 kubelet[2824]: E1213 13:22:17.803936 2824 projected.go:294] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found Dec 13 13:22:17.804063 kubelet[2824]: E1213 13:22:17.803963 2824 projected.go:200] Error preparing data for projected volume kube-api-access-m624w for pod kube-system/kube-proxy-8f7v2: configmap "kube-root-ca.crt" not found Dec 13 13:22:17.804063 kubelet[2824]: E1213 13:22:17.804038 2824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/353c6283-ae18-42a6-8707-59d30708af57-kube-api-access-m624w podName:353c6283-ae18-42a6-8707-59d30708af57 nodeName:}" failed. No retries permitted until 2024-12-13 13:22:18.304018981 +0000 UTC m=+14.711292507 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-m624w" (UniqueName: "kubernetes.io/projected/353c6283-ae18-42a6-8707-59d30708af57-kube-api-access-m624w") pod "kube-proxy-8f7v2" (UID: "353c6283-ae18-42a6-8707-59d30708af57") : configmap "kube-root-ca.crt" not found Dec 13 13:22:18.379757 kubelet[2824]: E1213 13:22:18.379655 2824 projected.go:294] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found Dec 13 13:22:18.379757 kubelet[2824]: E1213 13:22:18.379701 2824 projected.go:200] Error preparing data for projected volume kube-api-access-m624w for pod kube-system/kube-proxy-8f7v2: configmap "kube-root-ca.crt" not found Dec 13 13:22:18.379757 kubelet[2824]: E1213 13:22:18.379754 2824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/353c6283-ae18-42a6-8707-59d30708af57-kube-api-access-m624w podName:353c6283-ae18-42a6-8707-59d30708af57 nodeName:}" failed. No retries permitted until 2024-12-13 13:22:19.379728725 +0000 UTC m=+15.787002251 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-m624w" (UniqueName: "kubernetes.io/projected/353c6283-ae18-42a6-8707-59d30708af57-kube-api-access-m624w") pod "kube-proxy-8f7v2" (UID: "353c6283-ae18-42a6-8707-59d30708af57") : configmap "kube-root-ca.crt" not found Dec 13 13:22:18.380346 kubelet[2824]: E1213 13:22:18.380311 2824 projected.go:294] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found Dec 13 13:22:18.380346 kubelet[2824]: E1213 13:22:18.380334 2824 projected.go:200] Error preparing data for projected volume kube-api-access-tmf64 for pod kube-system/cilium-mzcnw: configmap "kube-root-ca.crt" not found Dec 13 13:22:18.380461 kubelet[2824]: E1213 13:22:18.380366 2824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/87af4ba6-4abb-49f4-8748-18d3ba478286-kube-api-access-tmf64 podName:87af4ba6-4abb-49f4-8748-18d3ba478286 nodeName:}" failed. No retries permitted until 2024-12-13 13:22:19.380355775 +0000 UTC m=+15.787629301 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-tmf64" (UniqueName: "kubernetes.io/projected/87af4ba6-4abb-49f4-8748-18d3ba478286-kube-api-access-tmf64") pod "cilium-mzcnw" (UID: "87af4ba6-4abb-49f4-8748-18d3ba478286") : configmap "kube-root-ca.crt" not found Dec 13 13:22:18.694899 kubelet[2824]: I1213 13:22:18.694732 2824 topology_manager.go:215] "Topology Admit Handler" podUID="d2fb0668-07bd-4a53-9b0f-ec12c129578d" podNamespace="kube-system" podName="cilium-operator-599987898-z5bgv" Dec 13 13:22:18.706895 systemd[1]: Created slice kubepods-besteffort-podd2fb0668_07bd_4a53_9b0f_ec12c129578d.slice - libcontainer container kubepods-besteffort-podd2fb0668_07bd_4a53_9b0f_ec12c129578d.slice. Dec 13 13:22:18.778710 kubelet[2824]: E1213 13:22:18.778652 2824 configmap.go:199] Couldn't get configMap kube-system/kube-proxy: failed to sync configmap cache: timed out waiting for the condition Dec 13 13:22:18.779166 kubelet[2824]: E1213 13:22:18.778804 2824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/353c6283-ae18-42a6-8707-59d30708af57-kube-proxy podName:353c6283-ae18-42a6-8707-59d30708af57 nodeName:}" failed. No retries permitted until 2024-12-13 13:22:19.278773366 +0000 UTC m=+15.686046932 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-proxy" (UniqueName: "kubernetes.io/configmap/353c6283-ae18-42a6-8707-59d30708af57-kube-proxy") pod "kube-proxy-8f7v2" (UID: "353c6283-ae18-42a6-8707-59d30708af57") : failed to sync configmap cache: timed out waiting for the condition Dec 13 13:22:18.783046 kubelet[2824]: I1213 13:22:18.782957 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/d2fb0668-07bd-4a53-9b0f-ec12c129578d-cilium-config-path\") pod \"cilium-operator-599987898-z5bgv\" (UID: \"d2fb0668-07bd-4a53-9b0f-ec12c129578d\") " pod="kube-system/cilium-operator-599987898-z5bgv" Dec 13 13:22:18.783046 kubelet[2824]: I1213 13:22:18.783002 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9s9s7\" (UniqueName: \"kubernetes.io/projected/d2fb0668-07bd-4a53-9b0f-ec12c129578d-kube-api-access-9s9s7\") pod \"cilium-operator-599987898-z5bgv\" (UID: \"d2fb0668-07bd-4a53-9b0f-ec12c129578d\") " pod="kube-system/cilium-operator-599987898-z5bgv" Dec 13 13:22:19.013982 containerd[1477]: time="2024-12-13T13:22:19.013428115Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-599987898-z5bgv,Uid:d2fb0668-07bd-4a53-9b0f-ec12c129578d,Namespace:kube-system,Attempt:0,}" Dec 13 13:22:19.044051 containerd[1477]: time="2024-12-13T13:22:19.043581521Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 13:22:19.044051 containerd[1477]: time="2024-12-13T13:22:19.043682203Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 13:22:19.044051 containerd[1477]: time="2024-12-13T13:22:19.043709243Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 13:22:19.044051 containerd[1477]: time="2024-12-13T13:22:19.043918046Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 13:22:19.077142 systemd[1]: Started cri-containerd-ae6a1363bbce38682e7e9f82b14004da2fb101205b6d8d6eb3974ea986c256a3.scope - libcontainer container ae6a1363bbce38682e7e9f82b14004da2fb101205b6d8d6eb3974ea986c256a3. Dec 13 13:22:19.112430 containerd[1477]: time="2024-12-13T13:22:19.112386180Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-599987898-z5bgv,Uid:d2fb0668-07bd-4a53-9b0f-ec12c129578d,Namespace:kube-system,Attempt:0,} returns sandbox id \"ae6a1363bbce38682e7e9f82b14004da2fb101205b6d8d6eb3974ea986c256a3\"" Dec 13 13:22:19.116176 containerd[1477]: time="2024-12-13T13:22:19.116133515Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Dec 13 13:22:19.451902 containerd[1477]: time="2024-12-13T13:22:19.451400998Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-mzcnw,Uid:87af4ba6-4abb-49f4-8748-18d3ba478286,Namespace:kube-system,Attempt:0,}" Dec 13 13:22:19.464064 containerd[1477]: time="2024-12-13T13:22:19.464024105Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-8f7v2,Uid:353c6283-ae18-42a6-8707-59d30708af57,Namespace:kube-system,Attempt:0,}" Dec 13 13:22:19.480722 containerd[1477]: time="2024-12-13T13:22:19.480597630Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 13:22:19.481180 containerd[1477]: time="2024-12-13T13:22:19.481008276Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 13:22:19.481180 containerd[1477]: time="2024-12-13T13:22:19.481052637Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 13:22:19.481360 containerd[1477]: time="2024-12-13T13:22:19.481266160Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 13:22:19.498316 containerd[1477]: time="2024-12-13T13:22:19.497934247Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 13:22:19.498316 containerd[1477]: time="2024-12-13T13:22:19.498012968Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 13:22:19.498316 containerd[1477]: time="2024-12-13T13:22:19.498030808Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 13:22:19.498316 containerd[1477]: time="2024-12-13T13:22:19.498140930Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 13:22:19.504784 systemd[1]: Started cri-containerd-08a2754e454ae96132c191313465a0ff2e9480b13d88079a4a438ec4f2571f81.scope - libcontainer container 08a2754e454ae96132c191313465a0ff2e9480b13d88079a4a438ec4f2571f81. Dec 13 13:22:19.522990 systemd[1]: Started cri-containerd-7d059d24982a0209774c10bb1682e6aa91e29d84c4865dbdf07a4e60327191f4.scope - libcontainer container 7d059d24982a0209774c10bb1682e6aa91e29d84c4865dbdf07a4e60327191f4. Dec 13 13:22:19.541427 containerd[1477]: time="2024-12-13T13:22:19.541286368Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-mzcnw,Uid:87af4ba6-4abb-49f4-8748-18d3ba478286,Namespace:kube-system,Attempt:0,} returns sandbox id \"08a2754e454ae96132c191313465a0ff2e9480b13d88079a4a438ec4f2571f81\"" Dec 13 13:22:19.560627 containerd[1477]: time="2024-12-13T13:22:19.560588814Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-8f7v2,Uid:353c6283-ae18-42a6-8707-59d30708af57,Namespace:kube-system,Attempt:0,} returns sandbox id \"7d059d24982a0209774c10bb1682e6aa91e29d84c4865dbdf07a4e60327191f4\"" Dec 13 13:22:19.565908 containerd[1477]: time="2024-12-13T13:22:19.565729770Z" level=info msg="CreateContainer within sandbox \"7d059d24982a0209774c10bb1682e6aa91e29d84c4865dbdf07a4e60327191f4\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Dec 13 13:22:19.582574 containerd[1477]: time="2024-12-13T13:22:19.582403857Z" level=info msg="CreateContainer within sandbox \"7d059d24982a0209774c10bb1682e6aa91e29d84c4865dbdf07a4e60327191f4\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"88fa657f50439f95e4f10cd531de9ae9b067c6f5c7fe5e1c00fdbac478b0fb2c\"" Dec 13 13:22:19.583844 containerd[1477]: time="2024-12-13T13:22:19.583433152Z" level=info msg="StartContainer for \"88fa657f50439f95e4f10cd531de9ae9b067c6f5c7fe5e1c00fdbac478b0fb2c\"" Dec 13 13:22:19.618919 systemd[1]: Started cri-containerd-88fa657f50439f95e4f10cd531de9ae9b067c6f5c7fe5e1c00fdbac478b0fb2c.scope - libcontainer container 88fa657f50439f95e4f10cd531de9ae9b067c6f5c7fe5e1c00fdbac478b0fb2c. Dec 13 13:22:19.655247 containerd[1477]: time="2024-12-13T13:22:19.655108653Z" level=info msg="StartContainer for \"88fa657f50439f95e4f10cd531de9ae9b067c6f5c7fe5e1c00fdbac478b0fb2c\" returns successfully" Dec 13 13:22:22.276403 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1614065331.mount: Deactivated successfully. Dec 13 13:22:23.176896 containerd[1477]: time="2024-12-13T13:22:23.176021956Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 13:22:23.178669 containerd[1477]: time="2024-12-13T13:22:23.178264347Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=17138294" Dec 13 13:22:23.180218 containerd[1477]: time="2024-12-13T13:22:23.180054811Z" level=info msg="ImageCreate event name:\"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 13:22:23.182045 containerd[1477]: time="2024-12-13T13:22:23.181860035Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"17128551\" in 4.065676399s" Dec 13 13:22:23.182045 containerd[1477]: time="2024-12-13T13:22:23.181915836Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\"" Dec 13 13:22:23.184214 containerd[1477]: time="2024-12-13T13:22:23.183971424Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Dec 13 13:22:23.185574 containerd[1477]: time="2024-12-13T13:22:23.185439444Z" level=info msg="CreateContainer within sandbox \"ae6a1363bbce38682e7e9f82b14004da2fb101205b6d8d6eb3974ea986c256a3\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Dec 13 13:22:23.206590 containerd[1477]: time="2024-12-13T13:22:23.206503810Z" level=info msg="CreateContainer within sandbox \"ae6a1363bbce38682e7e9f82b14004da2fb101205b6d8d6eb3974ea986c256a3\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"9fc7582bf65547c02d26deb9fc8a19b164c9cf14800312e5ddcac69a4829f858\"" Dec 13 13:22:23.207795 containerd[1477]: time="2024-12-13T13:22:23.207369661Z" level=info msg="StartContainer for \"9fc7582bf65547c02d26deb9fc8a19b164c9cf14800312e5ddcac69a4829f858\"" Dec 13 13:22:23.237995 systemd[1]: Started cri-containerd-9fc7582bf65547c02d26deb9fc8a19b164c9cf14800312e5ddcac69a4829f858.scope - libcontainer container 9fc7582bf65547c02d26deb9fc8a19b164c9cf14800312e5ddcac69a4829f858. Dec 13 13:22:23.266808 containerd[1477]: time="2024-12-13T13:22:23.266229659Z" level=info msg="StartContainer for \"9fc7582bf65547c02d26deb9fc8a19b164c9cf14800312e5ddcac69a4829f858\" returns successfully" Dec 13 13:22:23.713895 kubelet[2824]: I1213 13:22:23.713827 2824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-8f7v2" podStartSLOduration=6.713809807 podStartE2EDuration="6.713809807s" podCreationTimestamp="2024-12-13 13:22:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-12-13 13:22:19.824944807 +0000 UTC m=+16.232218333" watchObservedRunningTime="2024-12-13 13:22:23.713809807 +0000 UTC m=+20.121083333" Dec 13 13:22:31.398578 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount583141925.mount: Deactivated successfully. Dec 13 13:22:32.957355 containerd[1477]: time="2024-12-13T13:22:32.956373742Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 13:22:32.957355 containerd[1477]: time="2024-12-13T13:22:32.957288833Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=157649686" Dec 13 13:22:32.958440 containerd[1477]: time="2024-12-13T13:22:32.958391965Z" level=info msg="ImageCreate event name:\"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 13:22:32.960911 containerd[1477]: time="2024-12-13T13:22:32.960868073Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"157636062\" in 9.776848848s" Dec 13 13:22:32.961310 containerd[1477]: time="2024-12-13T13:22:32.960915514Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\"" Dec 13 13:22:32.966195 containerd[1477]: time="2024-12-13T13:22:32.966011131Z" level=info msg="CreateContainer within sandbox \"08a2754e454ae96132c191313465a0ff2e9480b13d88079a4a438ec4f2571f81\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Dec 13 13:22:32.986968 containerd[1477]: time="2024-12-13T13:22:32.986826246Z" level=info msg="CreateContainer within sandbox \"08a2754e454ae96132c191313465a0ff2e9480b13d88079a4a438ec4f2571f81\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"36c7bf433ad4194395ef0cec0b50e5506f334bf0955a6a0cea088d55207c9053\"" Dec 13 13:22:32.989879 containerd[1477]: time="2024-12-13T13:22:32.987404172Z" level=info msg="StartContainer for \"36c7bf433ad4194395ef0cec0b50e5506f334bf0955a6a0cea088d55207c9053\"" Dec 13 13:22:33.026185 systemd[1]: Started cri-containerd-36c7bf433ad4194395ef0cec0b50e5506f334bf0955a6a0cea088d55207c9053.scope - libcontainer container 36c7bf433ad4194395ef0cec0b50e5506f334bf0955a6a0cea088d55207c9053. Dec 13 13:22:33.059377 containerd[1477]: time="2024-12-13T13:22:33.059332210Z" level=info msg="StartContainer for \"36c7bf433ad4194395ef0cec0b50e5506f334bf0955a6a0cea088d55207c9053\" returns successfully" Dec 13 13:22:33.074373 systemd[1]: cri-containerd-36c7bf433ad4194395ef0cec0b50e5506f334bf0955a6a0cea088d55207c9053.scope: Deactivated successfully. Dec 13 13:22:33.312315 containerd[1477]: time="2024-12-13T13:22:33.311867442Z" level=info msg="shim disconnected" id=36c7bf433ad4194395ef0cec0b50e5506f334bf0955a6a0cea088d55207c9053 namespace=k8s.io Dec 13 13:22:33.312315 containerd[1477]: time="2024-12-13T13:22:33.311971923Z" level=warning msg="cleaning up after shim disconnected" id=36c7bf433ad4194395ef0cec0b50e5506f334bf0955a6a0cea088d55207c9053 namespace=k8s.io Dec 13 13:22:33.312315 containerd[1477]: time="2024-12-13T13:22:33.311992684Z" level=info msg="cleaning up dead shim" namespace=k8s.io Dec 13 13:22:33.326408 containerd[1477]: time="2024-12-13T13:22:33.325760756Z" level=warning msg="cleanup warnings time=\"2024-12-13T13:22:33Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Dec 13 13:22:33.857195 containerd[1477]: time="2024-12-13T13:22:33.856855108Z" level=info msg="CreateContainer within sandbox \"08a2754e454ae96132c191313465a0ff2e9480b13d88079a4a438ec4f2571f81\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Dec 13 13:22:33.869434 containerd[1477]: time="2024-12-13T13:22:33.869375166Z" level=info msg="CreateContainer within sandbox \"08a2754e454ae96132c191313465a0ff2e9480b13d88079a4a438ec4f2571f81\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"28003be96e8c1eaa0f2d3a766ba08814c4e37ed4b1a896dccf7c066b098fd8a0\"" Dec 13 13:22:33.872702 containerd[1477]: time="2024-12-13T13:22:33.871721512Z" level=info msg="StartContainer for \"28003be96e8c1eaa0f2d3a766ba08814c4e37ed4b1a896dccf7c066b098fd8a0\"" Dec 13 13:22:33.885647 kubelet[2824]: I1213 13:22:33.885197 2824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-599987898-z5bgv" podStartSLOduration=11.816552579 podStartE2EDuration="15.885180141s" podCreationTimestamp="2024-12-13 13:22:18 +0000 UTC" firstStartedPulling="2024-12-13 13:22:19.114528371 +0000 UTC m=+15.521801897" lastFinishedPulling="2024-12-13 13:22:23.183155933 +0000 UTC m=+19.590429459" observedRunningTime="2024-12-13 13:22:23.973575049 +0000 UTC m=+20.380848575" watchObservedRunningTime="2024-12-13 13:22:33.885180141 +0000 UTC m=+30.292453667" Dec 13 13:22:33.914091 systemd[1]: Started cri-containerd-28003be96e8c1eaa0f2d3a766ba08814c4e37ed4b1a896dccf7c066b098fd8a0.scope - libcontainer container 28003be96e8c1eaa0f2d3a766ba08814c4e37ed4b1a896dccf7c066b098fd8a0. Dec 13 13:22:33.948924 containerd[1477]: time="2024-12-13T13:22:33.947847834Z" level=info msg="StartContainer for \"28003be96e8c1eaa0f2d3a766ba08814c4e37ed4b1a896dccf7c066b098fd8a0\" returns successfully" Dec 13 13:22:33.963219 systemd[1]: systemd-sysctl.service: Deactivated successfully. Dec 13 13:22:33.964010 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Dec 13 13:22:33.964089 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables... Dec 13 13:22:33.971150 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Dec 13 13:22:33.971893 systemd[1]: cri-containerd-28003be96e8c1eaa0f2d3a766ba08814c4e37ed4b1a896dccf7c066b098fd8a0.scope: Deactivated successfully. Dec 13 13:22:33.981351 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-36c7bf433ad4194395ef0cec0b50e5506f334bf0955a6a0cea088d55207c9053-rootfs.mount: Deactivated successfully. Dec 13 13:22:33.998948 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-28003be96e8c1eaa0f2d3a766ba08814c4e37ed4b1a896dccf7c066b098fd8a0-rootfs.mount: Deactivated successfully. Dec 13 13:22:34.003268 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Dec 13 13:22:34.009497 containerd[1477]: time="2024-12-13T13:22:34.008288700Z" level=info msg="shim disconnected" id=28003be96e8c1eaa0f2d3a766ba08814c4e37ed4b1a896dccf7c066b098fd8a0 namespace=k8s.io Dec 13 13:22:34.009497 containerd[1477]: time="2024-12-13T13:22:34.008364781Z" level=warning msg="cleaning up after shim disconnected" id=28003be96e8c1eaa0f2d3a766ba08814c4e37ed4b1a896dccf7c066b098fd8a0 namespace=k8s.io Dec 13 13:22:34.009497 containerd[1477]: time="2024-12-13T13:22:34.008374221Z" level=info msg="cleaning up dead shim" namespace=k8s.io Dec 13 13:22:34.867178 containerd[1477]: time="2024-12-13T13:22:34.867081655Z" level=info msg="CreateContainer within sandbox \"08a2754e454ae96132c191313465a0ff2e9480b13d88079a4a438ec4f2571f81\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Dec 13 13:22:34.891161 containerd[1477]: time="2024-12-13T13:22:34.891064395Z" level=info msg="CreateContainer within sandbox \"08a2754e454ae96132c191313465a0ff2e9480b13d88079a4a438ec4f2571f81\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"ca0716e2d915f9418f463fc7052030f47974ebb73c009c58197e2bed939f0016\"" Dec 13 13:22:34.892253 containerd[1477]: time="2024-12-13T13:22:34.892213368Z" level=info msg="StartContainer for \"ca0716e2d915f9418f463fc7052030f47974ebb73c009c58197e2bed939f0016\"" Dec 13 13:22:34.927987 systemd[1]: Started cri-containerd-ca0716e2d915f9418f463fc7052030f47974ebb73c009c58197e2bed939f0016.scope - libcontainer container ca0716e2d915f9418f463fc7052030f47974ebb73c009c58197e2bed939f0016. Dec 13 13:22:34.963429 containerd[1477]: time="2024-12-13T13:22:34.963285859Z" level=info msg="StartContainer for \"ca0716e2d915f9418f463fc7052030f47974ebb73c009c58197e2bed939f0016\" returns successfully" Dec 13 13:22:34.972175 systemd[1]: cri-containerd-ca0716e2d915f9418f463fc7052030f47974ebb73c009c58197e2bed939f0016.scope: Deactivated successfully. Dec 13 13:22:34.999931 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-ca0716e2d915f9418f463fc7052030f47974ebb73c009c58197e2bed939f0016-rootfs.mount: Deactivated successfully. Dec 13 13:22:35.007342 containerd[1477]: time="2024-12-13T13:22:35.007154093Z" level=info msg="shim disconnected" id=ca0716e2d915f9418f463fc7052030f47974ebb73c009c58197e2bed939f0016 namespace=k8s.io Dec 13 13:22:35.007342 containerd[1477]: time="2024-12-13T13:22:35.007281415Z" level=warning msg="cleaning up after shim disconnected" id=ca0716e2d915f9418f463fc7052030f47974ebb73c009c58197e2bed939f0016 namespace=k8s.io Dec 13 13:22:35.007342 containerd[1477]: time="2024-12-13T13:22:35.007295415Z" level=info msg="cleaning up dead shim" namespace=k8s.io Dec 13 13:22:35.871796 containerd[1477]: time="2024-12-13T13:22:35.871730136Z" level=info msg="CreateContainer within sandbox \"08a2754e454ae96132c191313465a0ff2e9480b13d88079a4a438ec4f2571f81\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Dec 13 13:22:35.899328 containerd[1477]: time="2024-12-13T13:22:35.896412279Z" level=info msg="CreateContainer within sandbox \"08a2754e454ae96132c191313465a0ff2e9480b13d88079a4a438ec4f2571f81\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"590a9dd551ac1e339d1e595c191fa336cd7d22786ab1da6d8b47b543c8234a0b\"" Dec 13 13:22:35.899328 containerd[1477]: time="2024-12-13T13:22:35.897870294Z" level=info msg="StartContainer for \"590a9dd551ac1e339d1e595c191fa336cd7d22786ab1da6d8b47b543c8234a0b\"" Dec 13 13:22:35.929951 systemd[1]: Started cri-containerd-590a9dd551ac1e339d1e595c191fa336cd7d22786ab1da6d8b47b543c8234a0b.scope - libcontainer container 590a9dd551ac1e339d1e595c191fa336cd7d22786ab1da6d8b47b543c8234a0b. Dec 13 13:22:35.958461 systemd[1]: cri-containerd-590a9dd551ac1e339d1e595c191fa336cd7d22786ab1da6d8b47b543c8234a0b.scope: Deactivated successfully. Dec 13 13:22:35.960732 containerd[1477]: time="2024-12-13T13:22:35.960588802Z" level=info msg="StartContainer for \"590a9dd551ac1e339d1e595c191fa336cd7d22786ab1da6d8b47b543c8234a0b\" returns successfully" Dec 13 13:22:35.983987 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-590a9dd551ac1e339d1e595c191fa336cd7d22786ab1da6d8b47b543c8234a0b-rootfs.mount: Deactivated successfully. Dec 13 13:22:35.990470 containerd[1477]: time="2024-12-13T13:22:35.990158116Z" level=info msg="shim disconnected" id=590a9dd551ac1e339d1e595c191fa336cd7d22786ab1da6d8b47b543c8234a0b namespace=k8s.io Dec 13 13:22:35.990470 containerd[1477]: time="2024-12-13T13:22:35.990237077Z" level=warning msg="cleaning up after shim disconnected" id=590a9dd551ac1e339d1e595c191fa336cd7d22786ab1da6d8b47b543c8234a0b namespace=k8s.io Dec 13 13:22:35.990470 containerd[1477]: time="2024-12-13T13:22:35.990254477Z" level=info msg="cleaning up dead shim" namespace=k8s.io Dec 13 13:22:36.879843 containerd[1477]: time="2024-12-13T13:22:36.879657132Z" level=info msg="CreateContainer within sandbox \"08a2754e454ae96132c191313465a0ff2e9480b13d88079a4a438ec4f2571f81\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Dec 13 13:22:36.905646 containerd[1477]: time="2024-12-13T13:22:36.905082397Z" level=info msg="CreateContainer within sandbox \"08a2754e454ae96132c191313465a0ff2e9480b13d88079a4a438ec4f2571f81\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"7eef8c509f1c6c520e52af996fe6b3e92243ab283e5e5a2ee093a1c37553e171\"" Dec 13 13:22:36.907541 containerd[1477]: time="2024-12-13T13:22:36.907377021Z" level=info msg="StartContainer for \"7eef8c509f1c6c520e52af996fe6b3e92243ab283e5e5a2ee093a1c37553e171\"" Dec 13 13:22:36.941960 systemd[1]: Started cri-containerd-7eef8c509f1c6c520e52af996fe6b3e92243ab283e5e5a2ee093a1c37553e171.scope - libcontainer container 7eef8c509f1c6c520e52af996fe6b3e92243ab283e5e5a2ee093a1c37553e171. Dec 13 13:22:36.977770 containerd[1477]: time="2024-12-13T13:22:36.977435393Z" level=info msg="StartContainer for \"7eef8c509f1c6c520e52af996fe6b3e92243ab283e5e5a2ee093a1c37553e171\" returns successfully" Dec 13 13:22:37.140238 kubelet[2824]: I1213 13:22:37.140082 2824 kubelet_node_status.go:497] "Fast updating node status as it just became ready" Dec 13 13:22:37.179316 kubelet[2824]: I1213 13:22:37.179260 2824 topology_manager.go:215] "Topology Admit Handler" podUID="bcafff9a-8db6-4f56-b215-caee06539292" podNamespace="kube-system" podName="coredns-7db6d8ff4d-w6kfn" Dec 13 13:22:37.183919 kubelet[2824]: I1213 13:22:37.183878 2824 topology_manager.go:215] "Topology Admit Handler" podUID="4f9d26cd-c4d1-47f8-9a9a-b3321d7c8878" podNamespace="kube-system" podName="coredns-7db6d8ff4d-pt7lq" Dec 13 13:22:37.191689 systemd[1]: Created slice kubepods-burstable-podbcafff9a_8db6_4f56_b215_caee06539292.slice - libcontainer container kubepods-burstable-podbcafff9a_8db6_4f56_b215_caee06539292.slice. Dec 13 13:22:37.198529 systemd[1]: Created slice kubepods-burstable-pod4f9d26cd_c4d1_47f8_9a9a_b3321d7c8878.slice - libcontainer container kubepods-burstable-pod4f9d26cd_c4d1_47f8_9a9a_b3321d7c8878.slice. Dec 13 13:22:37.219467 kubelet[2824]: I1213 13:22:37.219397 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8gx8\" (UniqueName: \"kubernetes.io/projected/bcafff9a-8db6-4f56-b215-caee06539292-kube-api-access-k8gx8\") pod \"coredns-7db6d8ff4d-w6kfn\" (UID: \"bcafff9a-8db6-4f56-b215-caee06539292\") " pod="kube-system/coredns-7db6d8ff4d-w6kfn" Dec 13 13:22:37.219467 kubelet[2824]: I1213 13:22:37.219445 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4f9d26cd-c4d1-47f8-9a9a-b3321d7c8878-config-volume\") pod \"coredns-7db6d8ff4d-pt7lq\" (UID: \"4f9d26cd-c4d1-47f8-9a9a-b3321d7c8878\") " pod="kube-system/coredns-7db6d8ff4d-pt7lq" Dec 13 13:22:37.219467 kubelet[2824]: I1213 13:22:37.219467 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4wwb\" (UniqueName: \"kubernetes.io/projected/4f9d26cd-c4d1-47f8-9a9a-b3321d7c8878-kube-api-access-t4wwb\") pod \"coredns-7db6d8ff4d-pt7lq\" (UID: \"4f9d26cd-c4d1-47f8-9a9a-b3321d7c8878\") " pod="kube-system/coredns-7db6d8ff4d-pt7lq" Dec 13 13:22:37.219930 kubelet[2824]: I1213 13:22:37.219488 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bcafff9a-8db6-4f56-b215-caee06539292-config-volume\") pod \"coredns-7db6d8ff4d-w6kfn\" (UID: \"bcafff9a-8db6-4f56-b215-caee06539292\") " pod="kube-system/coredns-7db6d8ff4d-w6kfn" Dec 13 13:22:37.499106 containerd[1477]: time="2024-12-13T13:22:37.498966227Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-w6kfn,Uid:bcafff9a-8db6-4f56-b215-caee06539292,Namespace:kube-system,Attempt:0,}" Dec 13 13:22:37.504245 containerd[1477]: time="2024-12-13T13:22:37.504191041Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-pt7lq,Uid:4f9d26cd-c4d1-47f8-9a9a-b3321d7c8878,Namespace:kube-system,Attempt:0,}" Dec 13 13:22:39.165030 systemd-networkd[1383]: cilium_host: Link UP Dec 13 13:22:39.165245 systemd-networkd[1383]: cilium_net: Link UP Dec 13 13:22:39.165370 systemd-networkd[1383]: cilium_net: Gained carrier Dec 13 13:22:39.165478 systemd-networkd[1383]: cilium_host: Gained carrier Dec 13 13:22:39.281954 systemd-networkd[1383]: cilium_vxlan: Link UP Dec 13 13:22:39.281963 systemd-networkd[1383]: cilium_vxlan: Gained carrier Dec 13 13:22:39.494214 systemd-networkd[1383]: cilium_host: Gained IPv6LL Dec 13 13:22:39.576796 kernel: NET: Registered PF_ALG protocol family Dec 13 13:22:39.757093 systemd-networkd[1383]: cilium_net: Gained IPv6LL Dec 13 13:22:40.332899 systemd-networkd[1383]: lxc_health: Link UP Dec 13 13:22:40.343053 systemd-networkd[1383]: lxc_health: Gained carrier Dec 13 13:22:40.580538 systemd-networkd[1383]: lxcf617d94f825b: Link UP Dec 13 13:22:40.587401 systemd-networkd[1383]: lxc3dc8d169f83b: Link UP Dec 13 13:22:40.598790 kernel: eth0: renamed from tmpc4d19 Dec 13 13:22:40.607840 kernel: eth0: renamed from tmp44bc6 Dec 13 13:22:40.605825 systemd-networkd[1383]: lxcf617d94f825b: Gained carrier Dec 13 13:22:40.613852 systemd-networkd[1383]: lxc3dc8d169f83b: Gained carrier Dec 13 13:22:40.717360 systemd-networkd[1383]: cilium_vxlan: Gained IPv6LL Dec 13 13:22:41.479918 kubelet[2824]: I1213 13:22:41.478929 2824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-mzcnw" podStartSLOduration=11.060064064 podStartE2EDuration="24.478910552s" podCreationTimestamp="2024-12-13 13:22:17 +0000 UTC" firstStartedPulling="2024-12-13 13:22:19.543539162 +0000 UTC m=+15.950812688" lastFinishedPulling="2024-12-13 13:22:32.96238565 +0000 UTC m=+29.369659176" observedRunningTime="2024-12-13 13:22:37.904878671 +0000 UTC m=+34.312152237" watchObservedRunningTime="2024-12-13 13:22:41.478910552 +0000 UTC m=+37.886184078" Dec 13 13:22:41.613059 systemd-networkd[1383]: lxc_health: Gained IPv6LL Dec 13 13:22:41.804973 systemd-networkd[1383]: lxc3dc8d169f83b: Gained IPv6LL Dec 13 13:22:42.637003 systemd-networkd[1383]: lxcf617d94f825b: Gained IPv6LL Dec 13 13:22:44.623830 containerd[1477]: time="2024-12-13T13:22:44.623619982Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 13:22:44.623830 containerd[1477]: time="2024-12-13T13:22:44.623679503Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 13:22:44.623830 containerd[1477]: time="2024-12-13T13:22:44.623690623Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 13:22:44.623830 containerd[1477]: time="2024-12-13T13:22:44.623789464Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 13:22:44.651596 systemd[1]: run-containerd-runc-k8s.io-c4d190ca4b9c8f72b130470b71162e0c96651853101216611d96876dd750530d-runc.udRgh5.mount: Deactivated successfully. Dec 13 13:22:44.661969 systemd[1]: Started cri-containerd-c4d190ca4b9c8f72b130470b71162e0c96651853101216611d96876dd750530d.scope - libcontainer container c4d190ca4b9c8f72b130470b71162e0c96651853101216611d96876dd750530d. Dec 13 13:22:44.689726 containerd[1477]: time="2024-12-13T13:22:44.688866095Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 13:22:44.689726 containerd[1477]: time="2024-12-13T13:22:44.688925575Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 13:22:44.689726 containerd[1477]: time="2024-12-13T13:22:44.688937495Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 13:22:44.689726 containerd[1477]: time="2024-12-13T13:22:44.689017856Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 13:22:44.716665 systemd[1]: Started cri-containerd-44bc6978a37dacfdea65f71f2b2eb889c0464fb78af2431717cb410badce62b2.scope - libcontainer container 44bc6978a37dacfdea65f71f2b2eb889c0464fb78af2431717cb410badce62b2. Dec 13 13:22:44.740002 containerd[1477]: time="2024-12-13T13:22:44.739838478Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-w6kfn,Uid:bcafff9a-8db6-4f56-b215-caee06539292,Namespace:kube-system,Attempt:0,} returns sandbox id \"c4d190ca4b9c8f72b130470b71162e0c96651853101216611d96876dd750530d\"" Dec 13 13:22:44.747850 containerd[1477]: time="2024-12-13T13:22:44.747798670Z" level=info msg="CreateContainer within sandbox \"c4d190ca4b9c8f72b130470b71162e0c96651853101216611d96876dd750530d\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Dec 13 13:22:44.769241 containerd[1477]: time="2024-12-13T13:22:44.769105464Z" level=info msg="CreateContainer within sandbox \"c4d190ca4b9c8f72b130470b71162e0c96651853101216611d96876dd750530d\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"1ba2b73f131e0568ab97c8597788aa7871b77a11ddc24f28a5e9e071455c9041\"" Dec 13 13:22:44.771831 containerd[1477]: time="2024-12-13T13:22:44.771651087Z" level=info msg="StartContainer for \"1ba2b73f131e0568ab97c8597788aa7871b77a11ddc24f28a5e9e071455c9041\"" Dec 13 13:22:44.786252 containerd[1477]: time="2024-12-13T13:22:44.786200459Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-pt7lq,Uid:4f9d26cd-c4d1-47f8-9a9a-b3321d7c8878,Namespace:kube-system,Attempt:0,} returns sandbox id \"44bc6978a37dacfdea65f71f2b2eb889c0464fb78af2431717cb410badce62b2\"" Dec 13 13:22:44.791183 containerd[1477]: time="2024-12-13T13:22:44.791040983Z" level=info msg="CreateContainer within sandbox \"44bc6978a37dacfdea65f71f2b2eb889c0464fb78af2431717cb410badce62b2\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Dec 13 13:22:44.815965 containerd[1477]: time="2024-12-13T13:22:44.815913049Z" level=info msg="CreateContainer within sandbox \"44bc6978a37dacfdea65f71f2b2eb889c0464fb78af2431717cb410badce62b2\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"ddfa0666453ac363809ad8dd0b69115099fdb1f3126fc1116bf16cbf70ad5dd4\"" Dec 13 13:22:44.816580 containerd[1477]: time="2024-12-13T13:22:44.816549695Z" level=info msg="StartContainer for \"ddfa0666453ac363809ad8dd0b69115099fdb1f3126fc1116bf16cbf70ad5dd4\"" Dec 13 13:22:44.818195 systemd[1]: Started cri-containerd-1ba2b73f131e0568ab97c8597788aa7871b77a11ddc24f28a5e9e071455c9041.scope - libcontainer container 1ba2b73f131e0568ab97c8597788aa7871b77a11ddc24f28a5e9e071455c9041. Dec 13 13:22:44.867276 containerd[1477]: time="2024-12-13T13:22:44.867000313Z" level=info msg="StartContainer for \"1ba2b73f131e0568ab97c8597788aa7871b77a11ddc24f28a5e9e071455c9041\" returns successfully" Dec 13 13:22:44.869842 systemd[1]: Started cri-containerd-ddfa0666453ac363809ad8dd0b69115099fdb1f3126fc1116bf16cbf70ad5dd4.scope - libcontainer container ddfa0666453ac363809ad8dd0b69115099fdb1f3126fc1116bf16cbf70ad5dd4. Dec 13 13:22:44.917230 containerd[1477]: time="2024-12-13T13:22:44.917025487Z" level=info msg="StartContainer for \"ddfa0666453ac363809ad8dd0b69115099fdb1f3126fc1116bf16cbf70ad5dd4\" returns successfully" Dec 13 13:22:45.927950 kubelet[2824]: I1213 13:22:45.926806 2824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7db6d8ff4d-w6kfn" podStartSLOduration=27.926777762 podStartE2EDuration="27.926777762s" podCreationTimestamp="2024-12-13 13:22:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-12-13 13:22:44.937631394 +0000 UTC m=+41.344904920" watchObservedRunningTime="2024-12-13 13:22:45.926777762 +0000 UTC m=+42.334051328" Dec 13 13:22:45.948652 kubelet[2824]: I1213 13:22:45.947652 2824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7db6d8ff4d-pt7lq" podStartSLOduration=27.947629668 podStartE2EDuration="27.947629668s" podCreationTimestamp="2024-12-13 13:22:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-12-13 13:22:45.929803629 +0000 UTC m=+42.337077155" watchObservedRunningTime="2024-12-13 13:22:45.947629668 +0000 UTC m=+42.354903194" Dec 13 13:26:48.385834 update_engine[1468]: I20241213 13:26:48.385704 1468 prefs.cc:52] certificate-report-to-send-update not present in /var/lib/update_engine/prefs Dec 13 13:26:48.385834 update_engine[1468]: I20241213 13:26:48.385833 1468 prefs.cc:52] certificate-report-to-send-download not present in /var/lib/update_engine/prefs Dec 13 13:26:48.392160 update_engine[1468]: I20241213 13:26:48.386123 1468 prefs.cc:52] aleph-version not present in /var/lib/update_engine/prefs Dec 13 13:26:48.392160 update_engine[1468]: I20241213 13:26:48.386695 1468 omaha_request_params.cc:62] Current group set to alpha Dec 13 13:26:48.392160 update_engine[1468]: I20241213 13:26:48.386817 1468 update_attempter.cc:499] Already updated boot flags. Skipping. Dec 13 13:26:48.392160 update_engine[1468]: I20241213 13:26:48.386829 1468 update_attempter.cc:643] Scheduling an action processor start. Dec 13 13:26:48.392160 update_engine[1468]: I20241213 13:26:48.386848 1468 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction Dec 13 13:26:48.392160 update_engine[1468]: I20241213 13:26:48.386885 1468 prefs.cc:52] previous-version not present in /var/lib/update_engine/prefs Dec 13 13:26:48.392160 update_engine[1468]: I20241213 13:26:48.386941 1468 omaha_request_action.cc:271] Posting an Omaha request to disabled Dec 13 13:26:48.392160 update_engine[1468]: I20241213 13:26:48.386950 1468 omaha_request_action.cc:272] Request: Dec 13 13:26:48.392160 update_engine[1468]: Dec 13 13:26:48.392160 update_engine[1468]: Dec 13 13:26:48.392160 update_engine[1468]: Dec 13 13:26:48.392160 update_engine[1468]: Dec 13 13:26:48.392160 update_engine[1468]: Dec 13 13:26:48.392160 update_engine[1468]: Dec 13 13:26:48.392160 update_engine[1468]: Dec 13 13:26:48.392160 update_engine[1468]: Dec 13 13:26:48.392160 update_engine[1468]: I20241213 13:26:48.386957 1468 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Dec 13 13:26:48.392160 update_engine[1468]: I20241213 13:26:48.389549 1468 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Dec 13 13:26:48.392160 update_engine[1468]: I20241213 13:26:48.390129 1468 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Dec 13 13:26:48.392160 update_engine[1468]: E20241213 13:26:48.391794 1468 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Dec 13 13:26:48.392160 update_engine[1468]: I20241213 13:26:48.391870 1468 libcurl_http_fetcher.cc:283] No HTTP response, retry 1 Dec 13 13:26:48.392544 locksmithd[1498]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_CHECKING_FOR_UPDATE" NewVersion=0.0.0 NewSize=0 Dec 13 13:26:53.557241 systemd[1]: Started sshd@7-23.88.121.109:22-147.75.109.163:34532.service - OpenSSH per-connection server daemon (147.75.109.163:34532). Dec 13 13:26:54.555657 sshd[4235]: Accepted publickey for core from 147.75.109.163 port 34532 ssh2: RSA SHA256:mujnIWCHI6g7RhOIFUnU4L23td/nvZma7inzBKSUIRw Dec 13 13:26:54.557709 sshd-session[4235]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 13:26:54.563630 systemd-logind[1465]: New session 8 of user core. Dec 13 13:26:54.570036 systemd[1]: Started session-8.scope - Session 8 of User core. Dec 13 13:26:55.331193 sshd[4237]: Connection closed by 147.75.109.163 port 34532 Dec 13 13:26:55.332302 sshd-session[4235]: pam_unix(sshd:session): session closed for user core Dec 13 13:26:55.340390 systemd-logind[1465]: Session 8 logged out. Waiting for processes to exit. Dec 13 13:26:55.341181 systemd[1]: sshd@7-23.88.121.109:22-147.75.109.163:34532.service: Deactivated successfully. Dec 13 13:26:55.344445 systemd[1]: session-8.scope: Deactivated successfully. Dec 13 13:26:55.346110 systemd-logind[1465]: Removed session 8. Dec 13 13:26:58.288863 update_engine[1468]: I20241213 13:26:58.288295 1468 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Dec 13 13:26:58.288863 update_engine[1468]: I20241213 13:26:58.288671 1468 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Dec 13 13:26:58.289395 update_engine[1468]: I20241213 13:26:58.289095 1468 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Dec 13 13:26:58.289655 update_engine[1468]: E20241213 13:26:58.289592 1468 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Dec 13 13:26:58.289782 update_engine[1468]: I20241213 13:26:58.289672 1468 libcurl_http_fetcher.cc:283] No HTTP response, retry 2 Dec 13 13:27:00.507452 systemd[1]: Started sshd@8-23.88.121.109:22-147.75.109.163:59960.service - OpenSSH per-connection server daemon (147.75.109.163:59960). Dec 13 13:27:01.493197 sshd[4249]: Accepted publickey for core from 147.75.109.163 port 59960 ssh2: RSA SHA256:mujnIWCHI6g7RhOIFUnU4L23td/nvZma7inzBKSUIRw Dec 13 13:27:01.495690 sshd-session[4249]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 13:27:01.500906 systemd-logind[1465]: New session 9 of user core. Dec 13 13:27:01.505993 systemd[1]: Started session-9.scope - Session 9 of User core. Dec 13 13:27:02.258377 sshd[4251]: Connection closed by 147.75.109.163 port 59960 Dec 13 13:27:02.259279 sshd-session[4249]: pam_unix(sshd:session): session closed for user core Dec 13 13:27:02.265347 systemd[1]: sshd@8-23.88.121.109:22-147.75.109.163:59960.service: Deactivated successfully. Dec 13 13:27:02.269330 systemd[1]: session-9.scope: Deactivated successfully. Dec 13 13:27:02.271166 systemd-logind[1465]: Session 9 logged out. Waiting for processes to exit. Dec 13 13:27:02.273316 systemd-logind[1465]: Removed session 9. Dec 13 13:27:07.432859 systemd[1]: Started sshd@9-23.88.121.109:22-147.75.109.163:54546.service - OpenSSH per-connection server daemon (147.75.109.163:54546). Dec 13 13:27:08.291285 update_engine[1468]: I20241213 13:27:08.291180 1468 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Dec 13 13:27:08.291811 update_engine[1468]: I20241213 13:27:08.291451 1468 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Dec 13 13:27:08.291811 update_engine[1468]: I20241213 13:27:08.291782 1468 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Dec 13 13:27:08.292330 update_engine[1468]: E20241213 13:27:08.292253 1468 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Dec 13 13:27:08.292330 update_engine[1468]: I20241213 13:27:08.292333 1468 libcurl_http_fetcher.cc:283] No HTTP response, retry 3 Dec 13 13:27:08.418453 sshd[4264]: Accepted publickey for core from 147.75.109.163 port 54546 ssh2: RSA SHA256:mujnIWCHI6g7RhOIFUnU4L23td/nvZma7inzBKSUIRw Dec 13 13:27:08.420481 sshd-session[4264]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 13:27:08.425376 systemd-logind[1465]: New session 10 of user core. Dec 13 13:27:08.429988 systemd[1]: Started session-10.scope - Session 10 of User core. Dec 13 13:27:09.179005 sshd[4266]: Connection closed by 147.75.109.163 port 54546 Dec 13 13:27:09.179953 sshd-session[4264]: pam_unix(sshd:session): session closed for user core Dec 13 13:27:09.186362 systemd-logind[1465]: Session 10 logged out. Waiting for processes to exit. Dec 13 13:27:09.186988 systemd[1]: sshd@9-23.88.121.109:22-147.75.109.163:54546.service: Deactivated successfully. Dec 13 13:27:09.189215 systemd[1]: session-10.scope: Deactivated successfully. Dec 13 13:27:09.190489 systemd-logind[1465]: Removed session 10. Dec 13 13:27:14.358194 systemd[1]: Started sshd@10-23.88.121.109:22-147.75.109.163:54560.service - OpenSSH per-connection server daemon (147.75.109.163:54560). Dec 13 13:27:15.355093 sshd[4278]: Accepted publickey for core from 147.75.109.163 port 54560 ssh2: RSA SHA256:mujnIWCHI6g7RhOIFUnU4L23td/nvZma7inzBKSUIRw Dec 13 13:27:15.357200 sshd-session[4278]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 13:27:15.363018 systemd-logind[1465]: New session 11 of user core. Dec 13 13:27:15.369065 systemd[1]: Started session-11.scope - Session 11 of User core. Dec 13 13:27:16.124501 sshd[4280]: Connection closed by 147.75.109.163 port 54560 Dec 13 13:27:16.125577 sshd-session[4278]: pam_unix(sshd:session): session closed for user core Dec 13 13:27:16.129899 systemd[1]: sshd@10-23.88.121.109:22-147.75.109.163:54560.service: Deactivated successfully. Dec 13 13:27:16.133332 systemd[1]: session-11.scope: Deactivated successfully. Dec 13 13:27:16.138990 systemd-logind[1465]: Session 11 logged out. Waiting for processes to exit. Dec 13 13:27:16.140976 systemd-logind[1465]: Removed session 11. Dec 13 13:27:16.301242 systemd[1]: Started sshd@11-23.88.121.109:22-147.75.109.163:54984.service - OpenSSH per-connection server daemon (147.75.109.163:54984). Dec 13 13:27:17.295376 sshd[4291]: Accepted publickey for core from 147.75.109.163 port 54984 ssh2: RSA SHA256:mujnIWCHI6g7RhOIFUnU4L23td/nvZma7inzBKSUIRw Dec 13 13:27:17.297347 sshd-session[4291]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 13:27:17.302279 systemd-logind[1465]: New session 12 of user core. Dec 13 13:27:17.308021 systemd[1]: Started session-12.scope - Session 12 of User core. Dec 13 13:27:18.114359 sshd[4293]: Connection closed by 147.75.109.163 port 54984 Dec 13 13:27:18.115035 sshd-session[4291]: pam_unix(sshd:session): session closed for user core Dec 13 13:27:18.118390 systemd[1]: sshd@11-23.88.121.109:22-147.75.109.163:54984.service: Deactivated successfully. Dec 13 13:27:18.121539 systemd[1]: session-12.scope: Deactivated successfully. Dec 13 13:27:18.123656 systemd-logind[1465]: Session 12 logged out. Waiting for processes to exit. Dec 13 13:27:18.124687 systemd-logind[1465]: Removed session 12. Dec 13 13:27:18.286520 update_engine[1468]: I20241213 13:27:18.286309 1468 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Dec 13 13:27:18.288209 update_engine[1468]: I20241213 13:27:18.287102 1468 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Dec 13 13:27:18.288209 update_engine[1468]: I20241213 13:27:18.287319 1468 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Dec 13 13:27:18.288812 update_engine[1468]: E20241213 13:27:18.288344 1468 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Dec 13 13:27:18.288812 update_engine[1468]: I20241213 13:27:18.288409 1468 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded Dec 13 13:27:18.288812 update_engine[1468]: I20241213 13:27:18.288417 1468 omaha_request_action.cc:617] Omaha request response: Dec 13 13:27:18.288812 update_engine[1468]: E20241213 13:27:18.288499 1468 omaha_request_action.cc:636] Omaha request network transfer failed. Dec 13 13:27:18.288812 update_engine[1468]: I20241213 13:27:18.288516 1468 action_processor.cc:68] ActionProcessor::ActionComplete: OmahaRequestAction action failed. Aborting processing. Dec 13 13:27:18.288812 update_engine[1468]: I20241213 13:27:18.288522 1468 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Dec 13 13:27:18.288812 update_engine[1468]: I20241213 13:27:18.288525 1468 update_attempter.cc:306] Processing Done. Dec 13 13:27:18.288812 update_engine[1468]: E20241213 13:27:18.288539 1468 update_attempter.cc:619] Update failed. Dec 13 13:27:18.288812 update_engine[1468]: I20241213 13:27:18.288544 1468 utils.cc:600] Converting error code 2000 to kActionCodeOmahaErrorInHTTPResponse Dec 13 13:27:18.288812 update_engine[1468]: I20241213 13:27:18.288548 1468 payload_state.cc:97] Updating payload state for error code: 37 (kActionCodeOmahaErrorInHTTPResponse) Dec 13 13:27:18.288812 update_engine[1468]: I20241213 13:27:18.288553 1468 payload_state.cc:103] Ignoring failures until we get a valid Omaha response. Dec 13 13:27:18.288812 update_engine[1468]: I20241213 13:27:18.288614 1468 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction Dec 13 13:27:18.288812 update_engine[1468]: I20241213 13:27:18.288635 1468 omaha_request_action.cc:271] Posting an Omaha request to disabled Dec 13 13:27:18.288812 update_engine[1468]: I20241213 13:27:18.288640 1468 omaha_request_action.cc:272] Request: Dec 13 13:27:18.288812 update_engine[1468]: Dec 13 13:27:18.288812 update_engine[1468]: Dec 13 13:27:18.288812 update_engine[1468]: Dec 13 13:27:18.289154 update_engine[1468]: Dec 13 13:27:18.289154 update_engine[1468]: Dec 13 13:27:18.289154 update_engine[1468]: Dec 13 13:27:18.289154 update_engine[1468]: I20241213 13:27:18.288645 1468 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Dec 13 13:27:18.289511 update_engine[1468]: I20241213 13:27:18.289278 1468 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Dec 13 13:27:18.289511 update_engine[1468]: I20241213 13:27:18.289475 1468 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Dec 13 13:27:18.290096 update_engine[1468]: E20241213 13:27:18.289907 1468 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Dec 13 13:27:18.290096 update_engine[1468]: I20241213 13:27:18.289952 1468 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded Dec 13 13:27:18.290096 update_engine[1468]: I20241213 13:27:18.289958 1468 omaha_request_action.cc:617] Omaha request response: Dec 13 13:27:18.290096 update_engine[1468]: I20241213 13:27:18.289964 1468 action_processor.cc:65] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Dec 13 13:27:18.290096 update_engine[1468]: I20241213 13:27:18.289969 1468 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Dec 13 13:27:18.290096 update_engine[1468]: I20241213 13:27:18.289974 1468 update_attempter.cc:306] Processing Done. Dec 13 13:27:18.290096 update_engine[1468]: I20241213 13:27:18.289979 1468 update_attempter.cc:310] Error event sent. Dec 13 13:27:18.290096 update_engine[1468]: I20241213 13:27:18.289988 1468 update_check_scheduler.cc:74] Next update check in 40m50s Dec 13 13:27:18.290415 locksmithd[1498]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_REPORTING_ERROR_EVENT" NewVersion=0.0.0 NewSize=0 Dec 13 13:27:18.290415 locksmithd[1498]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_IDLE" NewVersion=0.0.0 NewSize=0 Dec 13 13:27:18.295271 systemd[1]: Started sshd@12-23.88.121.109:22-147.75.109.163:54990.service - OpenSSH per-connection server daemon (147.75.109.163:54990). Dec 13 13:27:19.282188 sshd[4302]: Accepted publickey for core from 147.75.109.163 port 54990 ssh2: RSA SHA256:mujnIWCHI6g7RhOIFUnU4L23td/nvZma7inzBKSUIRw Dec 13 13:27:19.284398 sshd-session[4302]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 13:27:19.289498 systemd-logind[1465]: New session 13 of user core. Dec 13 13:27:19.296005 systemd[1]: Started session-13.scope - Session 13 of User core. Dec 13 13:27:20.037625 sshd[4304]: Connection closed by 147.75.109.163 port 54990 Dec 13 13:27:20.038668 sshd-session[4302]: pam_unix(sshd:session): session closed for user core Dec 13 13:27:20.044351 systemd-logind[1465]: Session 13 logged out. Waiting for processes to exit. Dec 13 13:27:20.045298 systemd[1]: sshd@12-23.88.121.109:22-147.75.109.163:54990.service: Deactivated successfully. Dec 13 13:27:20.047811 systemd[1]: session-13.scope: Deactivated successfully. Dec 13 13:27:20.050063 systemd-logind[1465]: Removed session 13. Dec 13 13:27:25.212470 systemd[1]: Started sshd@13-23.88.121.109:22-147.75.109.163:55006.service - OpenSSH per-connection server daemon (147.75.109.163:55006). Dec 13 13:27:26.192141 sshd[4317]: Accepted publickey for core from 147.75.109.163 port 55006 ssh2: RSA SHA256:mujnIWCHI6g7RhOIFUnU4L23td/nvZma7inzBKSUIRw Dec 13 13:27:26.195512 sshd-session[4317]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 13:27:26.200164 systemd-logind[1465]: New session 14 of user core. Dec 13 13:27:26.207477 systemd[1]: Started session-14.scope - Session 14 of User core. Dec 13 13:27:26.940703 sshd[4319]: Connection closed by 147.75.109.163 port 55006 Dec 13 13:27:26.941684 sshd-session[4317]: pam_unix(sshd:session): session closed for user core Dec 13 13:27:26.946679 systemd[1]: sshd@13-23.88.121.109:22-147.75.109.163:55006.service: Deactivated successfully. Dec 13 13:27:26.950256 systemd[1]: session-14.scope: Deactivated successfully. Dec 13 13:27:26.951192 systemd-logind[1465]: Session 14 logged out. Waiting for processes to exit. Dec 13 13:27:26.952251 systemd-logind[1465]: Removed session 14. Dec 13 13:27:27.114053 systemd[1]: Started sshd@14-23.88.121.109:22-147.75.109.163:58184.service - OpenSSH per-connection server daemon (147.75.109.163:58184). Dec 13 13:27:28.093616 sshd[4330]: Accepted publickey for core from 147.75.109.163 port 58184 ssh2: RSA SHA256:mujnIWCHI6g7RhOIFUnU4L23td/nvZma7inzBKSUIRw Dec 13 13:27:28.095303 sshd-session[4330]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 13:27:28.102057 systemd-logind[1465]: New session 15 of user core. Dec 13 13:27:28.112327 systemd[1]: Started session-15.scope - Session 15 of User core. Dec 13 13:27:28.887127 sshd[4332]: Connection closed by 147.75.109.163 port 58184 Dec 13 13:27:28.886069 sshd-session[4330]: pam_unix(sshd:session): session closed for user core Dec 13 13:27:28.891038 systemd[1]: sshd@14-23.88.121.109:22-147.75.109.163:58184.service: Deactivated successfully. Dec 13 13:27:28.893238 systemd[1]: session-15.scope: Deactivated successfully. Dec 13 13:27:28.895690 systemd-logind[1465]: Session 15 logged out. Waiting for processes to exit. Dec 13 13:27:28.897006 systemd-logind[1465]: Removed session 15. Dec 13 13:27:29.059170 systemd[1]: Started sshd@15-23.88.121.109:22-147.75.109.163:58194.service - OpenSSH per-connection server daemon (147.75.109.163:58194). Dec 13 13:27:30.041123 sshd[4341]: Accepted publickey for core from 147.75.109.163 port 58194 ssh2: RSA SHA256:mujnIWCHI6g7RhOIFUnU4L23td/nvZma7inzBKSUIRw Dec 13 13:27:30.043200 sshd-session[4341]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 13:27:30.052428 systemd-logind[1465]: New session 16 of user core. Dec 13 13:27:30.058117 systemd[1]: Started session-16.scope - Session 16 of User core. Dec 13 13:27:32.414711 sshd[4343]: Connection closed by 147.75.109.163 port 58194 Dec 13 13:27:32.415204 sshd-session[4341]: pam_unix(sshd:session): session closed for user core Dec 13 13:27:32.420683 systemd[1]: sshd@15-23.88.121.109:22-147.75.109.163:58194.service: Deactivated successfully. Dec 13 13:27:32.425243 systemd[1]: session-16.scope: Deactivated successfully. Dec 13 13:27:32.426205 systemd-logind[1465]: Session 16 logged out. Waiting for processes to exit. Dec 13 13:27:32.427257 systemd-logind[1465]: Removed session 16. Dec 13 13:27:32.592279 systemd[1]: Started sshd@16-23.88.121.109:22-147.75.109.163:58196.service - OpenSSH per-connection server daemon (147.75.109.163:58196). Dec 13 13:27:33.576510 sshd[4359]: Accepted publickey for core from 147.75.109.163 port 58196 ssh2: RSA SHA256:mujnIWCHI6g7RhOIFUnU4L23td/nvZma7inzBKSUIRw Dec 13 13:27:33.578835 sshd-session[4359]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 13:27:33.583185 systemd-logind[1465]: New session 17 of user core. Dec 13 13:27:33.587949 systemd[1]: Started session-17.scope - Session 17 of User core. Dec 13 13:27:34.447804 sshd[4361]: Connection closed by 147.75.109.163 port 58196 Dec 13 13:27:34.447638 sshd-session[4359]: pam_unix(sshd:session): session closed for user core Dec 13 13:27:34.453601 systemd[1]: sshd@16-23.88.121.109:22-147.75.109.163:58196.service: Deactivated successfully. Dec 13 13:27:34.455894 systemd[1]: session-17.scope: Deactivated successfully. Dec 13 13:27:34.457513 systemd-logind[1465]: Session 17 logged out. Waiting for processes to exit. Dec 13 13:27:34.458460 systemd-logind[1465]: Removed session 17. Dec 13 13:27:34.627343 systemd[1]: Started sshd@17-23.88.121.109:22-147.75.109.163:58200.service - OpenSSH per-connection server daemon (147.75.109.163:58200). Dec 13 13:27:35.613328 sshd[4370]: Accepted publickey for core from 147.75.109.163 port 58200 ssh2: RSA SHA256:mujnIWCHI6g7RhOIFUnU4L23td/nvZma7inzBKSUIRw Dec 13 13:27:35.615275 sshd-session[4370]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 13:27:35.623965 systemd-logind[1465]: New session 18 of user core. Dec 13 13:27:35.628940 systemd[1]: Started session-18.scope - Session 18 of User core. Dec 13 13:27:36.373055 sshd[4372]: Connection closed by 147.75.109.163 port 58200 Dec 13 13:27:36.373961 sshd-session[4370]: pam_unix(sshd:session): session closed for user core Dec 13 13:27:36.378662 systemd[1]: sshd@17-23.88.121.109:22-147.75.109.163:58200.service: Deactivated successfully. Dec 13 13:27:36.381775 systemd[1]: session-18.scope: Deactivated successfully. Dec 13 13:27:36.384567 systemd-logind[1465]: Session 18 logged out. Waiting for processes to exit. Dec 13 13:27:36.385886 systemd-logind[1465]: Removed session 18. Dec 13 13:27:41.549163 systemd[1]: Started sshd@18-23.88.121.109:22-147.75.109.163:41278.service - OpenSSH per-connection server daemon (147.75.109.163:41278). Dec 13 13:27:42.532980 sshd[4387]: Accepted publickey for core from 147.75.109.163 port 41278 ssh2: RSA SHA256:mujnIWCHI6g7RhOIFUnU4L23td/nvZma7inzBKSUIRw Dec 13 13:27:42.535245 sshd-session[4387]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 13:27:42.545257 systemd-logind[1465]: New session 19 of user core. Dec 13 13:27:42.549031 systemd[1]: Started session-19.scope - Session 19 of User core. Dec 13 13:27:43.285429 sshd[4389]: Connection closed by 147.75.109.163 port 41278 Dec 13 13:27:43.286480 sshd-session[4387]: pam_unix(sshd:session): session closed for user core Dec 13 13:27:43.290336 systemd[1]: sshd@18-23.88.121.109:22-147.75.109.163:41278.service: Deactivated successfully. Dec 13 13:27:43.292564 systemd[1]: session-19.scope: Deactivated successfully. Dec 13 13:27:43.294329 systemd-logind[1465]: Session 19 logged out. Waiting for processes to exit. Dec 13 13:27:43.296122 systemd-logind[1465]: Removed session 19. Dec 13 13:27:48.465148 systemd[1]: Started sshd@19-23.88.121.109:22-147.75.109.163:59288.service - OpenSSH per-connection server daemon (147.75.109.163:59288). Dec 13 13:27:49.449492 sshd[4402]: Accepted publickey for core from 147.75.109.163 port 59288 ssh2: RSA SHA256:mujnIWCHI6g7RhOIFUnU4L23td/nvZma7inzBKSUIRw Dec 13 13:27:49.451663 sshd-session[4402]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 13:27:49.457681 systemd-logind[1465]: New session 20 of user core. Dec 13 13:27:49.464402 systemd[1]: Started session-20.scope - Session 20 of User core. Dec 13 13:27:50.200300 sshd[4404]: Connection closed by 147.75.109.163 port 59288 Dec 13 13:27:50.199583 sshd-session[4402]: pam_unix(sshd:session): session closed for user core Dec 13 13:27:50.204624 systemd[1]: sshd@19-23.88.121.109:22-147.75.109.163:59288.service: Deactivated successfully. Dec 13 13:27:50.208494 systemd[1]: session-20.scope: Deactivated successfully. Dec 13 13:27:50.209546 systemd-logind[1465]: Session 20 logged out. Waiting for processes to exit. Dec 13 13:27:50.211232 systemd-logind[1465]: Removed session 20. Dec 13 13:27:50.380237 systemd[1]: Started sshd@20-23.88.121.109:22-147.75.109.163:59298.service - OpenSSH per-connection server daemon (147.75.109.163:59298). Dec 13 13:27:51.373247 sshd[4417]: Accepted publickey for core from 147.75.109.163 port 59298 ssh2: RSA SHA256:mujnIWCHI6g7RhOIFUnU4L23td/nvZma7inzBKSUIRw Dec 13 13:27:51.375293 sshd-session[4417]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 13:27:51.381573 systemd-logind[1465]: New session 21 of user core. Dec 13 13:27:51.389140 systemd[1]: Started session-21.scope - Session 21 of User core. Dec 13 13:27:53.689572 systemd[1]: run-containerd-runc-k8s.io-7eef8c509f1c6c520e52af996fe6b3e92243ab283e5e5a2ee093a1c37553e171-runc.41oPlU.mount: Deactivated successfully. Dec 13 13:27:53.700076 containerd[1477]: time="2024-12-13T13:27:53.699136953Z" level=info msg="StopContainer for \"9fc7582bf65547c02d26deb9fc8a19b164c9cf14800312e5ddcac69a4829f858\" with timeout 30 (s)" Dec 13 13:27:53.702454 containerd[1477]: time="2024-12-13T13:27:53.701958609Z" level=info msg="Stop container \"9fc7582bf65547c02d26deb9fc8a19b164c9cf14800312e5ddcac69a4829f858\" with signal terminated" Dec 13 13:27:53.706782 containerd[1477]: time="2024-12-13T13:27:53.706050154Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Dec 13 13:27:53.717482 containerd[1477]: time="2024-12-13T13:27:53.717347861Z" level=info msg="StopContainer for \"7eef8c509f1c6c520e52af996fe6b3e92243ab283e5e5a2ee093a1c37553e171\" with timeout 2 (s)" Dec 13 13:27:53.717857 systemd[1]: cri-containerd-9fc7582bf65547c02d26deb9fc8a19b164c9cf14800312e5ddcac69a4829f858.scope: Deactivated successfully. Dec 13 13:27:53.720625 containerd[1477]: time="2024-12-13T13:27:53.719854675Z" level=info msg="Stop container \"7eef8c509f1c6c520e52af996fe6b3e92243ab283e5e5a2ee093a1c37553e171\" with signal terminated" Dec 13 13:27:53.732446 systemd-networkd[1383]: lxc_health: Link DOWN Dec 13 13:27:53.732453 systemd-networkd[1383]: lxc_health: Lost carrier Dec 13 13:27:53.757853 systemd[1]: cri-containerd-7eef8c509f1c6c520e52af996fe6b3e92243ab283e5e5a2ee093a1c37553e171.scope: Deactivated successfully. Dec 13 13:27:53.758514 systemd[1]: cri-containerd-7eef8c509f1c6c520e52af996fe6b3e92243ab283e5e5a2ee093a1c37553e171.scope: Consumed 7.881s CPU time. Dec 13 13:27:53.764394 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-9fc7582bf65547c02d26deb9fc8a19b164c9cf14800312e5ddcac69a4829f858-rootfs.mount: Deactivated successfully. Dec 13 13:27:53.784920 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-7eef8c509f1c6c520e52af996fe6b3e92243ab283e5e5a2ee093a1c37553e171-rootfs.mount: Deactivated successfully. Dec 13 13:27:53.799229 containerd[1477]: time="2024-12-13T13:27:53.799128824Z" level=info msg="shim disconnected" id=9fc7582bf65547c02d26deb9fc8a19b164c9cf14800312e5ddcac69a4829f858 namespace=k8s.io Dec 13 13:27:53.799469 containerd[1477]: time="2024-12-13T13:27:53.799233665Z" level=warning msg="cleaning up after shim disconnected" id=9fc7582bf65547c02d26deb9fc8a19b164c9cf14800312e5ddcac69a4829f858 namespace=k8s.io Dec 13 13:27:53.799469 containerd[1477]: time="2024-12-13T13:27:53.799257585Z" level=info msg="cleaning up dead shim" namespace=k8s.io Dec 13 13:27:53.799889 containerd[1477]: time="2024-12-13T13:27:53.799642388Z" level=info msg="shim disconnected" id=7eef8c509f1c6c520e52af996fe6b3e92243ab283e5e5a2ee093a1c37553e171 namespace=k8s.io Dec 13 13:27:53.799889 containerd[1477]: time="2024-12-13T13:27:53.799681148Z" level=warning msg="cleaning up after shim disconnected" id=7eef8c509f1c6c520e52af996fe6b3e92243ab283e5e5a2ee093a1c37553e171 namespace=k8s.io Dec 13 13:27:53.799889 containerd[1477]: time="2024-12-13T13:27:53.799689628Z" level=info msg="cleaning up dead shim" namespace=k8s.io Dec 13 13:27:53.824503 containerd[1477]: time="2024-12-13T13:27:53.824456974Z" level=warning msg="cleanup warnings time=\"2024-12-13T13:27:53Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Dec 13 13:27:53.836660 containerd[1477]: time="2024-12-13T13:27:53.836348765Z" level=info msg="StopContainer for \"9fc7582bf65547c02d26deb9fc8a19b164c9cf14800312e5ddcac69a4829f858\" returns successfully" Dec 13 13:27:53.838097 containerd[1477]: time="2024-12-13T13:27:53.837850254Z" level=info msg="StopContainer for \"7eef8c509f1c6c520e52af996fe6b3e92243ab283e5e5a2ee093a1c37553e171\" returns successfully" Dec 13 13:27:53.838658 containerd[1477]: time="2024-12-13T13:27:53.838394737Z" level=info msg="StopPodSandbox for \"ae6a1363bbce38682e7e9f82b14004da2fb101205b6d8d6eb3974ea986c256a3\"" Dec 13 13:27:53.838658 containerd[1477]: time="2024-12-13T13:27:53.838460977Z" level=info msg="Container to stop \"9fc7582bf65547c02d26deb9fc8a19b164c9cf14800312e5ddcac69a4829f858\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Dec 13 13:27:53.840481 containerd[1477]: time="2024-12-13T13:27:53.839123061Z" level=info msg="StopPodSandbox for \"08a2754e454ae96132c191313465a0ff2e9480b13d88079a4a438ec4f2571f81\"" Dec 13 13:27:53.840481 containerd[1477]: time="2024-12-13T13:27:53.839166981Z" level=info msg="Container to stop \"7eef8c509f1c6c520e52af996fe6b3e92243ab283e5e5a2ee093a1c37553e171\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Dec 13 13:27:53.840481 containerd[1477]: time="2024-12-13T13:27:53.839183301Z" level=info msg="Container to stop \"36c7bf433ad4194395ef0cec0b50e5506f334bf0955a6a0cea088d55207c9053\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Dec 13 13:27:53.840481 containerd[1477]: time="2024-12-13T13:27:53.839197982Z" level=info msg="Container to stop \"590a9dd551ac1e339d1e595c191fa336cd7d22786ab1da6d8b47b543c8234a0b\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Dec 13 13:27:53.840481 containerd[1477]: time="2024-12-13T13:27:53.839211102Z" level=info msg="Container to stop \"ca0716e2d915f9418f463fc7052030f47974ebb73c009c58197e2bed939f0016\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Dec 13 13:27:53.840481 containerd[1477]: time="2024-12-13T13:27:53.839224822Z" level=info msg="Container to stop \"28003be96e8c1eaa0f2d3a766ba08814c4e37ed4b1a896dccf7c066b098fd8a0\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Dec 13 13:27:53.842231 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-ae6a1363bbce38682e7e9f82b14004da2fb101205b6d8d6eb3974ea986c256a3-shm.mount: Deactivated successfully. Dec 13 13:27:53.849879 systemd[1]: cri-containerd-ae6a1363bbce38682e7e9f82b14004da2fb101205b6d8d6eb3974ea986c256a3.scope: Deactivated successfully. Dec 13 13:27:53.854384 systemd[1]: cri-containerd-08a2754e454ae96132c191313465a0ff2e9480b13d88079a4a438ec4f2571f81.scope: Deactivated successfully. Dec 13 13:27:53.886361 containerd[1477]: time="2024-12-13T13:27:53.886132619Z" level=info msg="shim disconnected" id=ae6a1363bbce38682e7e9f82b14004da2fb101205b6d8d6eb3974ea986c256a3 namespace=k8s.io Dec 13 13:27:53.886361 containerd[1477]: time="2024-12-13T13:27:53.886185460Z" level=warning msg="cleaning up after shim disconnected" id=ae6a1363bbce38682e7e9f82b14004da2fb101205b6d8d6eb3974ea986c256a3 namespace=k8s.io Dec 13 13:27:53.886361 containerd[1477]: time="2024-12-13T13:27:53.886192980Z" level=info msg="cleaning up dead shim" namespace=k8s.io Dec 13 13:27:53.887101 containerd[1477]: time="2024-12-13T13:27:53.886900824Z" level=info msg="shim disconnected" id=08a2754e454ae96132c191313465a0ff2e9480b13d88079a4a438ec4f2571f81 namespace=k8s.io Dec 13 13:27:53.887101 containerd[1477]: time="2024-12-13T13:27:53.886932384Z" level=warning msg="cleaning up after shim disconnected" id=08a2754e454ae96132c191313465a0ff2e9480b13d88079a4a438ec4f2571f81 namespace=k8s.io Dec 13 13:27:53.887101 containerd[1477]: time="2024-12-13T13:27:53.886939304Z" level=info msg="cleaning up dead shim" namespace=k8s.io Dec 13 13:27:53.904715 containerd[1477]: time="2024-12-13T13:27:53.904670249Z" level=warning msg="cleanup warnings time=\"2024-12-13T13:27:53Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Dec 13 13:27:53.905818 containerd[1477]: time="2024-12-13T13:27:53.905708855Z" level=info msg="TearDown network for sandbox \"ae6a1363bbce38682e7e9f82b14004da2fb101205b6d8d6eb3974ea986c256a3\" successfully" Dec 13 13:27:53.905818 containerd[1477]: time="2024-12-13T13:27:53.905802536Z" level=info msg="StopPodSandbox for \"ae6a1363bbce38682e7e9f82b14004da2fb101205b6d8d6eb3974ea986c256a3\" returns successfully" Dec 13 13:27:53.906764 kubelet[2824]: E1213 13:27:53.906469 2824 kubelet.go:2900] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Dec 13 13:27:53.907084 containerd[1477]: time="2024-12-13T13:27:53.906605980Z" level=info msg="TearDown network for sandbox \"08a2754e454ae96132c191313465a0ff2e9480b13d88079a4a438ec4f2571f81\" successfully" Dec 13 13:27:53.907084 containerd[1477]: time="2024-12-13T13:27:53.906628301Z" level=info msg="StopPodSandbox for \"08a2754e454ae96132c191313465a0ff2e9480b13d88079a4a438ec4f2571f81\" returns successfully" Dec 13 13:27:54.083265 kubelet[2824]: I1213 13:27:54.083063 2824 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-cilium-run\") pod \"87af4ba6-4abb-49f4-8748-18d3ba478286\" (UID: \"87af4ba6-4abb-49f4-8748-18d3ba478286\") " Dec 13 13:27:54.083265 kubelet[2824]: I1213 13:27:54.083119 2824 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-bpf-maps\") pod \"87af4ba6-4abb-49f4-8748-18d3ba478286\" (UID: \"87af4ba6-4abb-49f4-8748-18d3ba478286\") " Dec 13 13:27:54.083265 kubelet[2824]: I1213 13:27:54.083147 2824 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/87af4ba6-4abb-49f4-8748-18d3ba478286-cilium-config-path\") pod \"87af4ba6-4abb-49f4-8748-18d3ba478286\" (UID: \"87af4ba6-4abb-49f4-8748-18d3ba478286\") " Dec 13 13:27:54.083265 kubelet[2824]: I1213 13:27:54.083170 2824 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-host-proc-sys-net\") pod \"87af4ba6-4abb-49f4-8748-18d3ba478286\" (UID: \"87af4ba6-4abb-49f4-8748-18d3ba478286\") " Dec 13 13:27:54.083265 kubelet[2824]: I1213 13:27:54.083193 2824 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/d2fb0668-07bd-4a53-9b0f-ec12c129578d-cilium-config-path\") pod \"d2fb0668-07bd-4a53-9b0f-ec12c129578d\" (UID: \"d2fb0668-07bd-4a53-9b0f-ec12c129578d\") " Dec 13 13:27:54.083265 kubelet[2824]: I1213 13:27:54.083217 2824 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmf64\" (UniqueName: \"kubernetes.io/projected/87af4ba6-4abb-49f4-8748-18d3ba478286-kube-api-access-tmf64\") pod \"87af4ba6-4abb-49f4-8748-18d3ba478286\" (UID: \"87af4ba6-4abb-49f4-8748-18d3ba478286\") " Dec 13 13:27:54.084155 kubelet[2824]: I1213 13:27:54.083235 2824 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-lib-modules\") pod \"87af4ba6-4abb-49f4-8748-18d3ba478286\" (UID: \"87af4ba6-4abb-49f4-8748-18d3ba478286\") " Dec 13 13:27:54.084155 kubelet[2824]: I1213 13:27:54.083268 2824 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-etc-cni-netd\") pod \"87af4ba6-4abb-49f4-8748-18d3ba478286\" (UID: \"87af4ba6-4abb-49f4-8748-18d3ba478286\") " Dec 13 13:27:54.084155 kubelet[2824]: I1213 13:27:54.083329 2824 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-host-proc-sys-kernel\") pod \"87af4ba6-4abb-49f4-8748-18d3ba478286\" (UID: \"87af4ba6-4abb-49f4-8748-18d3ba478286\") " Dec 13 13:27:54.084155 kubelet[2824]: I1213 13:27:54.083361 2824 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/87af4ba6-4abb-49f4-8748-18d3ba478286-clustermesh-secrets\") pod \"87af4ba6-4abb-49f4-8748-18d3ba478286\" (UID: \"87af4ba6-4abb-49f4-8748-18d3ba478286\") " Dec 13 13:27:54.084155 kubelet[2824]: I1213 13:27:54.083379 2824 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-xtables-lock\") pod \"87af4ba6-4abb-49f4-8748-18d3ba478286\" (UID: \"87af4ba6-4abb-49f4-8748-18d3ba478286\") " Dec 13 13:27:54.084155 kubelet[2824]: I1213 13:27:54.083398 2824 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/87af4ba6-4abb-49f4-8748-18d3ba478286-hubble-tls\") pod \"87af4ba6-4abb-49f4-8748-18d3ba478286\" (UID: \"87af4ba6-4abb-49f4-8748-18d3ba478286\") " Dec 13 13:27:54.084445 kubelet[2824]: I1213 13:27:54.083417 2824 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-cni-path\") pod \"87af4ba6-4abb-49f4-8748-18d3ba478286\" (UID: \"87af4ba6-4abb-49f4-8748-18d3ba478286\") " Dec 13 13:27:54.084445 kubelet[2824]: I1213 13:27:54.083445 2824 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-hostproc\") pod \"87af4ba6-4abb-49f4-8748-18d3ba478286\" (UID: \"87af4ba6-4abb-49f4-8748-18d3ba478286\") " Dec 13 13:27:54.084445 kubelet[2824]: I1213 13:27:54.083467 2824 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-cilium-cgroup\") pod \"87af4ba6-4abb-49f4-8748-18d3ba478286\" (UID: \"87af4ba6-4abb-49f4-8748-18d3ba478286\") " Dec 13 13:27:54.084445 kubelet[2824]: I1213 13:27:54.083494 2824 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9s9s7\" (UniqueName: \"kubernetes.io/projected/d2fb0668-07bd-4a53-9b0f-ec12c129578d-kube-api-access-9s9s7\") pod \"d2fb0668-07bd-4a53-9b0f-ec12c129578d\" (UID: \"d2fb0668-07bd-4a53-9b0f-ec12c129578d\") " Dec 13 13:27:54.088491 kubelet[2824]: I1213 13:27:54.087667 2824 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "87af4ba6-4abb-49f4-8748-18d3ba478286" (UID: "87af4ba6-4abb-49f4-8748-18d3ba478286"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 13:27:54.088491 kubelet[2824]: I1213 13:27:54.087768 2824 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "87af4ba6-4abb-49f4-8748-18d3ba478286" (UID: "87af4ba6-4abb-49f4-8748-18d3ba478286"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 13:27:54.088491 kubelet[2824]: I1213 13:27:54.088235 2824 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "87af4ba6-4abb-49f4-8748-18d3ba478286" (UID: "87af4ba6-4abb-49f4-8748-18d3ba478286"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 13:27:54.088491 kubelet[2824]: I1213 13:27:54.088276 2824 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "87af4ba6-4abb-49f4-8748-18d3ba478286" (UID: "87af4ba6-4abb-49f4-8748-18d3ba478286"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 13:27:54.088491 kubelet[2824]: I1213 13:27:54.088345 2824 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "87af4ba6-4abb-49f4-8748-18d3ba478286" (UID: "87af4ba6-4abb-49f4-8748-18d3ba478286"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 13:27:54.088772 kubelet[2824]: I1213 13:27:54.088661 2824 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "87af4ba6-4abb-49f4-8748-18d3ba478286" (UID: "87af4ba6-4abb-49f4-8748-18d3ba478286"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 13:27:54.090686 kubelet[2824]: I1213 13:27:54.090648 2824 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2fb0668-07bd-4a53-9b0f-ec12c129578d-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "d2fb0668-07bd-4a53-9b0f-ec12c129578d" (UID: "d2fb0668-07bd-4a53-9b0f-ec12c129578d"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 13:27:54.092184 kubelet[2824]: I1213 13:27:54.091085 2824 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "87af4ba6-4abb-49f4-8748-18d3ba478286" (UID: "87af4ba6-4abb-49f4-8748-18d3ba478286"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 13:27:54.092184 kubelet[2824]: I1213 13:27:54.091887 2824 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-cni-path" (OuterVolumeSpecName: "cni-path") pod "87af4ba6-4abb-49f4-8748-18d3ba478286" (UID: "87af4ba6-4abb-49f4-8748-18d3ba478286"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 13:27:54.092184 kubelet[2824]: I1213 13:27:54.091923 2824 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-hostproc" (OuterVolumeSpecName: "hostproc") pod "87af4ba6-4abb-49f4-8748-18d3ba478286" (UID: "87af4ba6-4abb-49f4-8748-18d3ba478286"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 13:27:54.092184 kubelet[2824]: I1213 13:27:54.091938 2824 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "87af4ba6-4abb-49f4-8748-18d3ba478286" (UID: "87af4ba6-4abb-49f4-8748-18d3ba478286"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 13:27:54.092184 kubelet[2824]: I1213 13:27:54.092019 2824 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87af4ba6-4abb-49f4-8748-18d3ba478286-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "87af4ba6-4abb-49f4-8748-18d3ba478286" (UID: "87af4ba6-4abb-49f4-8748-18d3ba478286"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 13:27:54.092401 kubelet[2824]: I1213 13:27:54.092078 2824 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2fb0668-07bd-4a53-9b0f-ec12c129578d-kube-api-access-9s9s7" (OuterVolumeSpecName: "kube-api-access-9s9s7") pod "d2fb0668-07bd-4a53-9b0f-ec12c129578d" (UID: "d2fb0668-07bd-4a53-9b0f-ec12c129578d"). InnerVolumeSpecName "kube-api-access-9s9s7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 13:27:54.092401 kubelet[2824]: I1213 13:27:54.092132 2824 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87af4ba6-4abb-49f4-8748-18d3ba478286-kube-api-access-tmf64" (OuterVolumeSpecName: "kube-api-access-tmf64") pod "87af4ba6-4abb-49f4-8748-18d3ba478286" (UID: "87af4ba6-4abb-49f4-8748-18d3ba478286"). InnerVolumeSpecName "kube-api-access-tmf64". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 13:27:54.093605 kubelet[2824]: I1213 13:27:54.093558 2824 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87af4ba6-4abb-49f4-8748-18d3ba478286-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "87af4ba6-4abb-49f4-8748-18d3ba478286" (UID: "87af4ba6-4abb-49f4-8748-18d3ba478286"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 13:27:54.094627 kubelet[2824]: I1213 13:27:54.094577 2824 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87af4ba6-4abb-49f4-8748-18d3ba478286-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "87af4ba6-4abb-49f4-8748-18d3ba478286" (UID: "87af4ba6-4abb-49f4-8748-18d3ba478286"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 13:27:54.184123 kubelet[2824]: I1213 13:27:54.184058 2824 reconciler_common.go:289] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/87af4ba6-4abb-49f4-8748-18d3ba478286-hubble-tls\") on node \"ci-4186-0-0-9-3a4b2091e9\" DevicePath \"\"" Dec 13 13:27:54.184123 kubelet[2824]: I1213 13:27:54.184116 2824 reconciler_common.go:289] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-cni-path\") on node \"ci-4186-0-0-9-3a4b2091e9\" DevicePath \"\"" Dec 13 13:27:54.184123 kubelet[2824]: I1213 13:27:54.184136 2824 reconciler_common.go:289] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-hostproc\") on node \"ci-4186-0-0-9-3a4b2091e9\" DevicePath \"\"" Dec 13 13:27:54.184477 kubelet[2824]: I1213 13:27:54.184157 2824 reconciler_common.go:289] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-cilium-cgroup\") on node \"ci-4186-0-0-9-3a4b2091e9\" DevicePath \"\"" Dec 13 13:27:54.184477 kubelet[2824]: I1213 13:27:54.184177 2824 reconciler_common.go:289] "Volume detached for volume \"kube-api-access-9s9s7\" (UniqueName: \"kubernetes.io/projected/d2fb0668-07bd-4a53-9b0f-ec12c129578d-kube-api-access-9s9s7\") on node \"ci-4186-0-0-9-3a4b2091e9\" DevicePath \"\"" Dec 13 13:27:54.184477 kubelet[2824]: I1213 13:27:54.184194 2824 reconciler_common.go:289] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/d2fb0668-07bd-4a53-9b0f-ec12c129578d-cilium-config-path\") on node \"ci-4186-0-0-9-3a4b2091e9\" DevicePath \"\"" Dec 13 13:27:54.184477 kubelet[2824]: I1213 13:27:54.184210 2824 reconciler_common.go:289] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-cilium-run\") on node \"ci-4186-0-0-9-3a4b2091e9\" DevicePath \"\"" Dec 13 13:27:54.184477 kubelet[2824]: I1213 13:27:54.184226 2824 reconciler_common.go:289] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-bpf-maps\") on node \"ci-4186-0-0-9-3a4b2091e9\" DevicePath \"\"" Dec 13 13:27:54.184477 kubelet[2824]: I1213 13:27:54.184243 2824 reconciler_common.go:289] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/87af4ba6-4abb-49f4-8748-18d3ba478286-cilium-config-path\") on node \"ci-4186-0-0-9-3a4b2091e9\" DevicePath \"\"" Dec 13 13:27:54.184477 kubelet[2824]: I1213 13:27:54.184258 2824 reconciler_common.go:289] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-host-proc-sys-net\") on node \"ci-4186-0-0-9-3a4b2091e9\" DevicePath \"\"" Dec 13 13:27:54.184477 kubelet[2824]: I1213 13:27:54.184275 2824 reconciler_common.go:289] "Volume detached for volume \"kube-api-access-tmf64\" (UniqueName: \"kubernetes.io/projected/87af4ba6-4abb-49f4-8748-18d3ba478286-kube-api-access-tmf64\") on node \"ci-4186-0-0-9-3a4b2091e9\" DevicePath \"\"" Dec 13 13:27:54.184834 kubelet[2824]: I1213 13:27:54.184319 2824 reconciler_common.go:289] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-lib-modules\") on node \"ci-4186-0-0-9-3a4b2091e9\" DevicePath \"\"" Dec 13 13:27:54.184834 kubelet[2824]: I1213 13:27:54.184337 2824 reconciler_common.go:289] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-etc-cni-netd\") on node \"ci-4186-0-0-9-3a4b2091e9\" DevicePath \"\"" Dec 13 13:27:54.184834 kubelet[2824]: I1213 13:27:54.184352 2824 reconciler_common.go:289] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-host-proc-sys-kernel\") on node \"ci-4186-0-0-9-3a4b2091e9\" DevicePath \"\"" Dec 13 13:27:54.184834 kubelet[2824]: I1213 13:27:54.184367 2824 reconciler_common.go:289] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/87af4ba6-4abb-49f4-8748-18d3ba478286-clustermesh-secrets\") on node \"ci-4186-0-0-9-3a4b2091e9\" DevicePath \"\"" Dec 13 13:27:54.184834 kubelet[2824]: I1213 13:27:54.184382 2824 reconciler_common.go:289] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/87af4ba6-4abb-49f4-8748-18d3ba478286-xtables-lock\") on node \"ci-4186-0-0-9-3a4b2091e9\" DevicePath \"\"" Dec 13 13:27:54.676789 kubelet[2824]: I1213 13:27:54.674832 2824 scope.go:117] "RemoveContainer" containerID="7eef8c509f1c6c520e52af996fe6b3e92243ab283e5e5a2ee093a1c37553e171" Dec 13 13:27:54.682610 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-08a2754e454ae96132c191313465a0ff2e9480b13d88079a4a438ec4f2571f81-rootfs.mount: Deactivated successfully. Dec 13 13:27:54.682720 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-08a2754e454ae96132c191313465a0ff2e9480b13d88079a4a438ec4f2571f81-shm.mount: Deactivated successfully. Dec 13 13:27:54.682819 systemd[1]: var-lib-kubelet-pods-87af4ba6\x2d4abb\x2d49f4\x2d8748\x2d18d3ba478286-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dtmf64.mount: Deactivated successfully. Dec 13 13:27:54.682877 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-ae6a1363bbce38682e7e9f82b14004da2fb101205b6d8d6eb3974ea986c256a3-rootfs.mount: Deactivated successfully. Dec 13 13:27:54.682924 systemd[1]: var-lib-kubelet-pods-d2fb0668\x2d07bd\x2d4a53\x2d9b0f\x2dec12c129578d-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d9s9s7.mount: Deactivated successfully. Dec 13 13:27:54.682973 systemd[1]: var-lib-kubelet-pods-87af4ba6\x2d4abb\x2d49f4\x2d8748\x2d18d3ba478286-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Dec 13 13:27:54.683026 systemd[1]: var-lib-kubelet-pods-87af4ba6\x2d4abb\x2d49f4\x2d8748\x2d18d3ba478286-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Dec 13 13:27:54.687649 kubelet[2824]: E1213 13:27:54.687474 2824 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="kube-system/coredns-7db6d8ff4d-pt7lq" podUID="4f9d26cd-c4d1-47f8-9a9a-b3321d7c8878" Dec 13 13:27:54.691977 systemd[1]: Removed slice kubepods-burstable-pod87af4ba6_4abb_49f4_8748_18d3ba478286.slice - libcontainer container kubepods-burstable-pod87af4ba6_4abb_49f4_8748_18d3ba478286.slice. Dec 13 13:27:54.692201 systemd[1]: kubepods-burstable-pod87af4ba6_4abb_49f4_8748_18d3ba478286.slice: Consumed 7.976s CPU time. Dec 13 13:27:54.696458 containerd[1477]: time="2024-12-13T13:27:54.693351273Z" level=info msg="RemoveContainer for \"7eef8c509f1c6c520e52af996fe6b3e92243ab283e5e5a2ee093a1c37553e171\"" Dec 13 13:27:54.705078 containerd[1477]: time="2024-12-13T13:27:54.704934502Z" level=info msg="RemoveContainer for \"7eef8c509f1c6c520e52af996fe6b3e92243ab283e5e5a2ee093a1c37553e171\" returns successfully" Dec 13 13:27:54.706630 kubelet[2824]: I1213 13:27:54.706578 2824 scope.go:117] "RemoveContainer" containerID="590a9dd551ac1e339d1e595c191fa336cd7d22786ab1da6d8b47b543c8234a0b" Dec 13 13:27:54.708817 systemd[1]: Removed slice kubepods-besteffort-podd2fb0668_07bd_4a53_9b0f_ec12c129578d.slice - libcontainer container kubepods-besteffort-podd2fb0668_07bd_4a53_9b0f_ec12c129578d.slice. Dec 13 13:27:54.711960 containerd[1477]: time="2024-12-13T13:27:54.711925303Z" level=info msg="RemoveContainer for \"590a9dd551ac1e339d1e595c191fa336cd7d22786ab1da6d8b47b543c8234a0b\"" Dec 13 13:27:54.715357 containerd[1477]: time="2024-12-13T13:27:54.715305763Z" level=info msg="RemoveContainer for \"590a9dd551ac1e339d1e595c191fa336cd7d22786ab1da6d8b47b543c8234a0b\" returns successfully" Dec 13 13:27:54.716677 kubelet[2824]: I1213 13:27:54.715618 2824 scope.go:117] "RemoveContainer" containerID="ca0716e2d915f9418f463fc7052030f47974ebb73c009c58197e2bed939f0016" Dec 13 13:27:54.719239 containerd[1477]: time="2024-12-13T13:27:54.718929585Z" level=info msg="RemoveContainer for \"ca0716e2d915f9418f463fc7052030f47974ebb73c009c58197e2bed939f0016\"" Dec 13 13:27:54.723429 containerd[1477]: time="2024-12-13T13:27:54.722759727Z" level=info msg="RemoveContainer for \"ca0716e2d915f9418f463fc7052030f47974ebb73c009c58197e2bed939f0016\" returns successfully" Dec 13 13:27:54.725531 kubelet[2824]: I1213 13:27:54.725447 2824 scope.go:117] "RemoveContainer" containerID="28003be96e8c1eaa0f2d3a766ba08814c4e37ed4b1a896dccf7c066b098fd8a0" Dec 13 13:27:54.728707 containerd[1477]: time="2024-12-13T13:27:54.728667882Z" level=info msg="RemoveContainer for \"28003be96e8c1eaa0f2d3a766ba08814c4e37ed4b1a896dccf7c066b098fd8a0\"" Dec 13 13:27:54.732344 containerd[1477]: time="2024-12-13T13:27:54.732150263Z" level=info msg="RemoveContainer for \"28003be96e8c1eaa0f2d3a766ba08814c4e37ed4b1a896dccf7c066b098fd8a0\" returns successfully" Dec 13 13:27:54.732723 kubelet[2824]: I1213 13:27:54.732533 2824 scope.go:117] "RemoveContainer" containerID="36c7bf433ad4194395ef0cec0b50e5506f334bf0955a6a0cea088d55207c9053" Dec 13 13:27:54.735070 containerd[1477]: time="2024-12-13T13:27:54.734813718Z" level=info msg="RemoveContainer for \"36c7bf433ad4194395ef0cec0b50e5506f334bf0955a6a0cea088d55207c9053\"" Dec 13 13:27:54.738160 containerd[1477]: time="2024-12-13T13:27:54.738038578Z" level=info msg="RemoveContainer for \"36c7bf433ad4194395ef0cec0b50e5506f334bf0955a6a0cea088d55207c9053\" returns successfully" Dec 13 13:27:54.739033 kubelet[2824]: I1213 13:27:54.738990 2824 scope.go:117] "RemoveContainer" containerID="7eef8c509f1c6c520e52af996fe6b3e92243ab283e5e5a2ee093a1c37553e171" Dec 13 13:27:54.739547 containerd[1477]: time="2024-12-13T13:27:54.739513186Z" level=error msg="ContainerStatus for \"7eef8c509f1c6c520e52af996fe6b3e92243ab283e5e5a2ee093a1c37553e171\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"7eef8c509f1c6c520e52af996fe6b3e92243ab283e5e5a2ee093a1c37553e171\": not found" Dec 13 13:27:54.740931 kubelet[2824]: E1213 13:27:54.740841 2824 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"7eef8c509f1c6c520e52af996fe6b3e92243ab283e5e5a2ee093a1c37553e171\": not found" containerID="7eef8c509f1c6c520e52af996fe6b3e92243ab283e5e5a2ee093a1c37553e171" Dec 13 13:27:54.741076 kubelet[2824]: I1213 13:27:54.740964 2824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"7eef8c509f1c6c520e52af996fe6b3e92243ab283e5e5a2ee093a1c37553e171"} err="failed to get container status \"7eef8c509f1c6c520e52af996fe6b3e92243ab283e5e5a2ee093a1c37553e171\": rpc error: code = NotFound desc = an error occurred when try to find container \"7eef8c509f1c6c520e52af996fe6b3e92243ab283e5e5a2ee093a1c37553e171\": not found" Dec 13 13:27:54.741109 kubelet[2824]: I1213 13:27:54.741081 2824 scope.go:117] "RemoveContainer" containerID="590a9dd551ac1e339d1e595c191fa336cd7d22786ab1da6d8b47b543c8234a0b" Dec 13 13:27:54.741438 containerd[1477]: time="2024-12-13T13:27:54.741365357Z" level=error msg="ContainerStatus for \"590a9dd551ac1e339d1e595c191fa336cd7d22786ab1da6d8b47b543c8234a0b\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"590a9dd551ac1e339d1e595c191fa336cd7d22786ab1da6d8b47b543c8234a0b\": not found" Dec 13 13:27:54.741873 kubelet[2824]: E1213 13:27:54.741599 2824 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"590a9dd551ac1e339d1e595c191fa336cd7d22786ab1da6d8b47b543c8234a0b\": not found" containerID="590a9dd551ac1e339d1e595c191fa336cd7d22786ab1da6d8b47b543c8234a0b" Dec 13 13:27:54.741873 kubelet[2824]: I1213 13:27:54.741626 2824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"590a9dd551ac1e339d1e595c191fa336cd7d22786ab1da6d8b47b543c8234a0b"} err="failed to get container status \"590a9dd551ac1e339d1e595c191fa336cd7d22786ab1da6d8b47b543c8234a0b\": rpc error: code = NotFound desc = an error occurred when try to find container \"590a9dd551ac1e339d1e595c191fa336cd7d22786ab1da6d8b47b543c8234a0b\": not found" Dec 13 13:27:54.741873 kubelet[2824]: I1213 13:27:54.741643 2824 scope.go:117] "RemoveContainer" containerID="ca0716e2d915f9418f463fc7052030f47974ebb73c009c58197e2bed939f0016" Dec 13 13:27:54.742063 containerd[1477]: time="2024-12-13T13:27:54.742012201Z" level=error msg="ContainerStatus for \"ca0716e2d915f9418f463fc7052030f47974ebb73c009c58197e2bed939f0016\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"ca0716e2d915f9418f463fc7052030f47974ebb73c009c58197e2bed939f0016\": not found" Dec 13 13:27:54.742509 kubelet[2824]: E1213 13:27:54.742214 2824 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"ca0716e2d915f9418f463fc7052030f47974ebb73c009c58197e2bed939f0016\": not found" containerID="ca0716e2d915f9418f463fc7052030f47974ebb73c009c58197e2bed939f0016" Dec 13 13:27:54.742509 kubelet[2824]: I1213 13:27:54.742334 2824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"ca0716e2d915f9418f463fc7052030f47974ebb73c009c58197e2bed939f0016"} err="failed to get container status \"ca0716e2d915f9418f463fc7052030f47974ebb73c009c58197e2bed939f0016\": rpc error: code = NotFound desc = an error occurred when try to find container \"ca0716e2d915f9418f463fc7052030f47974ebb73c009c58197e2bed939f0016\": not found" Dec 13 13:27:54.742509 kubelet[2824]: I1213 13:27:54.742354 2824 scope.go:117] "RemoveContainer" containerID="28003be96e8c1eaa0f2d3a766ba08814c4e37ed4b1a896dccf7c066b098fd8a0" Dec 13 13:27:54.742966 containerd[1477]: time="2024-12-13T13:27:54.742863686Z" level=error msg="ContainerStatus for \"28003be96e8c1eaa0f2d3a766ba08814c4e37ed4b1a896dccf7c066b098fd8a0\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"28003be96e8c1eaa0f2d3a766ba08814c4e37ed4b1a896dccf7c066b098fd8a0\": not found" Dec 13 13:27:54.743230 kubelet[2824]: E1213 13:27:54.743207 2824 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"28003be96e8c1eaa0f2d3a766ba08814c4e37ed4b1a896dccf7c066b098fd8a0\": not found" containerID="28003be96e8c1eaa0f2d3a766ba08814c4e37ed4b1a896dccf7c066b098fd8a0" Dec 13 13:27:54.743827 kubelet[2824]: I1213 13:27:54.743480 2824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"28003be96e8c1eaa0f2d3a766ba08814c4e37ed4b1a896dccf7c066b098fd8a0"} err="failed to get container status \"28003be96e8c1eaa0f2d3a766ba08814c4e37ed4b1a896dccf7c066b098fd8a0\": rpc error: code = NotFound desc = an error occurred when try to find container \"28003be96e8c1eaa0f2d3a766ba08814c4e37ed4b1a896dccf7c066b098fd8a0\": not found" Dec 13 13:27:54.743827 kubelet[2824]: I1213 13:27:54.743515 2824 scope.go:117] "RemoveContainer" containerID="36c7bf433ad4194395ef0cec0b50e5506f334bf0955a6a0cea088d55207c9053" Dec 13 13:27:54.743934 containerd[1477]: time="2024-12-13T13:27:54.743714851Z" level=error msg="ContainerStatus for \"36c7bf433ad4194395ef0cec0b50e5506f334bf0955a6a0cea088d55207c9053\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"36c7bf433ad4194395ef0cec0b50e5506f334bf0955a6a0cea088d55207c9053\": not found" Dec 13 13:27:54.744319 kubelet[2824]: E1213 13:27:54.744113 2824 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"36c7bf433ad4194395ef0cec0b50e5506f334bf0955a6a0cea088d55207c9053\": not found" containerID="36c7bf433ad4194395ef0cec0b50e5506f334bf0955a6a0cea088d55207c9053" Dec 13 13:27:54.744319 kubelet[2824]: I1213 13:27:54.744195 2824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"36c7bf433ad4194395ef0cec0b50e5506f334bf0955a6a0cea088d55207c9053"} err="failed to get container status \"36c7bf433ad4194395ef0cec0b50e5506f334bf0955a6a0cea088d55207c9053\": rpc error: code = NotFound desc = an error occurred when try to find container \"36c7bf433ad4194395ef0cec0b50e5506f334bf0955a6a0cea088d55207c9053\": not found" Dec 13 13:27:54.744319 kubelet[2824]: I1213 13:27:54.744213 2824 scope.go:117] "RemoveContainer" containerID="9fc7582bf65547c02d26deb9fc8a19b164c9cf14800312e5ddcac69a4829f858" Dec 13 13:27:54.746048 containerd[1477]: time="2024-12-13T13:27:54.745976584Z" level=info msg="RemoveContainer for \"9fc7582bf65547c02d26deb9fc8a19b164c9cf14800312e5ddcac69a4829f858\"" Dec 13 13:27:54.749995 containerd[1477]: time="2024-12-13T13:27:54.749938768Z" level=info msg="RemoveContainer for \"9fc7582bf65547c02d26deb9fc8a19b164c9cf14800312e5ddcac69a4829f858\" returns successfully" Dec 13 13:27:54.750427 kubelet[2824]: I1213 13:27:54.750347 2824 scope.go:117] "RemoveContainer" containerID="9fc7582bf65547c02d26deb9fc8a19b164c9cf14800312e5ddcac69a4829f858" Dec 13 13:27:54.750790 containerd[1477]: time="2024-12-13T13:27:54.750699452Z" level=error msg="ContainerStatus for \"9fc7582bf65547c02d26deb9fc8a19b164c9cf14800312e5ddcac69a4829f858\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"9fc7582bf65547c02d26deb9fc8a19b164c9cf14800312e5ddcac69a4829f858\": not found" Dec 13 13:27:54.751062 kubelet[2824]: E1213 13:27:54.750940 2824 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"9fc7582bf65547c02d26deb9fc8a19b164c9cf14800312e5ddcac69a4829f858\": not found" containerID="9fc7582bf65547c02d26deb9fc8a19b164c9cf14800312e5ddcac69a4829f858" Dec 13 13:27:54.751062 kubelet[2824]: I1213 13:27:54.750990 2824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"9fc7582bf65547c02d26deb9fc8a19b164c9cf14800312e5ddcac69a4829f858"} err="failed to get container status \"9fc7582bf65547c02d26deb9fc8a19b164c9cf14800312e5ddcac69a4829f858\": rpc error: code = NotFound desc = an error occurred when try to find container \"9fc7582bf65547c02d26deb9fc8a19b164c9cf14800312e5ddcac69a4829f858\": not found" Dec 13 13:27:55.688234 kubelet[2824]: I1213 13:27:55.688156 2824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87af4ba6-4abb-49f4-8748-18d3ba478286" path="/var/lib/kubelet/pods/87af4ba6-4abb-49f4-8748-18d3ba478286/volumes" Dec 13 13:27:55.689051 kubelet[2824]: I1213 13:27:55.689025 2824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2fb0668-07bd-4a53-9b0f-ec12c129578d" path="/var/lib/kubelet/pods/d2fb0668-07bd-4a53-9b0f-ec12c129578d/volumes" Dec 13 13:27:55.771752 sshd[4419]: Connection closed by 147.75.109.163 port 59298 Dec 13 13:27:55.773200 sshd-session[4417]: pam_unix(sshd:session): session closed for user core Dec 13 13:27:55.777846 systemd-logind[1465]: Session 21 logged out. Waiting for processes to exit. Dec 13 13:27:55.778547 systemd[1]: sshd@20-23.88.121.109:22-147.75.109.163:59298.service: Deactivated successfully. Dec 13 13:27:55.782083 systemd[1]: session-21.scope: Deactivated successfully. Dec 13 13:27:55.782716 systemd[1]: session-21.scope: Consumed 1.130s CPU time. Dec 13 13:27:55.783821 systemd-logind[1465]: Removed session 21. Dec 13 13:27:55.950184 systemd[1]: Started sshd@21-23.88.121.109:22-147.75.109.163:59300.service - OpenSSH per-connection server daemon (147.75.109.163:59300). Dec 13 13:27:56.685595 kubelet[2824]: E1213 13:27:56.685457 2824 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="kube-system/coredns-7db6d8ff4d-pt7lq" podUID="4f9d26cd-c4d1-47f8-9a9a-b3321d7c8878" Dec 13 13:27:56.932115 sshd[4585]: Accepted publickey for core from 147.75.109.163 port 59300 ssh2: RSA SHA256:mujnIWCHI6g7RhOIFUnU4L23td/nvZma7inzBKSUIRw Dec 13 13:27:56.933988 sshd-session[4585]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 13:27:56.938991 systemd-logind[1465]: New session 22 of user core. Dec 13 13:27:56.949008 systemd[1]: Started session-22.scope - Session 22 of User core. Dec 13 13:27:58.605045 kubelet[2824]: I1213 13:27:58.604997 2824 topology_manager.go:215] "Topology Admit Handler" podUID="3dcd3714-4603-447f-a7b7-77789925d92a" podNamespace="kube-system" podName="cilium-xwfn4" Dec 13 13:27:58.605497 kubelet[2824]: E1213 13:27:58.605062 2824 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="87af4ba6-4abb-49f4-8748-18d3ba478286" containerName="cilium-agent" Dec 13 13:27:58.605497 kubelet[2824]: E1213 13:27:58.605075 2824 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="d2fb0668-07bd-4a53-9b0f-ec12c129578d" containerName="cilium-operator" Dec 13 13:27:58.605497 kubelet[2824]: E1213 13:27:58.605080 2824 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="87af4ba6-4abb-49f4-8748-18d3ba478286" containerName="apply-sysctl-overwrites" Dec 13 13:27:58.605497 kubelet[2824]: E1213 13:27:58.605087 2824 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="87af4ba6-4abb-49f4-8748-18d3ba478286" containerName="clean-cilium-state" Dec 13 13:27:58.605497 kubelet[2824]: E1213 13:27:58.605092 2824 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="87af4ba6-4abb-49f4-8748-18d3ba478286" containerName="mount-cgroup" Dec 13 13:27:58.605497 kubelet[2824]: E1213 13:27:58.605098 2824 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="87af4ba6-4abb-49f4-8748-18d3ba478286" containerName="mount-bpf-fs" Dec 13 13:27:58.605497 kubelet[2824]: I1213 13:27:58.605122 2824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2fb0668-07bd-4a53-9b0f-ec12c129578d" containerName="cilium-operator" Dec 13 13:27:58.605497 kubelet[2824]: I1213 13:27:58.605129 2824 memory_manager.go:354] "RemoveStaleState removing state" podUID="87af4ba6-4abb-49f4-8748-18d3ba478286" containerName="cilium-agent" Dec 13 13:27:58.617603 systemd[1]: Created slice kubepods-burstable-pod3dcd3714_4603_447f_a7b7_77789925d92a.slice - libcontainer container kubepods-burstable-pod3dcd3714_4603_447f_a7b7_77789925d92a.slice. Dec 13 13:27:58.685415 kubelet[2824]: E1213 13:27:58.685263 2824 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="kube-system/coredns-7db6d8ff4d-pt7lq" podUID="4f9d26cd-c4d1-47f8-9a9a-b3321d7c8878" Dec 13 13:27:58.708668 kubelet[2824]: I1213 13:27:58.708244 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/3dcd3714-4603-447f-a7b7-77789925d92a-xtables-lock\") pod \"cilium-xwfn4\" (UID: \"3dcd3714-4603-447f-a7b7-77789925d92a\") " pod="kube-system/cilium-xwfn4" Dec 13 13:27:58.708858 kubelet[2824]: I1213 13:27:58.708684 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/3dcd3714-4603-447f-a7b7-77789925d92a-host-proc-sys-net\") pod \"cilium-xwfn4\" (UID: \"3dcd3714-4603-447f-a7b7-77789925d92a\") " pod="kube-system/cilium-xwfn4" Dec 13 13:27:58.708858 kubelet[2824]: I1213 13:27:58.708729 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/3dcd3714-4603-447f-a7b7-77789925d92a-host-proc-sys-kernel\") pod \"cilium-xwfn4\" (UID: \"3dcd3714-4603-447f-a7b7-77789925d92a\") " pod="kube-system/cilium-xwfn4" Dec 13 13:27:58.708858 kubelet[2824]: I1213 13:27:58.708788 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/3dcd3714-4603-447f-a7b7-77789925d92a-cilium-cgroup\") pod \"cilium-xwfn4\" (UID: \"3dcd3714-4603-447f-a7b7-77789925d92a\") " pod="kube-system/cilium-xwfn4" Dec 13 13:27:58.708858 kubelet[2824]: I1213 13:27:58.708851 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/3dcd3714-4603-447f-a7b7-77789925d92a-hostproc\") pod \"cilium-xwfn4\" (UID: \"3dcd3714-4603-447f-a7b7-77789925d92a\") " pod="kube-system/cilium-xwfn4" Dec 13 13:27:58.709098 kubelet[2824]: I1213 13:27:58.708887 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/3dcd3714-4603-447f-a7b7-77789925d92a-cilium-ipsec-secrets\") pod \"cilium-xwfn4\" (UID: \"3dcd3714-4603-447f-a7b7-77789925d92a\") " pod="kube-system/cilium-xwfn4" Dec 13 13:27:58.709098 kubelet[2824]: I1213 13:27:58.708917 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/3dcd3714-4603-447f-a7b7-77789925d92a-cilium-run\") pod \"cilium-xwfn4\" (UID: \"3dcd3714-4603-447f-a7b7-77789925d92a\") " pod="kube-system/cilium-xwfn4" Dec 13 13:27:58.709098 kubelet[2824]: I1213 13:27:58.708951 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckcgd\" (UniqueName: \"kubernetes.io/projected/3dcd3714-4603-447f-a7b7-77789925d92a-kube-api-access-ckcgd\") pod \"cilium-xwfn4\" (UID: \"3dcd3714-4603-447f-a7b7-77789925d92a\") " pod="kube-system/cilium-xwfn4" Dec 13 13:27:58.709098 kubelet[2824]: I1213 13:27:58.708978 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/3dcd3714-4603-447f-a7b7-77789925d92a-bpf-maps\") pod \"cilium-xwfn4\" (UID: \"3dcd3714-4603-447f-a7b7-77789925d92a\") " pod="kube-system/cilium-xwfn4" Dec 13 13:27:58.709098 kubelet[2824]: I1213 13:27:58.709016 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/3dcd3714-4603-447f-a7b7-77789925d92a-cilium-config-path\") pod \"cilium-xwfn4\" (UID: \"3dcd3714-4603-447f-a7b7-77789925d92a\") " pod="kube-system/cilium-xwfn4" Dec 13 13:27:58.709098 kubelet[2824]: I1213 13:27:58.709048 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/3dcd3714-4603-447f-a7b7-77789925d92a-cni-path\") pod \"cilium-xwfn4\" (UID: \"3dcd3714-4603-447f-a7b7-77789925d92a\") " pod="kube-system/cilium-xwfn4" Dec 13 13:27:58.709492 kubelet[2824]: I1213 13:27:58.709080 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/3dcd3714-4603-447f-a7b7-77789925d92a-clustermesh-secrets\") pod \"cilium-xwfn4\" (UID: \"3dcd3714-4603-447f-a7b7-77789925d92a\") " pod="kube-system/cilium-xwfn4" Dec 13 13:27:58.709492 kubelet[2824]: I1213 13:27:58.709114 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3dcd3714-4603-447f-a7b7-77789925d92a-etc-cni-netd\") pod \"cilium-xwfn4\" (UID: \"3dcd3714-4603-447f-a7b7-77789925d92a\") " pod="kube-system/cilium-xwfn4" Dec 13 13:27:58.709492 kubelet[2824]: I1213 13:27:58.709144 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3dcd3714-4603-447f-a7b7-77789925d92a-lib-modules\") pod \"cilium-xwfn4\" (UID: \"3dcd3714-4603-447f-a7b7-77789925d92a\") " pod="kube-system/cilium-xwfn4" Dec 13 13:27:58.709492 kubelet[2824]: I1213 13:27:58.709174 2824 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/3dcd3714-4603-447f-a7b7-77789925d92a-hubble-tls\") pod \"cilium-xwfn4\" (UID: \"3dcd3714-4603-447f-a7b7-77789925d92a\") " pod="kube-system/cilium-xwfn4" Dec 13 13:27:58.773090 sshd[4587]: Connection closed by 147.75.109.163 port 59300 Dec 13 13:27:58.773837 sshd-session[4585]: pam_unix(sshd:session): session closed for user core Dec 13 13:27:58.779333 systemd[1]: sshd@21-23.88.121.109:22-147.75.109.163:59300.service: Deactivated successfully. Dec 13 13:27:58.782674 systemd[1]: session-22.scope: Deactivated successfully. Dec 13 13:27:58.783047 systemd[1]: session-22.scope: Consumed 1.039s CPU time. Dec 13 13:27:58.784452 systemd-logind[1465]: Session 22 logged out. Waiting for processes to exit. Dec 13 13:27:58.785645 systemd-logind[1465]: Removed session 22. Dec 13 13:27:58.908326 kubelet[2824]: E1213 13:27:58.908194 2824 kubelet.go:2900] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Dec 13 13:27:58.923244 containerd[1477]: time="2024-12-13T13:27:58.923129253Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-xwfn4,Uid:3dcd3714-4603-447f-a7b7-77789925d92a,Namespace:kube-system,Attempt:0,}" Dec 13 13:27:58.946159 systemd[1]: Started sshd@22-23.88.121.109:22-147.75.109.163:38726.service - OpenSSH per-connection server daemon (147.75.109.163:38726). Dec 13 13:27:58.950673 containerd[1477]: time="2024-12-13T13:27:58.950219333Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 13:27:58.950673 containerd[1477]: time="2024-12-13T13:27:58.950297933Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 13:27:58.950673 containerd[1477]: time="2024-12-13T13:27:58.950315134Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 13:27:58.950673 containerd[1477]: time="2024-12-13T13:27:58.950399334Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 13:27:58.973035 systemd[1]: Started cri-containerd-43ecc77c0c32d8dc21773b24537ce7f2847b14a848f76b0a8e992a55c8d72cf6.scope - libcontainer container 43ecc77c0c32d8dc21773b24537ce7f2847b14a848f76b0a8e992a55c8d72cf6. Dec 13 13:27:58.999442 containerd[1477]: time="2024-12-13T13:27:58.999383863Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-xwfn4,Uid:3dcd3714-4603-447f-a7b7-77789925d92a,Namespace:kube-system,Attempt:0,} returns sandbox id \"43ecc77c0c32d8dc21773b24537ce7f2847b14a848f76b0a8e992a55c8d72cf6\"" Dec 13 13:27:59.003928 containerd[1477]: time="2024-12-13T13:27:59.003891090Z" level=info msg="CreateContainer within sandbox \"43ecc77c0c32d8dc21773b24537ce7f2847b14a848f76b0a8e992a55c8d72cf6\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Dec 13 13:27:59.017046 containerd[1477]: time="2024-12-13T13:27:59.016999887Z" level=info msg="CreateContainer within sandbox \"43ecc77c0c32d8dc21773b24537ce7f2847b14a848f76b0a8e992a55c8d72cf6\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"859086e1031ffee5d66d7633955f8d11ef28a478618108a5ea5964dc33319fc1\"" Dec 13 13:27:59.017959 containerd[1477]: time="2024-12-13T13:27:59.017926212Z" level=info msg="StartContainer for \"859086e1031ffee5d66d7633955f8d11ef28a478618108a5ea5964dc33319fc1\"" Dec 13 13:27:59.042964 systemd[1]: Started cri-containerd-859086e1031ffee5d66d7633955f8d11ef28a478618108a5ea5964dc33319fc1.scope - libcontainer container 859086e1031ffee5d66d7633955f8d11ef28a478618108a5ea5964dc33319fc1. Dec 13 13:27:59.070243 containerd[1477]: time="2024-12-13T13:27:59.070198641Z" level=info msg="StartContainer for \"859086e1031ffee5d66d7633955f8d11ef28a478618108a5ea5964dc33319fc1\" returns successfully" Dec 13 13:27:59.081961 systemd[1]: cri-containerd-859086e1031ffee5d66d7633955f8d11ef28a478618108a5ea5964dc33319fc1.scope: Deactivated successfully. Dec 13 13:27:59.123927 containerd[1477]: time="2024-12-13T13:27:59.123864197Z" level=info msg="shim disconnected" id=859086e1031ffee5d66d7633955f8d11ef28a478618108a5ea5964dc33319fc1 namespace=k8s.io Dec 13 13:27:59.124377 containerd[1477]: time="2024-12-13T13:27:59.124176559Z" level=warning msg="cleaning up after shim disconnected" id=859086e1031ffee5d66d7633955f8d11ef28a478618108a5ea5964dc33319fc1 namespace=k8s.io Dec 13 13:27:59.124464 containerd[1477]: time="2024-12-13T13:27:59.124234119Z" level=info msg="cleaning up dead shim" namespace=k8s.io Dec 13 13:27:59.714568 containerd[1477]: time="2024-12-13T13:27:59.714414919Z" level=info msg="CreateContainer within sandbox \"43ecc77c0c32d8dc21773b24537ce7f2847b14a848f76b0a8e992a55c8d72cf6\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Dec 13 13:27:59.734029 containerd[1477]: time="2024-12-13T13:27:59.733867914Z" level=info msg="CreateContainer within sandbox \"43ecc77c0c32d8dc21773b24537ce7f2847b14a848f76b0a8e992a55c8d72cf6\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"2c7be16cb9b1a1923ac31b25217460b7d2bc5ca0597b65a5c5e82c1a51a27681\"" Dec 13 13:27:59.734834 containerd[1477]: time="2024-12-13T13:27:59.734788839Z" level=info msg="StartContainer for \"2c7be16cb9b1a1923ac31b25217460b7d2bc5ca0597b65a5c5e82c1a51a27681\"" Dec 13 13:27:59.763058 systemd[1]: Started cri-containerd-2c7be16cb9b1a1923ac31b25217460b7d2bc5ca0597b65a5c5e82c1a51a27681.scope - libcontainer container 2c7be16cb9b1a1923ac31b25217460b7d2bc5ca0597b65a5c5e82c1a51a27681. Dec 13 13:27:59.794595 containerd[1477]: time="2024-12-13T13:27:59.794438311Z" level=info msg="StartContainer for \"2c7be16cb9b1a1923ac31b25217460b7d2bc5ca0597b65a5c5e82c1a51a27681\" returns successfully" Dec 13 13:27:59.804448 systemd[1]: cri-containerd-2c7be16cb9b1a1923ac31b25217460b7d2bc5ca0597b65a5c5e82c1a51a27681.scope: Deactivated successfully. Dec 13 13:27:59.834251 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-2c7be16cb9b1a1923ac31b25217460b7d2bc5ca0597b65a5c5e82c1a51a27681-rootfs.mount: Deactivated successfully. Dec 13 13:27:59.840181 containerd[1477]: time="2024-12-13T13:27:59.840080140Z" level=info msg="shim disconnected" id=2c7be16cb9b1a1923ac31b25217460b7d2bc5ca0597b65a5c5e82c1a51a27681 namespace=k8s.io Dec 13 13:27:59.840540 containerd[1477]: time="2024-12-13T13:27:59.840254781Z" level=warning msg="cleaning up after shim disconnected" id=2c7be16cb9b1a1923ac31b25217460b7d2bc5ca0597b65a5c5e82c1a51a27681 namespace=k8s.io Dec 13 13:27:59.840540 containerd[1477]: time="2024-12-13T13:27:59.840281262Z" level=info msg="cleaning up dead shim" namespace=k8s.io Dec 13 13:27:59.930908 sshd[4610]: Accepted publickey for core from 147.75.109.163 port 38726 ssh2: RSA SHA256:mujnIWCHI6g7RhOIFUnU4L23td/nvZma7inzBKSUIRw Dec 13 13:27:59.932829 sshd-session[4610]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 13:27:59.938021 systemd-logind[1465]: New session 23 of user core. Dec 13 13:27:59.945120 systemd[1]: Started session-23.scope - Session 23 of User core. Dec 13 13:28:00.603169 sshd[4771]: Connection closed by 147.75.109.163 port 38726 Dec 13 13:28:00.604292 sshd-session[4610]: pam_unix(sshd:session): session closed for user core Dec 13 13:28:00.611100 systemd[1]: sshd@22-23.88.121.109:22-147.75.109.163:38726.service: Deactivated successfully. Dec 13 13:28:00.614135 systemd[1]: session-23.scope: Deactivated successfully. Dec 13 13:28:00.616032 systemd-logind[1465]: Session 23 logged out. Waiting for processes to exit. Dec 13 13:28:00.617125 systemd-logind[1465]: Removed session 23. Dec 13 13:28:00.687143 kubelet[2824]: E1213 13:28:00.685619 2824 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="kube-system/coredns-7db6d8ff4d-pt7lq" podUID="4f9d26cd-c4d1-47f8-9a9a-b3321d7c8878" Dec 13 13:28:00.720828 containerd[1477]: time="2024-12-13T13:28:00.720099167Z" level=info msg="CreateContainer within sandbox \"43ecc77c0c32d8dc21773b24537ce7f2847b14a848f76b0a8e992a55c8d72cf6\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Dec 13 13:28:00.749770 containerd[1477]: time="2024-12-13T13:28:00.749712302Z" level=info msg="CreateContainer within sandbox \"43ecc77c0c32d8dc21773b24537ce7f2847b14a848f76b0a8e992a55c8d72cf6\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"a042887b56cd4ec05a83bbead4189bb29043d6478af6d21f50999174354fd7cd\"" Dec 13 13:28:00.750978 containerd[1477]: time="2024-12-13T13:28:00.750946349Z" level=info msg="StartContainer for \"a042887b56cd4ec05a83bbead4189bb29043d6478af6d21f50999174354fd7cd\"" Dec 13 13:28:00.790041 systemd[1]: Started cri-containerd-a042887b56cd4ec05a83bbead4189bb29043d6478af6d21f50999174354fd7cd.scope - libcontainer container a042887b56cd4ec05a83bbead4189bb29043d6478af6d21f50999174354fd7cd. Dec 13 13:28:00.793887 systemd[1]: Started sshd@23-23.88.121.109:22-147.75.109.163:38730.service - OpenSSH per-connection server daemon (147.75.109.163:38730). Dec 13 13:28:00.832345 containerd[1477]: time="2024-12-13T13:28:00.832084067Z" level=info msg="StartContainer for \"a042887b56cd4ec05a83bbead4189bb29043d6478af6d21f50999174354fd7cd\" returns successfully" Dec 13 13:28:00.836627 systemd[1]: cri-containerd-a042887b56cd4ec05a83bbead4189bb29043d6478af6d21f50999174354fd7cd.scope: Deactivated successfully. Dec 13 13:28:00.870984 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a042887b56cd4ec05a83bbead4189bb29043d6478af6d21f50999174354fd7cd-rootfs.mount: Deactivated successfully. Dec 13 13:28:00.878640 containerd[1477]: time="2024-12-13T13:28:00.878548221Z" level=info msg="shim disconnected" id=a042887b56cd4ec05a83bbead4189bb29043d6478af6d21f50999174354fd7cd namespace=k8s.io Dec 13 13:28:00.878869 containerd[1477]: time="2024-12-13T13:28:00.878640022Z" level=warning msg="cleaning up after shim disconnected" id=a042887b56cd4ec05a83bbead4189bb29043d6478af6d21f50999174354fd7cd namespace=k8s.io Dec 13 13:28:00.878869 containerd[1477]: time="2024-12-13T13:28:00.878662302Z" level=info msg="cleaning up dead shim" namespace=k8s.io Dec 13 13:28:01.725170 containerd[1477]: time="2024-12-13T13:28:01.724852006Z" level=info msg="CreateContainer within sandbox \"43ecc77c0c32d8dc21773b24537ce7f2847b14a848f76b0a8e992a55c8d72cf6\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Dec 13 13:28:01.744881 containerd[1477]: time="2024-12-13T13:28:01.744447242Z" level=info msg="CreateContainer within sandbox \"43ecc77c0c32d8dc21773b24537ce7f2847b14a848f76b0a8e992a55c8d72cf6\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"7c005b6326fc885ccdd7f4c4ffca1696b15f527611eaccbc7f97c98a1b9db2ec\"" Dec 13 13:28:01.745797 containerd[1477]: time="2024-12-13T13:28:01.745318607Z" level=info msg="StartContainer for \"7c005b6326fc885ccdd7f4c4ffca1696b15f527611eaccbc7f97c98a1b9db2ec\"" Dec 13 13:28:01.781135 systemd[1]: Started cri-containerd-7c005b6326fc885ccdd7f4c4ffca1696b15f527611eaccbc7f97c98a1b9db2ec.scope - libcontainer container 7c005b6326fc885ccdd7f4c4ffca1696b15f527611eaccbc7f97c98a1b9db2ec. Dec 13 13:28:01.800772 sshd[4794]: Accepted publickey for core from 147.75.109.163 port 38730 ssh2: RSA SHA256:mujnIWCHI6g7RhOIFUnU4L23td/nvZma7inzBKSUIRw Dec 13 13:28:01.803784 sshd-session[4794]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 13:28:01.813987 systemd-logind[1465]: New session 24 of user core. Dec 13 13:28:01.818581 systemd[1]: cri-containerd-7c005b6326fc885ccdd7f4c4ffca1696b15f527611eaccbc7f97c98a1b9db2ec.scope: Deactivated successfully. Dec 13 13:28:01.824023 systemd[1]: Started session-24.scope - Session 24 of User core. Dec 13 13:28:01.828230 containerd[1477]: time="2024-12-13T13:28:01.827234689Z" level=info msg="StartContainer for \"7c005b6326fc885ccdd7f4c4ffca1696b15f527611eaccbc7f97c98a1b9db2ec\" returns successfully" Dec 13 13:28:01.851442 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-7c005b6326fc885ccdd7f4c4ffca1696b15f527611eaccbc7f97c98a1b9db2ec-rootfs.mount: Deactivated successfully. Dec 13 13:28:01.857988 containerd[1477]: time="2024-12-13T13:28:01.857755789Z" level=info msg="shim disconnected" id=7c005b6326fc885ccdd7f4c4ffca1696b15f527611eaccbc7f97c98a1b9db2ec namespace=k8s.io Dec 13 13:28:01.857988 containerd[1477]: time="2024-12-13T13:28:01.857811189Z" level=warning msg="cleaning up after shim disconnected" id=7c005b6326fc885ccdd7f4c4ffca1696b15f527611eaccbc7f97c98a1b9db2ec namespace=k8s.io Dec 13 13:28:01.857988 containerd[1477]: time="2024-12-13T13:28:01.857819269Z" level=info msg="cleaning up dead shim" namespace=k8s.io Dec 13 13:28:02.368776 kubelet[2824]: I1213 13:28:02.368702 2824 setters.go:580] "Node became not ready" node="ci-4186-0-0-9-3a4b2091e9" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2024-12-13T13:28:02Z","lastTransitionTime":"2024-12-13T13:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} Dec 13 13:28:02.685658 kubelet[2824]: E1213 13:28:02.684730 2824 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="kube-system/coredns-7db6d8ff4d-pt7lq" podUID="4f9d26cd-c4d1-47f8-9a9a-b3321d7c8878" Dec 13 13:28:02.732052 containerd[1477]: time="2024-12-13T13:28:02.731657294Z" level=info msg="CreateContainer within sandbox \"43ecc77c0c32d8dc21773b24537ce7f2847b14a848f76b0a8e992a55c8d72cf6\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Dec 13 13:28:02.763344 containerd[1477]: time="2024-12-13T13:28:02.763174560Z" level=info msg="CreateContainer within sandbox \"43ecc77c0c32d8dc21773b24537ce7f2847b14a848f76b0a8e992a55c8d72cf6\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"62509089b5c4aaeae037512fc0e1b5f6b2744c6290562b9cde25f9c2e7209a76\"" Dec 13 13:28:02.764691 containerd[1477]: time="2024-12-13T13:28:02.763835884Z" level=info msg="StartContainer for \"62509089b5c4aaeae037512fc0e1b5f6b2744c6290562b9cde25f9c2e7209a76\"" Dec 13 13:28:02.796007 systemd[1]: Started cri-containerd-62509089b5c4aaeae037512fc0e1b5f6b2744c6290562b9cde25f9c2e7209a76.scope - libcontainer container 62509089b5c4aaeae037512fc0e1b5f6b2744c6290562b9cde25f9c2e7209a76. Dec 13 13:28:02.833918 containerd[1477]: time="2024-12-13T13:28:02.833871296Z" level=info msg="StartContainer for \"62509089b5c4aaeae037512fc0e1b5f6b2744c6290562b9cde25f9c2e7209a76\" returns successfully" Dec 13 13:28:03.153865 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aes-ce)) Dec 13 13:28:03.767720 containerd[1477]: time="2024-12-13T13:28:03.767651911Z" level=info msg="StopPodSandbox for \"ae6a1363bbce38682e7e9f82b14004da2fb101205b6d8d6eb3974ea986c256a3\"" Dec 13 13:28:03.768232 containerd[1477]: time="2024-12-13T13:28:03.768130834Z" level=info msg="TearDown network for sandbox \"ae6a1363bbce38682e7e9f82b14004da2fb101205b6d8d6eb3974ea986c256a3\" successfully" Dec 13 13:28:03.768232 containerd[1477]: time="2024-12-13T13:28:03.768149514Z" level=info msg="StopPodSandbox for \"ae6a1363bbce38682e7e9f82b14004da2fb101205b6d8d6eb3974ea986c256a3\" returns successfully" Dec 13 13:28:03.770106 containerd[1477]: time="2024-12-13T13:28:03.768805198Z" level=info msg="RemovePodSandbox for \"ae6a1363bbce38682e7e9f82b14004da2fb101205b6d8d6eb3974ea986c256a3\"" Dec 13 13:28:03.770106 containerd[1477]: time="2024-12-13T13:28:03.768834518Z" level=info msg="Forcibly stopping sandbox \"ae6a1363bbce38682e7e9f82b14004da2fb101205b6d8d6eb3974ea986c256a3\"" Dec 13 13:28:03.770106 containerd[1477]: time="2024-12-13T13:28:03.768884438Z" level=info msg="TearDown network for sandbox \"ae6a1363bbce38682e7e9f82b14004da2fb101205b6d8d6eb3974ea986c256a3\" successfully" Dec 13 13:28:03.772881 containerd[1477]: time="2024-12-13T13:28:03.772836941Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"ae6a1363bbce38682e7e9f82b14004da2fb101205b6d8d6eb3974ea986c256a3\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Dec 13 13:28:03.773069 containerd[1477]: time="2024-12-13T13:28:03.773049463Z" level=info msg="RemovePodSandbox \"ae6a1363bbce38682e7e9f82b14004da2fb101205b6d8d6eb3974ea986c256a3\" returns successfully" Dec 13 13:28:03.773730 containerd[1477]: time="2024-12-13T13:28:03.773708267Z" level=info msg="StopPodSandbox for \"08a2754e454ae96132c191313465a0ff2e9480b13d88079a4a438ec4f2571f81\"" Dec 13 13:28:03.774036 containerd[1477]: time="2024-12-13T13:28:03.774006308Z" level=info msg="TearDown network for sandbox \"08a2754e454ae96132c191313465a0ff2e9480b13d88079a4a438ec4f2571f81\" successfully" Dec 13 13:28:03.774126 containerd[1477]: time="2024-12-13T13:28:03.774110749Z" level=info msg="StopPodSandbox for \"08a2754e454ae96132c191313465a0ff2e9480b13d88079a4a438ec4f2571f81\" returns successfully" Dec 13 13:28:03.774860 containerd[1477]: time="2024-12-13T13:28:03.774552551Z" level=info msg="RemovePodSandbox for \"08a2754e454ae96132c191313465a0ff2e9480b13d88079a4a438ec4f2571f81\"" Dec 13 13:28:03.774860 containerd[1477]: time="2024-12-13T13:28:03.774575752Z" level=info msg="Forcibly stopping sandbox \"08a2754e454ae96132c191313465a0ff2e9480b13d88079a4a438ec4f2571f81\"" Dec 13 13:28:03.774860 containerd[1477]: time="2024-12-13T13:28:03.774615432Z" level=info msg="TearDown network for sandbox \"08a2754e454ae96132c191313465a0ff2e9480b13d88079a4a438ec4f2571f81\" successfully" Dec 13 13:28:03.778968 containerd[1477]: time="2024-12-13T13:28:03.778928017Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"08a2754e454ae96132c191313465a0ff2e9480b13d88079a4a438ec4f2571f81\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Dec 13 13:28:03.779313 containerd[1477]: time="2024-12-13T13:28:03.779131618Z" level=info msg="RemovePodSandbox \"08a2754e454ae96132c191313465a0ff2e9480b13d88079a4a438ec4f2571f81\" returns successfully" Dec 13 13:28:06.128005 systemd-networkd[1383]: lxc_health: Link UP Dec 13 13:28:06.136942 systemd-networkd[1383]: lxc_health: Gained carrier Dec 13 13:28:06.717382 systemd[1]: run-containerd-runc-k8s.io-62509089b5c4aaeae037512fc0e1b5f6b2744c6290562b9cde25f9c2e7209a76-runc.afDxbf.mount: Deactivated successfully. Dec 13 13:28:06.951572 kubelet[2824]: I1213 13:28:06.951194 2824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-xwfn4" podStartSLOduration=8.951174195 podStartE2EDuration="8.951174195s" podCreationTimestamp="2024-12-13 13:27:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-12-13 13:28:03.764483292 +0000 UTC m=+360.171756818" watchObservedRunningTime="2024-12-13 13:28:06.951174195 +0000 UTC m=+363.358447681" Dec 13 13:28:07.375480 systemd-networkd[1383]: lxc_health: Gained IPv6LL Dec 13 13:28:09.025725 systemd[1]: run-containerd-runc-k8s.io-62509089b5c4aaeae037512fc0e1b5f6b2744c6290562b9cde25f9c2e7209a76-runc.fbrvCo.mount: Deactivated successfully. Dec 13 13:28:13.587952 sshd[4865]: Connection closed by 147.75.109.163 port 38730 Dec 13 13:28:13.589112 sshd-session[4794]: pam_unix(sshd:session): session closed for user core Dec 13 13:28:13.594232 systemd[1]: sshd@23-23.88.121.109:22-147.75.109.163:38730.service: Deactivated successfully. Dec 13 13:28:13.596826 systemd[1]: session-24.scope: Deactivated successfully. Dec 13 13:28:13.600376 systemd-logind[1465]: Session 24 logged out. Waiting for processes to exit. Dec 13 13:28:13.602118 systemd-logind[1465]: Removed session 24.